torch>=2.0
transformers>=4.51.0
safetensors>=0.4
accelerate>=0.34.0

[dev]
pytest
ruff

[fla]
flash-linear-attention
