transformers>=4.36.0
huggingface-hub>=0.20.0
torch>=2.0.0

[all]
llama-cpp-python>=0.2.0
auto-gptq>=0.7.0
autoawq>=0.2.0
datasets>=2.14.0

[awq]
autoawq>=0.2.0

[gguf]
llama-cpp-python>=0.2.0

[gptq]
auto-gptq>=0.7.0
datasets>=2.14.0
