huggingface-hub>=1.4.1
llama-cpp-python[server]>=0.3.16
mlx-lm>=0.30.7
mlx-vlm>=0.3.12
ollama>=0.6.1
openai>=2.24.0
pillow>=12.1.1
pyaml>=26.2.1
pymupdf>=1.27.1
requests>=2.32.5
torchvision>=0.25.0
twine>=6.2.0
