httpx
llama-stack-client<0.4,>=0.3.5

[dev]
pytest>=7.0
pytest-cov>=4.0
ruff>=0.1.0
build>=1.0
python-dotenv>=1.0

[langgraph]
langgraph>=0.2
