
[all]
prompt-protector[anthropic,ollama,onnx,openai,otel,presidio,redis,schema,transformers,vault,yaml]

[anthropic]
anthropic>=0.40

[dev]
pytest>=8
pytest-asyncio>=0.23
respx>=0.21
pyyaml>=6
ruff>=0.6
mypy>=1.10

[llamacpp]
llama-cpp-python>=0.2

[local-all]
prompt-protector[ollama,onnx,presidio,transformers]

[ollama]
httpx>=0.27

[onnx]
onnxruntime>=1.17

[openai]
openai>=1.40

[otel]
opentelemetry-api>=1.20

[presidio]
presidio-analyzer>=2.2
presidio-anonymizer>=2.2
spacy>=3.7

[redis]
redis>=5

[schema]
jsonschema>=4
pydantic>=2

[server]
fastapi>=0.110
uvicorn>=0.30
pydantic>=2
pyyaml>=6

[spacy]
spacy>=3.7

[transformers]
transformers>=4.40
torch>=2.0

[vault]
cryptography>=42

[yaml]
pyyaml>=6
