Metadata-Version: 2.4
Name: olive-ai
Version: 0.12.1
Summary: Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs.
Home-page: https://microsoft.github.io/Olive/
Download-URL: https://github.com/microsoft/Olive/tags
Author: Microsoft Corporation
Author-email: olivedevteam@microsoft.com
License: MIT License
Classifier: Development Status :: 3 - Alpha
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: MIT License
Classifier: Operating System :: POSIX :: Linux
Classifier: Operating System :: Microsoft :: Windows
Classifier: Topic :: Scientific/Engineering
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Classifier: Topic :: Software Development
Classifier: Topic :: Software Development :: Libraries
Classifier: Topic :: Software Development :: Libraries :: Python Modules
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 3 :: Only
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: 3.11
Classifier: Programming Language :: Python :: 3.12
Classifier: Programming Language :: Python :: 3.13
Requires-Python: >=3.10
License-File: LICENSE
License-File: NOTICE.txt
Requires-Dist: hf-xet
Requires-Dist: numpy
Requires-Dist: onnx
Requires-Dist: onnx_ir>=0.1.2
Requires-Dist: onnxscript>=0.5.3
Requires-Dist: opentelemetry-sdk>=1.39.1
Requires-Dist: optuna
Requires-Dist: pandas
Requires-Dist: pydantic>=2.0
Requires-Dist: pyyaml
Requires-Dist: questionary
Requires-Dist: torch
Requires-Dist: torchmetrics>=1.0.0
Requires-Dist: transformers
Provides-Extra: aimet-onnx
Requires-Dist: aimet-onnx>=2.15.0; extra == "aimet-onnx"
Provides-Extra: auto-opt
Requires-Dist: optimum; extra == "auto-opt"
Provides-Extra: azureml
Requires-Dist: azure-ai-ml>=1.11.1; extra == "azureml"
Requires-Dist: azure-identity; extra == "azureml"
Provides-Extra: bnb
Requires-Dist: bitsandbytes; extra == "bnb"
Requires-Dist: triton; extra == "bnb"
Provides-Extra: capture-onnx-graph
Requires-Dist: optimum; extra == "capture-onnx-graph"
Provides-Extra: cpu
Requires-Dist: onnxruntime; extra == "cpu"
Provides-Extra: directml
Requires-Dist: onnxruntime-directml; extra == "directml"
Provides-Extra: docker
Requires-Dist: docker; extra == "docker"
Provides-Extra: shared-cache
Requires-Dist: azure-identity; extra == "shared-cache"
Requires-Dist: azure-storage-blob; extra == "shared-cache"
Provides-Extra: finetune
Requires-Dist: optimum; extra == "finetune"
Requires-Dist: accelerate>=0.30.0; extra == "finetune"
Requires-Dist: peft; extra == "finetune"
Requires-Dist: scipy; extra == "finetune"
Requires-Dist: bitsandbytes; extra == "finetune"
Requires-Dist: triton; extra == "finetune"
Provides-Extra: flash-attn
Requires-Dist: flash_attn; extra == "flash-attn"
Provides-Extra: gpu
Requires-Dist: onnxruntime-gpu; extra == "gpu"
Provides-Extra: inc
Requires-Dist: neural-compressor; extra == "inc"
Provides-Extra: lora
Requires-Dist: accelerate>=0.30.0; extra == "lora"
Requires-Dist: peft; extra == "lora"
Requires-Dist: scipy; extra == "lora"
Provides-Extra: diffusers
Requires-Dist: accelerate>=0.30.0; extra == "diffusers"
Requires-Dist: peft; extra == "diffusers"
Requires-Dist: diffusers; extra == "diffusers"
Provides-Extra: nvmo
Requires-Dist: nvidia-modelopt[onnx]; extra == "nvmo"
Provides-Extra: openvino
Requires-Dist: openvino>=2025.4.1; extra == "openvino"
Requires-Dist: nncf>=2.19.0; extra == "openvino"
Requires-Dist: numpy<2.0; extra == "openvino"
Requires-Dist: optimum[openvino]>=2.1.0; extra == "openvino"
Requires-Dist: onnxruntime-openvino; extra == "openvino"
Provides-Extra: optimum
Requires-Dist: optimum; extra == "optimum"
Provides-Extra: qairt
Requires-Dist: qairt-dev[onnx]; extra == "qairt"
Provides-Extra: qnn
Requires-Dist: onnxruntime-qnn; extra == "qnn"
Provides-Extra: tf
Requires-Dist: tensorflow==1.15.0; extra == "tf"
Provides-Extra: torch-tensorrt
Requires-Dist: torch-tensorrt; extra == "torch-tensorrt"
Provides-Extra: tune-session-params
Requires-Dist: psutil; extra == "tune-session-params"
Dynamic: author
Dynamic: author-email
Dynamic: classifier
Dynamic: description
Dynamic: download-url
Dynamic: home-page
Dynamic: license
Dynamic: license-file
Dynamic: provides-extra
Dynamic: requires-dist
Dynamic: requires-python
Dynamic: summary

Olive: Simplify ML Model Finetuning, Conversion, Quantization, and Optimization for CPUs, GPUs and NPUs
