llama-cpp-python>=0.3.19
requests>=2.33.0
