# file: /Users/mfwolffe/GithubOrgs/tenseleyFlow/DocumentLanguageModel/src/dlm/export/targets/llama_server.py
# hypothesis_version: 6.152.1

[493, '$SCRIPT_DIR/', '--api-key', '--chat-template-file', '--ctx-size', '--host', '--lora', '--model', '--port', '127.0.0.1', '8000', 'adapter.gguf', 'adapter_dir', 'adapter_gguf_path', 'artifacts', 'base.', 'chat-template.jinja', 'chat_template', 'context_length', 'disabled', 'llama-server', 'model_path', 'spec', 'utf-8', 'vendor_override']