# file: /Users/mfwolffe/GithubOrgs/tenseleyFlow/DocumentLanguageModel/src/dlm/export/targets/vllm.py
# hypothesis_version: 6.152.1

[493, 8000, '$SCRIPT_DIR/', '*', '--dtype', '--enable-lora', '--host', '--lora-modules', '--max-model-len', '--port', '--revision', '--served-model-name', '0', '127.0.0.1', '8000', '=', '=$SCRIPT_DIR/', 'adapter', 'adapter_version', 'adapters', 'arm64', 'auto', 'context_length', 'darwin', 'dtype', 'environment', 'export_manifest.json', 'hf', 'host', 'lora_modules', 'max_model_len', 'mixed', 'model', 'module_specs', 'name', 'path', 'port', 'revision', 'runtime_env', 'serve', 'served_model_name', 'target', 'v', 'vllm', 'vllm_config.json', 'vllm_launch.sh']