# file: /Users/mfwolffe/GithubOrgs/tenseleyFlow/DocumentLanguageModel/src/dlm/base_models/registry.py
# hypothesis_version: 6.152.1

[0.27, 0.72, 1.0, 2.5, 3.1, 3.4, 4.0, 4.4, 4.5, 5.2, 6.2, 6.5, 7.6, 8.0, 14.6, 15.5, 16.0, 16.5, 18.0, 30.0, 48.0, 93.4, 224, 256, 448, 576, 672, 750, 1024, 1540, 2048, 3025, 4096, 8192, 16000, 32768, 65536, 131072, 135000000, 360000000, 500000000, 1000000000, 1500000000, 1700000000, 2000000000, 2200000000, 2600000000, 2900000000, 3000000000, 3800000000, 4000000000, 7000000000, 8000000000, 8400000000, 9000000000, 24000000000, 46700000000, '<IMG_CONTEXT>', '<image>', '<|AUDIO|>', '<|image_pad|>', 'Apache-2.0', 'Gemma', 'Gemma2ForCausalLM', 'InternVLChatModel', 'LlamaForCausalLM', 'MIT', 'MixtralForCausalLM', 'Olmo2ForCausalLM', 'Other', 'Phi3ForCausalLM', 'Qwen/Qwen3-1.7B', 'Qwen/Qwen3-4B', 'Qwen/Qwen3-8B', 'Qwen2ForCausalLM', 'Qwen3ForCausalLM', 'SmolLM3ForCausalLM', '[IMG]', 'audio-language', 'chatml', 'down_proj', 'dynamic', 'fixed', 'gate_up_proj', 'gemma', 'gemma-2-2b-it', 'gemma-2-9b-it', 'gemma2', 'google/gemma-2-2b-it', 'google/gemma-2-9b-it', 'internvl2', 'internvl2-2b', 'internvl3', 'internvl3-2b', 'k_proj', 'llama', 'llama-3.2-1b', 'llama-3.2-3b', 'llama-bpe', 'llama3', 'mistral', 'mistral3', 'o_proj', 'olmo-2-7b-instruct', 'olmo2', 'paligemma', 'paligemma-3b-mix-224', 'phi-2', 'phi-3.5-mini', 'phi-4-mini-reasoning', 'phi3', 'phi4mini', 'q_proj', 'qkv_proj', 'qwen2', 'qwen2-audio', 'qwen2-vl', 'qwen2-vl-2b-instruct', 'qwen2.5-0.5b', 'qwen2.5-1.5b', 'qwen2.5-3b', 'qwen2.5-coder-1.5b', 'qwen3', 'qwen3-1.7b', 'qwen3-1.7b-thinking', 'qwen3-4b', 'qwen3-8b', 'qwen3thinking', 'smollm', 'smollm2-1.7b', 'smollm2-135m', 'smollm2-360m', 'smollm3', 'smollm3-3b', 'superbpe', 'tekken', 'text-moe', 'v_proj', 'vision-language']