# file: /Users/mfwolffe/GithubOrgs/tenseleyFlow/DocumentLanguageModel/src/dlm/doc/schema.py
# hypothesis_version: 6.152.1

[0.0, 5e-06, 0.0002, 0.0003, 0.01, 0.05, 0.1, 0.5, 1.0, 2.0, 200, 256, 1024, 2048, 10000, 32768, '**/*', 'Q4_K_M', 'Q5_K_M', 'Q6_K', 'Q8_0', 'adamw_bnb_8bit', 'adamw_torch', 'adapter', 'adapters', 'after', 'auto', 'base', 'bf16', 'constant', 'cosine', 'dapt', 'dlm_id', 'dlm_version', 'dora', 'dpo', 'forbid', 'fp16', 'fp32', 'galore_adamw', 'galore_adamw_8bit', 'grad_accum', 'hinge', 'ipo', 'learning_rate', 'linear', 'lora', 'lora_alpha', 'lora_dropout', 'lora_r', 'micro_batch_size', 'orpo', 'paged_adamw_8bit', 'permissive', 'pre_adapter', 'qlora', 'sft', 'sigmoid', 'strict', 'target_modules']