{ "model": "unsloth/gemma-2-9b", "training_type": "SFT and DPO", "max_seq_length": 1024, "dtype": "float16", "training_args": { "sft": { "learning_rate": 2e-06, "batch_size": 4, "gradient_accumulation_steps": 4, "num_train_epochs": 3, "scheduler": "cosine", "optim": "adamw_8bit", "seed": 2802 } } }