{ "base_model_name_or_path": "/workspace/miluser/kimura/LoRAMoE/output/20250219_test_for_math/sft_lora_model/models--llm-jp--llm-jp-3-13b-instruct3/snapshots/76aa77ba36e2c2a17d94635cea5ade03777c9869", "bias": "none", "blc_alpha": 0.3, "blc_weight": 0.3, "enable_lora": null, "fan_in_fan_out": false, "inference_mode": true, "lora_alpha": 32.0, "lora_dropout": 0.05, "lora_nums": 6, "merge_weights": false, "modules_to_save": null, "peft_type": "LORA", "r": 4, "target_modules": [ "gate_proj", "down_proj", "up_proj" ], "task_type": "CAUSAL_LM" }