Muhammad7865253
commited on
Create model_config.json
Browse files- model_config.json +62 -0
model_config.json
ADDED
@@ -0,0 +1,62 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_name": "unsloth/Qwen2-1.5B-Instruct",
|
3 |
+
"max_seq_length": 2048,
|
4 |
+
"dtype": null,
|
5 |
+
"load_in_4bit": false,
|
6 |
+
"lora_params": {
|
7 |
+
"r": 16,
|
8 |
+
"target_modules": [
|
9 |
+
"q_proj",
|
10 |
+
"k_proj",
|
11 |
+
"v_proj",
|
12 |
+
"o_proj",
|
13 |
+
"gate_proj",
|
14 |
+
"up_proj",
|
15 |
+
"down_proj"
|
16 |
+
],
|
17 |
+
"lora_alpha": 16,
|
18 |
+
"lora_dropout": 0,
|
19 |
+
"bias": "none",
|
20 |
+
"use_gradient_checkpointing": "unsloth",
|
21 |
+
"random_state": 3407,
|
22 |
+
"use_rslora": false,
|
23 |
+
"loftq_config": null
|
24 |
+
},
|
25 |
+
"training_args": {
|
26 |
+
"per_device_train_batch_size": 4,
|
27 |
+
"gradient_accumulation_steps": 4,
|
28 |
+
"warmup_steps": 5,
|
29 |
+
"num_train_epochs": 2,
|
30 |
+
"learning_rate": 0.0002,
|
31 |
+
"fp16": false,
|
32 |
+
"bf16": true,
|
33 |
+
"logging_steps": 5,
|
34 |
+
"optim": "adamw_8bit",
|
35 |
+
"weight_decay": 0.01,
|
36 |
+
"lr_scheduler_type": "linear",
|
37 |
+
"seed": 3407,
|
38 |
+
"output_dir": "outputs",
|
39 |
+
"report_to": "none"
|
40 |
+
},
|
41 |
+
"data_config": {
|
42 |
+
"dataset_name": "ruslanmv/ai-medical-chatbot",
|
43 |
+
"dataset_split": "train",
|
44 |
+
"dataset_rows": 20000,
|
45 |
+
"conversation_format": {
|
46 |
+
"instruction": "Provide an instruction to the model, e.g., 'Provide advice on coping with stress.'",
|
47 |
+
"input": "Include user details or context, e.g., 'I've been feeling overwhelmed with work and personal responsibilities.'",
|
48 |
+
"response": "The model's response based on the instruction and input."
|
49 |
+
},
|
50 |
+
"max_seq_length": 2048,
|
51 |
+
"train_on_responses_only": false
|
52 |
+
},
|
53 |
+
"inference_config": {
|
54 |
+
"temperature": 1.0,
|
55 |
+
"min_p": 0.9,
|
56 |
+
"max_new_tokens": 512
|
57 |
+
},
|
58 |
+
"alpaca_prompt": {
|
59 |
+
"template": "If you are a medical chatbot, plz provide answer to user for his query\n\n### Instruction:\n{}\n\n### Input:\n{}\n\n### Response:\n{}",
|
60 |
+
"eos_token": true
|
61 |
+
}
|
62 |
+
}
|