PEFT
Safetensors
qwen2
axolotl
Generated from Trainer
Files changed (1) hide show
  1. README.md +155 -141
README.md CHANGED
@@ -1,142 +1,156 @@
1
- ---
2
- library_name: peft
3
- license: apache-2.0
4
- base_model: Qwen/Qwen2.5-7B
5
- tags:
6
- - axolotl
7
- - generated_from_trainer
8
- datasets:
9
- - sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data
10
- model-index:
11
- - name: reasoning-v0.2-qwen2.5-7b
12
- results: []
13
- ---
14
-
15
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
- should probably proofread and complete it, then remove this comment. -->
17
-
18
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
19
- <details><summary>See axolotl config</summary>
20
-
21
- axolotl version: `0.7.0`
22
- ```yaml
23
- base_model: Qwen/Qwen2.5-7B
24
- hub_model_id: sumukshashidhar-testing/reasoning-v0.2-qwen2.5-7b
25
- trust_remote_code: true
26
-
27
- load_in_8bit: false
28
- load_in_4bit: false
29
- strict: false
30
- bf16: true
31
- hf_use_auth_token: true
32
-
33
- plugins:
34
- - axolotl.integrations.liger.LigerPlugin
35
- liger_rope: true
36
- liger_rms_norm: true
37
- liger_glu_activation: true
38
- liger_layer_norm: true
39
- liger_fused_linear_cross_entropy: true
40
- save_safetensors:
41
-
42
- datasets:
43
- - path: sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data
44
- type: completion
45
- field: text
46
- dataset_prepared_path: .axolotl_cache_data/reasoning-rerankers
47
- shuffle_merged_datasets: true
48
- # dataset_exact_deduplication: true
49
- val_set_size: 0.05
50
- output_dir: /scratch/reasoning-reankers/reasoning-v0.1-qwen2.5-7b
51
- push_dataset_to_hub: sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data-in-progress
52
-
53
- sequence_length: 2048
54
- sample_packing: true
55
- pad_to_sequence_len: true
56
-
57
- adapter: lora
58
- lora_r: 256
59
- lora_alpha: 32
60
- lora_dropout: 0.05
61
- peft_use_rslora: true
62
- lora_target_linear: true
63
-
64
- gradient_accumulation_steps: 1
65
- micro_batch_size: 32
66
- eval_batch_size: 1
67
- num_epochs: 3
68
- learning_rate: 5e-4
69
- warmup_ratio: 0.05
70
- evals_per_epoch: 2
71
- saves_per_epoch: 2
72
- gradient_checkpointing: true
73
- lr_scheduler: cosine
74
- optimizer: paged_adamw_8bit
75
-
76
- profiler_steps: 100
77
- save_safetensors: true
78
- train_on_inputs: true
79
- wandb_project: reasoning-rerankers
80
- wandb_name: rr-qwen-7b
81
- deepspeed: zero1.json
82
-
83
- ```
84
-
85
- </details><br>
86
-
87
- # reasoning-v0.2-qwen2.5-7b
88
-
89
- This model is a fine-tuned version of [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) on the sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data dataset.
90
- It achieves the following results on the evaluation set:
91
- - Loss: 0.4119
92
-
93
- ## Model description
94
-
95
- More information needed
96
-
97
- ## Intended uses & limitations
98
-
99
- More information needed
100
-
101
- ## Training and evaluation data
102
-
103
- More information needed
104
-
105
- ## Training procedure
106
-
107
- ### Training hyperparameters
108
-
109
- The following hyperparameters were used during training:
110
- - learning_rate: 0.0005
111
- - train_batch_size: 32
112
- - eval_batch_size: 1
113
- - seed: 42
114
- - distributed_type: multi-GPU
115
- - num_devices: 8
116
- - total_train_batch_size: 256
117
- - total_eval_batch_size: 8
118
- - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
119
- - lr_scheduler_type: cosine
120
- - lr_scheduler_warmup_steps: 49
121
- - num_epochs: 3.0
122
-
123
- ### Training results
124
-
125
- | Training Loss | Epoch | Step | Validation Loss |
126
- |:-------------:|:------:|:----:|:---------------:|
127
- | No log | 0.0030 | 1 | 2.2497 |
128
- | 0.51 | 0.5 | 166 | 0.7306 |
129
- | 0.2733 | 1.0 | 332 | 0.5004 |
130
- | 0.1938 | 1.5 | 498 | 0.4445 |
131
- | 0.1783 | 2.0 | 664 | 0.4152 |
132
- | 0.1446 | 2.5 | 830 | 0.4147 |
133
- | 0.1424 | 3.0 | 996 | 0.4119 |
134
-
135
-
136
- ### Framework versions
137
-
138
- - PEFT 0.14.0
139
- - Transformers 4.48.3
140
- - Pytorch 2.4.0
141
- - Datasets 3.2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
  - Tokenizers 0.21.1
 
1
+ ---
2
+ library_name: peft
3
+ license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-7B
5
+ tags:
6
+ - axolotl
7
+ - generated_from_trainer
8
+ datasets:
9
+ - sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data
10
+ language:
11
+ - zho
12
+ - eng
13
+ - fra
14
+ - spa
15
+ - por
16
+ - deu
17
+ - ita
18
+ - rus
19
+ - jpn
20
+ - kor
21
+ - vie
22
+ - tha
23
+ - ara
24
+ model-index:
25
+ - name: reasoning-v0.2-qwen2.5-7b
26
+ results: []
27
+ ---
28
+
29
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
+ should probably proofread and complete it, then remove this comment. -->
31
+
32
+ [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
33
+ <details><summary>See axolotl config</summary>
34
+
35
+ axolotl version: `0.7.0`
36
+ ```yaml
37
+ base_model: Qwen/Qwen2.5-7B
38
+ hub_model_id: sumukshashidhar-testing/reasoning-v0.2-qwen2.5-7b
39
+ trust_remote_code: true
40
+
41
+ load_in_8bit: false
42
+ load_in_4bit: false
43
+ strict: false
44
+ bf16: true
45
+ hf_use_auth_token: true
46
+
47
+ plugins:
48
+ - axolotl.integrations.liger.LigerPlugin
49
+ liger_rope: true
50
+ liger_rms_norm: true
51
+ liger_glu_activation: true
52
+ liger_layer_norm: true
53
+ liger_fused_linear_cross_entropy: true
54
+ save_safetensors:
55
+
56
+ datasets:
57
+ - path: sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data
58
+ type: completion
59
+ field: text
60
+ dataset_prepared_path: .axolotl_cache_data/reasoning-rerankers
61
+ shuffle_merged_datasets: true
62
+ # dataset_exact_deduplication: true
63
+ val_set_size: 0.05
64
+ output_dir: /scratch/reasoning-reankers/reasoning-v0.1-qwen2.5-7b
65
+ push_dataset_to_hub: sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data-in-progress
66
+
67
+ sequence_length: 2048
68
+ sample_packing: true
69
+ pad_to_sequence_len: true
70
+
71
+ adapter: lora
72
+ lora_r: 256
73
+ lora_alpha: 32
74
+ lora_dropout: 0.05
75
+ peft_use_rslora: true
76
+ lora_target_linear: true
77
+
78
+ gradient_accumulation_steps: 1
79
+ micro_batch_size: 32
80
+ eval_batch_size: 1
81
+ num_epochs: 3
82
+ learning_rate: 5e-4
83
+ warmup_ratio: 0.05
84
+ evals_per_epoch: 2
85
+ saves_per_epoch: 2
86
+ gradient_checkpointing: true
87
+ lr_scheduler: cosine
88
+ optimizer: paged_adamw_8bit
89
+
90
+ profiler_steps: 100
91
+ save_safetensors: true
92
+ train_on_inputs: true
93
+ wandb_project: reasoning-rerankers
94
+ wandb_name: rr-qwen-7b
95
+ deepspeed: zero1.json
96
+
97
+ ```
98
+
99
+ </details><br>
100
+
101
+ # reasoning-v0.2-qwen2.5-7b
102
+
103
+ This model is a fine-tuned version of [Qwen/Qwen2.5-7B](https://huggingface.co/Qwen/Qwen2.5-7B) on the sumukshashidhar-testing/reasoning-rerankers-relevance-sft-data dataset.
104
+ It achieves the following results on the evaluation set:
105
+ - Loss: 0.4119
106
+
107
+ ## Model description
108
+
109
+ More information needed
110
+
111
+ ## Intended uses & limitations
112
+
113
+ More information needed
114
+
115
+ ## Training and evaluation data
116
+
117
+ More information needed
118
+
119
+ ## Training procedure
120
+
121
+ ### Training hyperparameters
122
+
123
+ The following hyperparameters were used during training:
124
+ - learning_rate: 0.0005
125
+ - train_batch_size: 32
126
+ - eval_batch_size: 1
127
+ - seed: 42
128
+ - distributed_type: multi-GPU
129
+ - num_devices: 8
130
+ - total_train_batch_size: 256
131
+ - total_eval_batch_size: 8
132
+ - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
133
+ - lr_scheduler_type: cosine
134
+ - lr_scheduler_warmup_steps: 49
135
+ - num_epochs: 3.0
136
+
137
+ ### Training results
138
+
139
+ | Training Loss | Epoch | Step | Validation Loss |
140
+ |:-------------:|:------:|:----:|:---------------:|
141
+ | No log | 0.0030 | 1 | 2.2497 |
142
+ | 0.51 | 0.5 | 166 | 0.7306 |
143
+ | 0.2733 | 1.0 | 332 | 0.5004 |
144
+ | 0.1938 | 1.5 | 498 | 0.4445 |
145
+ | 0.1783 | 2.0 | 664 | 0.4152 |
146
+ | 0.1446 | 2.5 | 830 | 0.4147 |
147
+ | 0.1424 | 3.0 | 996 | 0.4119 |
148
+
149
+
150
+ ### Framework versions
151
+
152
+ - PEFT 0.14.0
153
+ - Transformers 4.48.3
154
+ - Pytorch 2.4.0
155
+ - Datasets 3.2.0
156
  - Tokenizers 0.21.1