error577 commited on
Commit
2eb9bf8
·
verified ·
1 Parent(s): 34f9ab1

End of training

Browse files
README.md CHANGED
@@ -42,7 +42,7 @@ early_stopping_patience: null
42
  eval_max_new_tokens: 128
43
  eval_table_size: null
44
  evals_per_epoch: 1
45
- flash_attention: false
46
  fp16: null
47
  fsdp: null
48
  fsdp_config: null
@@ -102,7 +102,7 @@ xformers_attention: null
102
 
103
  This model is a fine-tuned version of [unsloth/Qwen2-7B](https://huggingface.co/unsloth/Qwen2-7B) on the None dataset.
104
  It achieves the following results on the evaluation set:
105
- - Loss: nan
106
 
107
  ## Model description
108
 
@@ -136,11 +136,11 @@ The following hyperparameters were used during training:
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
- | 0.0 | 0.0044 | 1 | nan |
140
- | 0.0 | 0.1106 | 25 | nan |
141
- | 0.0 | 0.2211 | 50 | nan |
142
- | 0.0 | 0.3317 | 75 | nan |
143
- | 0.0 | 0.4422 | 100 | nan |
144
 
145
 
146
  ### Framework versions
 
42
  eval_max_new_tokens: 128
43
  eval_table_size: null
44
  evals_per_epoch: 1
45
+ flash_attention: true
46
  fp16: null
47
  fsdp: null
48
  fsdp_config: null
 
102
 
103
  This model is a fine-tuned version of [unsloth/Qwen2-7B](https://huggingface.co/unsloth/Qwen2-7B) on the None dataset.
104
  It achieves the following results on the evaluation set:
105
+ - Loss: 0.8507
106
 
107
  ## Model description
108
 
 
136
 
137
  | Training Loss | Epoch | Step | Validation Loss |
138
  |:-------------:|:------:|:----:|:---------------:|
139
+ | 1.8272 | 0.0044 | 1 | 1.9260 |
140
+ | 0.8706 | 0.1106 | 25 | 0.9072 |
141
+ | 0.9693 | 0.2211 | 50 | 0.8707 |
142
+ | 0.7109 | 0.3317 | 75 | 0.8530 |
143
+ | 0.6786 | 0.4422 | 100 | 0.8507 |
144
 
145
 
146
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "v_proj",
24
  "o_proj",
 
 
25
  "gate_proj",
26
- "k_proj",
27
- "q_proj",
28
  "up_proj",
29
- "down_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "k_proj",
24
  "o_proj",
25
+ "v_proj",
26
+ "down_proj",
27
  "gate_proj",
 
 
28
  "up_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9606a4718cbf605fac2f9f3a7bd9f241c176317d4e2e367804b2d84c1210b79b
3
  size 323103018
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d44aa5af6009b5829f29130b49aa01c47d83e67dbe1c0ca8bbe4382d9a61c689
3
  size 323103018
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22f5f7fd591b8777ebb93ab7cdbca3a92cf15051b9528de45a8917b17a81d33d
3
  size 323014168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bf3a39e95c63fd317b4511b04b2b99d1f0c8e2c21346f3119f4e957cbbfc4d3
3
  size 323014168
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b0f8735a73365b512b9f48f10b65433a4cbab77cedd6badac3f38ce1ab2b45ce
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b5a647679363a8a6842c95e00e29abca1fa0c4d3c9b3a2d819be7d3f45d4cd
3
  size 6776