error577 commited on
Commit
38ea6b4
·
verified ·
1 Parent(s): d2c46d6

End of training

Browse files
README.md CHANGED
@@ -43,7 +43,7 @@ early_stopping_patience: null
43
  eval_max_new_tokens: 128
44
  eval_table_size: null
45
  evals_per_epoch: 1
46
- flash_attention: false
47
  fp16: null
48
  fsdp: null
49
  fsdp_config: null
@@ -103,7 +103,7 @@ xformers_attention: null
103
 
104
  This model is a fine-tuned version of [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: nan
107
 
108
  ## Model description
109
 
@@ -137,11 +137,11 @@ The following hyperparameters were used during training:
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
- | 0.0 | 0.0002 | 1 | nan |
141
- | 0.1966 | 0.0055 | 25 | nan |
142
- | 4.8102 | 0.0111 | 50 | nan |
143
- | 2.9287 | 0.0166 | 75 | nan |
144
- | 3.5741 | 0.0221 | 100 | nan |
145
 
146
 
147
  ### Framework versions
 
43
  eval_max_new_tokens: 128
44
  eval_table_size: null
45
  evals_per_epoch: 1
46
+ flash_attention: true
47
  fp16: null
48
  fsdp: null
49
  fsdp_config: null
 
103
 
104
  This model is a fine-tuned version of [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 0.9404
107
 
108
  ## Model description
109
 
 
137
 
138
  | Training Loss | Epoch | Step | Validation Loss |
139
  |:-------------:|:------:|:----:|:---------------:|
140
+ | 1.7563 | 0.0002 | 1 | 3.0182 |
141
+ | 1.531 | 0.0055 | 25 | 1.2687 |
142
+ | 1.1124 | 0.0111 | 50 | 1.0121 |
143
+ | 0.5517 | 0.0166 | 75 | 0.9560 |
144
+ | 0.3658 | 0.0221 | 100 | 0.9404 |
145
 
146
 
147
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "k_proj",
24
- "o_proj",
25
- "q_proj",
26
- "down_proj",
27
  "gate_proj",
 
 
28
  "up_proj",
29
- "v_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
 
23
  "gate_proj",
24
+ "q_proj",
25
+ "v_proj",
26
  "up_proj",
27
+ "o_proj",
28
+ "down_proj",
29
+ "k_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:858c3d4df2784005ed0cd000a01797eb9b52c0c425686bf7aaaaec1e018f7844
3
  size 90258378
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:abf7dce0f2eb241853271543c313030dcb7fa9fb841ff314e90c8222190df620
3
  size 90258378
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9cdd439a76295fd6d09be599ad07a660ed6668b045a9dd94068eb8b9e1779ca3
3
  size 90207248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3abd600633004f737757a13247a23693caa58d5002d763d8e51a00c693dab67b
3
  size 90207248
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05530ac48b6077bf2d56f539fa4393812b10b3dc652ed678c61e51a2fe9ac582
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74fb7478fc1012b216a24ea1fb3ee5f5f6fc81742a0ea3f826b6fc1e8aae82d5
3
  size 6776