besimray commited on
Commit
322ebcd
·
verified ·
1 Parent(s): 9c66697

End of training

Browse files
Files changed (2) hide show
  1. README.md +16 -10
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: peft
3
  license: llama3.2
4
- base_model: unsloth/Llama-3.2-1B-Instruct
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
@@ -19,13 +19,19 @@ should probably proofread and complete it, then remove this comment. -->
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
- base_model: unsloth/Llama-3.2-1B-Instruct
23
  bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
  - path: mhenrichsen/alpaca_2k_test
28
- type: alpaca
 
 
 
 
 
 
29
  debug: null
30
  deepspeed: null
31
  early_stopping_patience: null
@@ -77,7 +83,7 @@ wandb_entity: besimray24-rayon
77
  wandb_mode: online
78
  wandb_project: Public_TuningSN
79
  wandb_run: miner_id_24
80
- wandb_runid: 383a850e-bb15-45a2-8f4b-fc96eb001a74
81
  warmup_steps: 10
82
  weight_decay: 0.0
83
  xformers_attention: null
@@ -88,9 +94,9 @@ xformers_attention: null
88
 
89
  # test
90
 
91
- This model is a fine-tuned version of [unsloth/Llama-3.2-1B-Instruct](https://huggingface.co/unsloth/Llama-3.2-1B-Instruct) on the None dataset.
92
  It achieves the following results on the evaluation set:
93
- - Loss: 1.2167
94
 
95
  ## Model description
96
 
@@ -124,10 +130,10 @@ The following hyperparameters were used during training:
124
 
125
  | Training Loss | Epoch | Step | Validation Loss |
126
  |:-------------:|:------:|:----:|:---------------:|
127
- | 1.3218 | 0.0042 | 1 | 1.2625 |
128
- | 1.3071 | 0.0126 | 3 | 1.2579 |
129
- | 1.4942 | 0.0253 | 6 | 1.2140 |
130
- | 1.277 | 0.0379 | 9 | 1.2167 |
131
 
132
 
133
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  license: llama3.2
4
+ base_model: unsloth/Llama-3.2-3B-Instruct
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
+ base_model: unsloth/Llama-3.2-3B-Instruct
23
  bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
  - path: mhenrichsen/alpaca_2k_test
28
+ type:
29
+ field_input: input
30
+ field_instruction: instruction
31
+ field_output: output
32
+ field_system: text
33
+ system_format: '{system}'
34
+ system_prompt: you are helpful
35
  debug: null
36
  deepspeed: null
37
  early_stopping_patience: null
 
83
  wandb_mode: online
84
  wandb_project: Public_TuningSN
85
  wandb_run: miner_id_24
86
+ wandb_runid: 123e4567-e89b-12d3-a456-426614174000
87
  warmup_steps: 10
88
  weight_decay: 0.0
89
  xformers_attention: null
 
94
 
95
  # test
96
 
97
+ This model is a fine-tuned version of [unsloth/Llama-3.2-3B-Instruct](https://huggingface.co/unsloth/Llama-3.2-3B-Instruct) on the None dataset.
98
  It achieves the following results on the evaluation set:
99
+ - Loss: 0.0439
100
 
101
  ## Model description
102
 
 
130
 
131
  | Training Loss | Epoch | Step | Validation Loss |
132
  |:-------------:|:------:|:----:|:---------------:|
133
+ | 0.1517 | 0.0042 | 1 | 0.2442 |
134
+ | 0.1181 | 0.0126 | 3 | 0.2362 |
135
+ | 0.3502 | 0.0253 | 6 | 0.1496 |
136
+ | 0.0495 | 0.0379 | 9 | 0.0439 |
137
 
138
 
139
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61ba026082313c7f15d589fe716210021e2d1334718de7cb4e272b7552bcf546
3
- size 45169354
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43bf6fd3cc6bb3b2f61eb935b2c760530f315b059c95c5db87d7f3f2698f317b
3
+ size 97396522