ghost613 commited on
Commit
ef63775
·
verified ·
1 Parent(s): 10afecf

phi3_on_korean_summary

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- license: other
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
@@ -12,12 +12,11 @@ model-index:
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/ghost_in_the_lab/Korean-fine-tune-models/runs/g1zumzc4?nw=nwuserghost_in_the_lab)
16
  # phi3_on_korean_summary
17
 
18
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.5903
21
 
22
  ## Model description
23
 
@@ -37,24 +36,52 @@ More information needed
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 5e-05
40
- - train_batch_size: 1
41
- - eval_batch_size: 1
42
  - seed: 42
43
  - gradient_accumulation_steps: 5
44
- - total_train_batch_size: 5
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 50
48
- - training_steps: 1400
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
 
54
  ### Framework versions
55
 
56
- - PEFT 0.7.1
57
- - Transformers 4.41.0.dev0
58
- - Pytorch 2.1.2
59
- - Datasets 2.15.0
60
- - Tokenizers 0.19.1
 
1
  ---
2
+ license: mit
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
 
15
  # phi3_on_korean_summary
16
 
17
  This model is a fine-tuned version of [microsoft/Phi-3-mini-4k-instruct](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct) on the None dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.6044
20
 
21
  ## Model description
22
 
 
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 5e-05
39
+ - train_batch_size: 2
40
+ - eval_batch_size: 2
41
  - seed: 42
42
  - gradient_accumulation_steps: 5
43
+ - total_train_batch_size: 10
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: linear
46
  - lr_scheduler_warmup_steps: 50
47
+ - training_steps: 500
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Training results
51
 
52
+ | Training Loss | Epoch | Step | Validation Loss |
53
+ |:-------------:|:-----:|:----:|:---------------:|
54
+ | 1.2363 | 0.26 | 20 | 1.1356 |
55
+ | 1.0765 | 0.53 | 40 | 1.0056 |
56
+ | 0.9569 | 0.79 | 60 | 0.8976 |
57
+ | 0.8639 | 1.05 | 80 | 0.8329 |
58
+ | 0.8059 | 1.32 | 100 | 0.7890 |
59
+ | 0.7728 | 1.58 | 120 | 0.7568 |
60
+ | 0.7474 | 1.84 | 140 | 0.7310 |
61
+ | 0.7053 | 2.11 | 160 | 0.7122 |
62
+ | 0.6654 | 2.37 | 180 | 0.6928 |
63
+ | 0.6464 | 2.63 | 200 | 0.6775 |
64
+ | 0.6407 | 2.89 | 220 | 0.6652 |
65
+ | 0.6192 | 3.16 | 240 | 0.6582 |
66
+ | 0.5811 | 3.42 | 260 | 0.6478 |
67
+ | 0.5817 | 3.68 | 280 | 0.6382 |
68
+ | 0.5717 | 3.95 | 300 | 0.6308 |
69
+ | 0.5493 | 4.21 | 320 | 0.6277 |
70
+ | 0.5352 | 4.47 | 340 | 0.6202 |
71
+ | 0.5287 | 4.74 | 360 | 0.6155 |
72
+ | 0.5204 | 5.0 | 380 | 0.6105 |
73
+ | 0.4992 | 5.26 | 400 | 0.6132 |
74
+ | 0.4891 | 5.53 | 420 | 0.6090 |
75
+ | 0.4872 | 5.79 | 440 | 0.6060 |
76
+ | 0.4875 | 6.05 | 460 | 0.6039 |
77
+ | 0.4629 | 6.32 | 480 | 0.6061 |
78
+ | 0.4782 | 6.58 | 500 | 0.6044 |
79
+
80
 
81
  ### Framework versions
82
 
83
+ - PEFT 0.8.2
84
+ - Transformers 4.38.2
85
+ - Pytorch 2.1.2+cu121
86
+ - Datasets 2.17.0
87
+ - Tokenizers 0.15.0
adapter_config.json CHANGED
@@ -19,10 +19,11 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "o_proj",
23
  "qkv_proj",
24
  "down_proj",
 
25
  "gate_up_proj"
26
  ],
27
- "task_type": "CAUSAL_LM"
 
28
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "qkv_proj",
23
  "down_proj",
24
+ "o_proj",
25
  "gate_up_proj"
26
  ],
27
+ "task_type": "CAUSAL_LM",
28
+ "use_rslora": false
29
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d93fd911e76ba4aaf4dc925090fbe94c3e8f5ffde030ed875ddffb6098233679
3
  size 100697728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dd2d1d50d827e6f916b726a5f4e46e1e97935ef0652d28306cec87b3d86a934
3
  size 100697728
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d48a4bfd03c5b213eadb0d4ea8bc6e9b8912d11c9e1ec2de5cca71a449f7138b
3
- size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f04837b222bab4c6b3f0d7d3bf4b2968b57842b3a1400556cb6b7fd37c2f35cd
3
+ size 4920