Tippawan commited on
Commit
34bcd31
·
verified ·
1 Parent(s): e999a61

End of training

Browse files
Files changed (2) hide show
  1. README.md +12 -11
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -1,12 +1,12 @@
1
  ---
2
  library_name: peft
3
  license: other
4
- base_model: SeaLLMs/SeaLLM3-7B-Chat
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
- - name: proof-reading-SeaLLM3-7B-Chat-3090-v11
10
  results: []
11
  ---
12
 
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  axolotl version: `0.5.0`
20
  ```yaml
21
- base_model: SeaLLMs/SeaLLM3-7B-Chat
22
  trust_remote_code: true
23
 
24
  load_in_8bit: false
@@ -26,7 +26,7 @@ load_in_4bit: true
26
  strict: false
27
 
28
  datasets:
29
- - path: Tippawan/p11-seallm
30
  type: chat_template
31
  conversation: chatml
32
  field_messages: messages
@@ -41,7 +41,7 @@ eval_sample_packing: false
41
  pad_to_sequence_len: false
42
 
43
  push_to_hub: true
44
- hub_model_id: Tippawan/proof-reading-SeaLLM3-7B-Chat-3090-v11 # Replace with your Hugging Face repo ID
45
  use_auth_token: true # Ensure you have set your Hugging Face API token in the environment
46
  hub_private_repo: true # Set to true if you want the repository to be private
47
  hub_strategy: all_checkpoints
@@ -49,14 +49,14 @@ save_total_limit: 3
49
  load_best_model_at_end: true
50
 
51
  adapter: lora
52
- lora_model_dir: Tippawan/proof-reading-SeaLLM3-7B-Chat-3090-v9
53
  lora_r: 16
54
  lora_alpha: 32
55
  lora_dropout: 0.05
56
  lora_target_linear: true
57
  lora_fan_in_fan_out:
58
 
59
- wandb_project: proof-reading-SeaLLM3-7B-Chat-3090-v11
60
  wandb_entity:
61
  wandb_watch:
62
  wandb_name:
@@ -64,7 +64,7 @@ wandb_log_model:
64
 
65
  gradient_accumulation_steps: 4
66
  micro_batch_size: 2
67
- num_epochs: 1 #editted 3
68
  optimizer: adamw_torch
69
  lr_scheduler: cosine
70
  learning_rate: 0.0002
@@ -92,13 +92,14 @@ weight_decay: 0.0
92
  fsdp:
93
  fsdp_config:
94
  special_tokens:
 
95
  ```
96
 
97
  </details><br>
98
 
99
- # proof-reading-SeaLLM3-7B-Chat-3090-v11
100
 
101
- This model is a fine-tuned version of [SeaLLMs/SeaLLM3-7B-Chat](https://huggingface.co/SeaLLMs/SeaLLM3-7B-Chat) on the None dataset.
102
 
103
  ## Model description
104
 
@@ -126,7 +127,7 @@ The following hyperparameters were used during training:
126
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
127
  - lr_scheduler_type: cosine
128
  - lr_scheduler_warmup_steps: 10
129
- - num_epochs: 1
130
 
131
  ### Training results
132
 
 
1
  ---
2
  library_name: peft
3
  license: other
4
+ base_model: SeaLLMs/SeaLLMs-v3-7B-Chat
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
8
  model-index:
9
+ - name: proof-reading-hl-v1
10
  results: []
11
  ---
12
 
 
18
 
19
  axolotl version: `0.5.0`
20
  ```yaml
21
+ base_model: SeaLLMs/SeaLLMs-v3-7B-Chat
22
  trust_remote_code: true
23
 
24
  load_in_8bit: false
 
26
  strict: false
27
 
28
  datasets:
29
+ - path: Tippawan/hl-2
30
  type: chat_template
31
  conversation: chatml
32
  field_messages: messages
 
41
  pad_to_sequence_len: false
42
 
43
  push_to_hub: true
44
+ hub_model_id: Tippawan/proof-reading-hl-v1 # Replace with your Hugging Face repo ID
45
  use_auth_token: true # Ensure you have set your Hugging Face API token in the environment
46
  hub_private_repo: true # Set to true if you want the repository to be private
47
  hub_strategy: all_checkpoints
 
49
  load_best_model_at_end: true
50
 
51
  adapter: lora
52
+ lora_model_dir:
53
  lora_r: 16
54
  lora_alpha: 32
55
  lora_dropout: 0.05
56
  lora_target_linear: true
57
  lora_fan_in_fan_out:
58
 
59
+ wandb_project: proof-reading-hl-v1
60
  wandb_entity:
61
  wandb_watch:
62
  wandb_name:
 
64
 
65
  gradient_accumulation_steps: 4
66
  micro_batch_size: 2
67
+ num_epochs: 10 #editted 3
68
  optimizer: adamw_torch
69
  lr_scheduler: cosine
70
  learning_rate: 0.0002
 
92
  fsdp:
93
  fsdp_config:
94
  special_tokens:
95
+
96
  ```
97
 
98
  </details><br>
99
 
100
+ # proof-reading-hl-v1
101
 
102
+ This model is a fine-tuned version of [SeaLLMs/SeaLLMs-v3-7B-Chat](https://huggingface.co/SeaLLMs/SeaLLMs-v3-7B-Chat) on the None dataset.
103
 
104
  ## Model description
105
 
 
127
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
128
  - lr_scheduler_type: cosine
129
  - lr_scheduler_warmup_steps: 10
130
+ - num_epochs: 10
131
 
132
  ### Training results
133
 
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5be0710fe18dc5c0bce44d297de83dbbff8402c49f8a3cf6c7284c445680f90f
3
  size 161621802
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:005442da11bab27e4487afab889afb5a1983f44da1667266a9bd4c6f9cb933d1
3
  size 161621802