SourCoachSauers commited on
Commit
edeee06
·
verified ·
1 Parent(s): a14e87f

color_descriptions

Browse files
README.md CHANGED
@@ -39,6 +39,8 @@ The following hyperparameters were used during training:
39
  - train_batch_size: 4
40
  - eval_batch_size: 4
41
  - seed: 42
 
 
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
  - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.03
 
39
  - train_batch_size: 4
40
  - eval_batch_size: 4
41
  - seed: 42
42
+ - gradient_accumulation_steps: 4
43
+ - total_train_batch_size: 16
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.03
adapter_config.json CHANGED
@@ -16,7 +16,7 @@
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 32,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:538fdbff1d625944a1d58b882700b018e3989acea3756f273c076d5dca8fc455
3
- size 28865136
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c774b1d1dbab6275f08d4c335a48e408a3a60d6f29feac6c148d9b8634d4c6e2
3
+ size 57701064
runs/Jul17_02-57-24_ip-172-18-164-229.us-west-2.compute.internal/events.out.tfevents.1721185059.ip-172-18-164-229.us-west-2.compute.internal.13851.5 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:393bc8cf20d1a35c4aa4359d36431aa52dad9259b0f0137b67e01986a07b6243
3
+ size 6723
runs/Jul17_02-57-55_ip-172-18-164-229.us-west-2.compute.internal/events.out.tfevents.1721185079.ip-172-18-164-229.us-west-2.compute.internal.13851.6 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4087ce3b2e928dbfb468aff886d6de1b93540c528dded30e93d65de939e98511
3
+ size 8379
runs/Jul17_03-10-32_ip-172-18-164-229.us-west-2.compute.internal/events.out.tfevents.1721185837.ip-172-18-164-229.us-west-2.compute.internal.13851.7 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb08145a7af7d586f1e3e6a9866aaf0f48bd7573b9b98672dabcdcf6ba383de9
3
+ size 452432
tokenizer.json CHANGED
@@ -67,12 +67,6 @@
67
  "id": "A",
68
  "type_id": 0
69
  }
70
- },
71
- {
72
- "SpecialToken": {
73
- "id": "</s>",
74
- "type_id": 0
75
- }
76
  }
77
  ],
78
  "pair": [
@@ -88,12 +82,6 @@
88
  "type_id": 0
89
  }
90
  },
91
- {
92
- "SpecialToken": {
93
- "id": "</s>",
94
- "type_id": 0
95
- }
96
- },
97
  {
98
  "SpecialToken": {
99
  "id": "<s>",
@@ -105,24 +93,9 @@
105
  "id": "B",
106
  "type_id": 1
107
  }
108
- },
109
- {
110
- "SpecialToken": {
111
- "id": "</s>",
112
- "type_id": 1
113
- }
114
  }
115
  ],
116
  "special_tokens": {
117
- "</s>": {
118
- "id": "</s>",
119
- "ids": [
120
- 2
121
- ],
122
- "tokens": [
123
- "</s>"
124
- ]
125
- },
126
  "<s>": {
127
  "id": "<s>",
128
  "ids": [
 
67
  "id": "A",
68
  "type_id": 0
69
  }
 
 
 
 
 
 
70
  }
71
  ],
72
  "pair": [
 
82
  "type_id": 0
83
  }
84
  },
 
 
 
 
 
 
85
  {
86
  "SpecialToken": {
87
  "id": "<s>",
 
93
  "id": "B",
94
  "type_id": 1
95
  }
 
 
 
 
 
 
96
  }
97
  ],
98
  "special_tokens": {
 
 
 
 
 
 
 
 
 
99
  "<s>": {
100
  "id": "<s>",
101
  "ids": [
tokenizer_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "add_bos_token": true,
3
- "add_eos_token": true,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
 
1
  {
2
  "add_bos_token": true,
3
+ "add_eos_token": false,
4
  "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19f5daeac2a52babf39bbbd785344377fed2a39aa23a1208f0751347b5c9e72f
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c51b4f76cd7106c343ff68fe7e955f3a205f65d0661177b6c663fc3211c67ae
3
  size 5368