natsu0wu commited on
Commit
2ee8906
·
verified ·
1 Parent(s): 9874da4

Upload folder using huggingface_hub

Browse files
adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
  "up_proj",
25
- "gate_proj",
26
  "Wqkv",
 
27
  "down_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
24
+ "o_proj",
25
  "Wqkv",
26
+ "gate_proj",
27
  "down_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d69d87ffe679162fcd6c1417586ae283f7aa05de424e6e247c272b183730204
3
  size 340808816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c653323e8da28fbb868ad2f81198f303bd89f299eb88397203332e532218a1
3
  size 340808816
phi_sft/checkpoint-21/adapter_config.json CHANGED
@@ -20,10 +20,10 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
  "up_proj",
25
- "gate_proj",
26
  "Wqkv",
 
27
  "down_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
24
+ "o_proj",
25
  "Wqkv",
26
+ "gate_proj",
27
  "down_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
phi_sft/checkpoint-21/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0d69d87ffe679162fcd6c1417586ae283f7aa05de424e6e247c272b183730204
3
  size 340808816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d4c653323e8da28fbb868ad2f81198f303bd89f299eb88397203332e532218a1
3
  size 340808816
phi_sft/checkpoint-21/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd8b1a4e3405a74c3ce40cab1b6f0dbebc3c43b8ce9531f37c3f194d5261ae26
3
  size 173249466
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47f7b91246b53fb888d3f8ab9983dbcf0286787d69149a95745755a73390b06b
3
  size 173249466
phi_sft/checkpoint-21/trainer_state.json CHANGED
@@ -10,16 +10,16 @@
10
  "log_history": [
11
  {
12
  "epoch": 1.4035087719298245,
13
- "grad_norm": 0.3721645474433899,
14
  "learning_rate": 1.0000000000000002e-06,
15
  "loss": 1.1129,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 2.807017543859649,
20
- "grad_norm": 0.25281819701194763,
21
  "learning_rate": 2.0000000000000003e-06,
22
- "loss": 1.1335,
23
  "step": 20
24
  }
25
  ],
 
10
  "log_history": [
11
  {
12
  "epoch": 1.4035087719298245,
13
+ "grad_norm": 0.2112729400396347,
14
  "learning_rate": 1.0000000000000002e-06,
15
  "loss": 1.1129,
16
  "step": 10
17
  },
18
  {
19
  "epoch": 2.807017543859649,
20
+ "grad_norm": 0.24816565215587616,
21
  "learning_rate": 2.0000000000000003e-06,
22
+ "loss": 1.1337,
23
  "step": 20
24
  }
25
  ],
phi_sft/checkpoint-21/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:285bd8fb225ee0af23eb0b593556bb1aee77579379c6e8735ea8c4ea028ec6eb
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:660e00c8924c1ca69c5c471e9f192e23d4be5cc016db2182b7d705a67e822cf5
3
  size 5432
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:285bd8fb225ee0af23eb0b593556bb1aee77579379c6e8735ea8c4ea028ec6eb
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:660e00c8924c1ca69c5c471e9f192e23d4be5cc016db2182b7d705a67e822cf5
3
  size 5432