dyang415 commited on
Commit
9f8eb20
·
verified ·
1 Parent(s): 00a8ec9

Training in progress, epoch 0

Browse files
adapter_config.json CHANGED
@@ -19,10 +19,10 @@
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "q_proj",
23
  "v_proj",
24
  "o_proj",
25
- "k_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
 
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
22
  "v_proj",
23
  "o_proj",
24
+ "k_proj",
25
+ "q_proj"
26
  ],
27
  "task_type": "CAUSAL_LM"
28
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2a126b3dff573bc42c185fdc14d10f4007712a89a09afdec39645fda79b8dac
3
  size 109086416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffc51da30d7ee80dcd5c370b0c5edffe4e852ffa5c4f39653383a6e6859844fa
3
  size 109086416
config.json CHANGED
@@ -17,7 +17,7 @@
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
  "num_local_experts": 8,
20
- "output_router_logits": true,
21
  "quantization_config": {
22
  "bnb_4bit_compute_dtype": "bfloat16",
23
  "bnb_4bit_quant_type": "nf4",
 
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
  "num_local_experts": 8,
20
+ "output_router_logits": false,
21
  "quantization_config": {
22
  "bnb_4bit_compute_dtype": "bfloat16",
23
  "bnb_4bit_quant_type": "nf4",
runs/Feb29_20-51-41_azure-jap/events.out.tfevents.1709239903.azure-jap.6633.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b603ea22b78be3982ed28d43c9b08c46b0a20d9f0dcfedb6e0a0a8fab683f3c
3
+ size 15241
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3f4e75195439e3cdd0b24ae52a89f22f6c2bf1aae525a1b48ac39d438d6902c
3
  size 4859
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a876d3ad1a3da34bc655ebf92da2818115160b7167df404045d8b809728134d
3
  size 4859