somaia02 commited on
Commit
65e80ca
·
1 Parent(s): d058be0

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -12,12 +12,12 @@
12
  "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
- "r": 8,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "q_proj",
20
  "k_proj",
 
21
  "v_proj"
22
  ],
23
  "task_type": "CAUSAL_LM"
 
12
  "lora_dropout": 0.05,
13
  "modules_to_save": null,
14
  "peft_type": "LORA",
15
+ "r": 16,
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
 
19
  "k_proj",
20
+ "q_proj",
21
  "v_proj"
22
  ],
23
  "task_type": "CAUSAL_LM"
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cd42a6a3b012049dbfcf0d4ad62b630ef8dca1676e140ca7ca180a24e13c910d
3
- size 2669168
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a8a809056b973f3df74a8e0401d2f6b3f48855ca88f92c074a282b2fc872456
3
+ size 5323528
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef926b73ee7f8ea582bb0c8e88b44eeac71091525992f56ddfde5d64524b7acf
3
  size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c82f8daa53495bfef11faca7a8d954f8023465412fa138e9c80f9e8382c966a8
3
  size 4600