mistyAbbish commited on
Commit
b522f17
·
verified ·
1 Parent(s): 235ab8d

Training in progress, step 20

Browse files
adapter_config.json CHANGED
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "v_proj",
27
- "q_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "q_proj",
27
+ "v_proj"
28
  ],
29
  "task_type": "CAUSAL_LM",
30
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:547331fe40ff72c01c4929ba3a7d3eab5f05aa0cae1da0166ddacda72b3a8f67
3
  size 10107280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb4fa4bb2d2af1e69f992f16127ea520a11cef8cbbf45dc1a38be61d055928c7
3
  size 10107280
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f945e9dcfeee93b9412e3f2a6bcc945a69702c787e44402ec734a8c47bc0aa72
3
  size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:266561a0de9936fafe969d9ddae8a9a31693a03d7fb773b71f20d739d43f6256
3
  size 5624