washimneupane commited on
Commit
fb9d8bd
·
verified ·
1 Parent(s): ee9ee28

Training in progress, step 6000

Browse files
adapter_config.json CHANGED
@@ -1,28 +1,51 @@
1
  {
2
- "alpha_pattern": {},
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "gpt2",
5
- "bias": "none",
6
  "fan_in_fan_out": true,
7
- "inference_mode": false,
8
- "init_lora_weights": true,
9
- "layers_pattern": null,
10
- "layers_to_transform": null,
11
- "loftq_config": {},
12
- "lora_alpha": 32,
13
- "lora_dropout": 0.01,
14
- "megatron_config": null,
15
- "megatron_core": "megatron.core",
 
 
 
 
 
 
 
16
  "modules_to_save": null,
17
- "peft_type": "LORA",
18
- "r": 8,
19
- "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
22
- "c_proj",
23
- "c_fc",
24
- "c_attn"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
25
  ],
26
- "task_type": "CAUSAL_LM",
27
- "use_rslora": false
28
  }
 
1
  {
 
2
  "auto_mapping": null,
3
  "base_model_name_or_path": "gpt2",
 
4
  "fan_in_fan_out": true,
5
+ "feedforward_modules": [
6
+ "transformer.h.0.attn.c_proj",
7
+ "transformer.h.5.attn.c_proj",
8
+ "transformer.h.11.attn.c_proj",
9
+ "transformer.h.9.attn.c_proj",
10
+ "transformer.h.2.attn.c_proj",
11
+ "transformer.h.8.attn.c_proj",
12
+ "transformer.h.10.attn.c_proj",
13
+ "transformer.h.1.attn.c_proj",
14
+ "transformer.h.6.attn.c_proj",
15
+ "transformer.h.3.attn.c_proj",
16
+ "transformer.h.4.attn.c_proj",
17
+ "transformer.h.7.attn.c_proj"
18
+ ],
19
+ "inference_mode": true,
20
+ "init_ia3_weights": true,
21
  "modules_to_save": null,
22
+ "peft_type": "IA3",
 
 
23
  "revision": null,
24
  "target_modules": [
25
+ "transformer.h.3.attn.c_attn",
26
+ "transformer.h.5.attn.c_proj",
27
+ "transformer.h.6.attn.c_proj",
28
+ "transformer.h.4.attn.c_attn",
29
+ "transformer.h.11.attn.c_proj",
30
+ "transformer.h.5.attn.c_attn",
31
+ "transformer.h.8.attn.c_attn",
32
+ "transformer.h.0.attn.c_attn",
33
+ "transformer.h.0.attn.c_proj",
34
+ "transformer.h.6.attn.c_attn",
35
+ "transformer.h.9.attn.c_proj",
36
+ "transformer.h.2.attn.c_proj",
37
+ "transformer.h.1.attn.c_proj",
38
+ "transformer.h.11.attn.c_attn",
39
+ "transformer.h.1.attn.c_attn",
40
+ "transformer.h.7.attn.c_proj",
41
+ "transformer.h.9.attn.c_attn",
42
+ "transformer.h.7.attn.c_attn",
43
+ "transformer.h.8.attn.c_proj",
44
+ "transformer.h.10.attn.c_proj",
45
+ "transformer.h.2.attn.c_attn",
46
+ "transformer.h.3.attn.c_proj",
47
+ "transformer.h.4.attn.c_proj",
48
+ "transformer.h.10.attn.c_attn"
49
  ],
50
+ "task_type": "CAUSAL_LM"
 
51
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e13ef76a1f210827fb843156b1b57d78d50037a4131c82d2a7ad9d458a589960
3
- size 4730632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7a79aafd8a3ded5c8022b7bd591a9b2733484f1509b02860beafed6decdcdff
3
+ size 150264
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a76dc550c0bd6965f6490ca703882eb8539c9f8cddc3d56fb829b9f1e12ef167
3
- size 4600
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de088ca0b011df1e98d8aba3ad48e1acb1c1af504ac7bc03cf919b6dfb166165
3
+ size 4664