3v324v23 commited on
Commit
8872684
·
1 Parent(s): 7c3265a
Files changed (5) hide show
  1. .gitattributes +1 -0
  2. config.json +51 -0
  3. model.safetensors +3 -0
  4. quant_log.csv +169 -0
  5. quantize_config.json +21 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ model.safetensors filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "/root/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/7ae557604adf67be50417f59c2c2f167def9a775",
4
+ "architectures": [
5
+ "Qwen2ForCausalLM"
6
+ ],
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 151643,
9
+ "eos_token_id": 151645,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 896,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4864,
14
+ "max_position_embeddings": 32768,
15
+ "max_window_layers": 21,
16
+ "model_type": "qwen2",
17
+ "num_attention_heads": 14,
18
+ "num_hidden_layers": 24,
19
+ "num_key_value_heads": 2,
20
+ "quantization_config": {
21
+ "bits": 2,
22
+ "checkpoint_format": "gptq",
23
+ "desc_act": false,
24
+ "dynamic": null,
25
+ "group_size": 128,
26
+ "lm_head": false,
27
+ "meta": {
28
+ "damp_auto_increment": 0.0025,
29
+ "damp_percent": 0.01,
30
+ "mse": 0.0,
31
+ "quantizer": [
32
+ "gptqmodel:1.5.1"
33
+ ],
34
+ "static_groups": false,
35
+ "true_sequential": true,
36
+ "uri": "https://github.com/modelcloud/gptqmodel"
37
+ },
38
+ "quant_method": "gptq",
39
+ "sym": true
40
+ },
41
+ "rms_norm_eps": 1e-06,
42
+ "rope_scaling": null,
43
+ "rope_theta": 1000000.0,
44
+ "sliding_window": null,
45
+ "tie_word_embeddings": true,
46
+ "torch_dtype": "bfloat16",
47
+ "transformers_version": "4.47.1",
48
+ "use_cache": true,
49
+ "use_sliding_window": false,
50
+ "vocab_size": 151936
51
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8572721112376195fbcc8595a6a204b3bfad41f90101ba5d4e36d205c718e79f
3
+ size 641496752
quant_log.csv ADDED
@@ -0,0 +1,169 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ layer,module,loss,damp,time
2
+ 0,self_attn.k_proj,0.00138,0.01000,0.372
3
+ 0,self_attn.v_proj,0.00003,0.01000,0.143
4
+ 0,self_attn.q_proj,0.01032,0.01000,0.146
5
+ 0,self_attn.o_proj,0.00001,0.01000,0.141
6
+ 0,mlp.up_proj,0.09639,0.01000,0.147
7
+ 0,mlp.gate_proj,0.19389,0.01000,0.144
8
+ 0,mlp.down_proj,0.00145,0.01000,0.760
9
+ 1,self_attn.k_proj,0.00983,0.01000,0.141
10
+ 1,self_attn.v_proj,0.00101,0.01000,0.140
11
+ 1,self_attn.q_proj,0.03416,0.01000,0.145
12
+ 1,self_attn.o_proj,0.00206,0.01000,0.142
13
+ 1,mlp.up_proj,0.10123,0.01000,0.152
14
+ 1,mlp.gate_proj,0.19488,0.01000,0.147
15
+ 1,mlp.down_proj,0.00114,0.01000,0.763
16
+ 2,self_attn.k_proj,0.01729,0.01000,0.143
17
+ 2,self_attn.v_proj,0.00211,0.01000,0.141
18
+ 2,self_attn.q_proj,0.05901,0.01000,0.143
19
+ 2,self_attn.o_proj,0.00084,0.01000,0.144
20
+ 2,mlp.up_proj,0.16170,0.01000,0.144
21
+ 2,mlp.gate_proj,0.30168,0.01000,0.144
22
+ 2,mlp.down_proj,3.40212,0.01000,0.777
23
+ 3,self_attn.k_proj,0.02164,0.01000,0.146
24
+ 3,self_attn.v_proj,0.00550,0.01000,0.144
25
+ 3,self_attn.q_proj,0.08517,0.01000,0.140
26
+ 3,self_attn.o_proj,0.00069,0.01000,0.145
27
+ 3,mlp.up_proj,1.55705,0.01000,0.149
28
+ 3,mlp.gate_proj,14.22241,0.01000,0.153
29
+ 3,mlp.down_proj,10.47623,0.01000,0.771
30
+ 4,self_attn.k_proj,0.01608,0.01000,0.141
31
+ 4,self_attn.v_proj,0.00733,0.01000,0.140
32
+ 4,self_attn.q_proj,0.07782,0.01000,0.141
33
+ 4,self_attn.o_proj,0.00357,0.01000,0.142
34
+ 4,mlp.up_proj,0.26308,0.01000,0.146
35
+ 4,mlp.gate_proj,0.44777,0.01000,0.146
36
+ 4,mlp.down_proj,0.00393,0.01000,0.773
37
+ 5,self_attn.k_proj,0.01548,0.01000,0.141
38
+ 5,self_attn.v_proj,0.00733,0.01000,0.143
39
+ 5,self_attn.q_proj,0.07514,0.01000,0.142
40
+ 5,self_attn.o_proj,0.00126,0.01000,0.142
41
+ 5,mlp.up_proj,0.32005,0.01000,0.147
42
+ 5,mlp.gate_proj,0.75867,0.01000,0.143
43
+ 5,mlp.down_proj,0.00588,0.01000,0.782
44
+ 6,self_attn.k_proj,0.02100,0.01000,0.142
45
+ 6,self_attn.v_proj,0.00581,0.01000,0.142
46
+ 6,self_attn.q_proj,0.08125,0.01000,0.143
47
+ 6,self_attn.o_proj,0.00086,0.01000,0.142
48
+ 6,mlp.up_proj,0.25930,0.01000,0.146
49
+ 6,mlp.gate_proj,0.43217,0.01000,0.147
50
+ 6,mlp.down_proj,0.00290,0.01000,0.776
51
+ 7,self_attn.k_proj,0.02048,0.01000,0.140
52
+ 7,self_attn.v_proj,0.00687,0.01000,0.139
53
+ 7,self_attn.q_proj,0.08387,0.01000,0.140
54
+ 7,self_attn.o_proj,0.00102,0.01000,0.143
55
+ 7,mlp.up_proj,0.27253,0.01000,0.144
56
+ 7,mlp.gate_proj,0.38177,0.01000,0.150
57
+ 7,mlp.down_proj,0.00328,0.01000,0.786
58
+ 8,self_attn.k_proj,0.01920,0.01000,0.144
59
+ 8,self_attn.v_proj,0.00666,0.01000,0.139
60
+ 8,self_attn.q_proj,0.08629,0.01000,0.142
61
+ 8,self_attn.o_proj,0.00223,0.01000,0.145
62
+ 8,mlp.up_proj,0.24209,0.01000,0.145
63
+ 8,mlp.gate_proj,0.34371,0.01000,0.147
64
+ 8,mlp.down_proj,0.00234,0.01000,0.770
65
+ 9,self_attn.k_proj,0.03560,0.01000,0.146
66
+ 9,self_attn.v_proj,0.01327,0.01000,0.146
67
+ 9,self_attn.q_proj,0.15881,0.01000,0.142
68
+ 9,self_attn.o_proj,0.00107,0.01000,0.141
69
+ 9,mlp.up_proj,0.24913,0.01000,0.145
70
+ 9,mlp.gate_proj,0.31356,0.01000,0.153
71
+ 9,mlp.down_proj,0.00273,0.01000,0.770
72
+ 10,self_attn.k_proj,0.01898,0.01000,0.140
73
+ 10,self_attn.v_proj,0.01226,0.01000,0.142
74
+ 10,self_attn.q_proj,0.09338,0.01000,0.146
75
+ 10,self_attn.o_proj,0.00256,0.01000,0.139
76
+ 10,mlp.up_proj,0.19837,0.01000,0.146
77
+ 10,mlp.gate_proj,0.25931,0.01000,0.152
78
+ 10,mlp.down_proj,0.00200,0.01000,0.781
79
+ 11,self_attn.k_proj,0.03752,0.01000,0.142
80
+ 11,self_attn.v_proj,0.01387,0.01000,0.139
81
+ 11,self_attn.q_proj,0.17363,0.01000,0.144
82
+ 11,self_attn.o_proj,0.00194,0.01000,0.142
83
+ 11,mlp.up_proj,0.24216,0.01000,0.147
84
+ 11,mlp.gate_proj,0.26251,0.01000,0.147
85
+ 11,mlp.down_proj,0.00267,0.01000,0.776
86
+ 12,self_attn.k_proj,0.01698,0.01000,0.142
87
+ 12,self_attn.v_proj,0.00979,0.01000,0.142
88
+ 12,self_attn.q_proj,0.08870,0.01000,0.143
89
+ 12,self_attn.o_proj,0.00231,0.01000,0.141
90
+ 12,mlp.up_proj,0.20748,0.01000,0.146
91
+ 12,mlp.gate_proj,0.22905,0.01000,0.148
92
+ 12,mlp.down_proj,0.00239,0.01000,0.772
93
+ 13,self_attn.k_proj,0.02764,0.01000,0.143
94
+ 13,self_attn.v_proj,0.01191,0.01000,0.145
95
+ 13,self_attn.q_proj,0.14133,0.01000,0.146
96
+ 13,self_attn.o_proj,0.00121,0.01000,0.144
97
+ 13,mlp.up_proj,0.24341,0.01000,0.149
98
+ 13,mlp.gate_proj,0.24755,0.01000,0.147
99
+ 13,mlp.down_proj,0.00313,0.01000,0.776
100
+ 14,self_attn.k_proj,0.02157,0.01000,0.144
101
+ 14,self_attn.v_proj,0.01694,0.01000,0.143
102
+ 14,self_attn.q_proj,0.12651,0.01000,0.145
103
+ 14,self_attn.o_proj,0.00391,0.01000,0.143
104
+ 14,mlp.up_proj,0.23182,0.01000,0.145
105
+ 14,mlp.gate_proj,0.26455,0.01000,0.157
106
+ 14,mlp.down_proj,0.00381,0.01000,0.783
107
+ 15,self_attn.k_proj,0.01991,0.01000,0.145
108
+ 15,self_attn.v_proj,0.01372,0.01000,0.144
109
+ 15,self_attn.q_proj,0.10473,0.01000,0.146
110
+ 15,self_attn.o_proj,0.00366,0.01000,0.145
111
+ 15,mlp.up_proj,0.29403,0.01000,0.149
112
+ 15,mlp.gate_proj,0.30643,0.01000,0.145
113
+ 15,mlp.down_proj,0.00572,0.01000,0.765
114
+ 16,self_attn.k_proj,0.03065,0.01000,0.144
115
+ 16,self_attn.v_proj,0.02840,0.01000,0.140
116
+ 16,self_attn.q_proj,0.21804,0.01000,0.146
117
+ 16,self_attn.o_proj,0.00136,0.01000,0.143
118
+ 16,mlp.up_proj,0.34904,0.01000,0.146
119
+ 16,mlp.gate_proj,0.45293,0.01000,0.147
120
+ 16,mlp.down_proj,0.01072,0.01000,0.766
121
+ 17,self_attn.k_proj,0.02077,0.01000,0.145
122
+ 17,self_attn.v_proj,0.02172,0.01000,0.145
123
+ 17,self_attn.q_proj,0.13563,0.01000,0.144
124
+ 17,self_attn.o_proj,0.00379,0.01000,0.142
125
+ 17,mlp.up_proj,0.40654,0.01000,0.150
126
+ 17,mlp.gate_proj,0.63069,0.01000,0.153
127
+ 17,mlp.down_proj,0.00786,0.01000,0.771
128
+ 18,self_attn.k_proj,0.02198,0.01000,0.149
129
+ 18,self_attn.v_proj,0.01774,0.01000,0.141
130
+ 18,self_attn.q_proj,0.10674,0.01000,0.145
131
+ 18,self_attn.o_proj,0.00129,0.01000,0.143
132
+ 18,mlp.up_proj,0.38269,0.01000,0.148
133
+ 18,mlp.gate_proj,0.47625,0.01000,0.147
134
+ 18,mlp.down_proj,0.00765,0.01000,0.777
135
+ 19,self_attn.k_proj,0.01776,0.01000,0.144
136
+ 19,self_attn.v_proj,0.01706,0.01000,0.142
137
+ 19,self_attn.q_proj,0.11138,0.01000,0.143
138
+ 19,self_attn.o_proj,0.00345,0.01000,0.143
139
+ 19,mlp.up_proj,0.57152,0.01000,0.146
140
+ 19,mlp.gate_proj,0.68538,0.01000,0.148
141
+ 19,mlp.down_proj,0.01574,0.01000,0.783
142
+ 20,self_attn.k_proj,0.02558,0.01000,0.145
143
+ 20,self_attn.v_proj,0.05435,0.01000,0.146
144
+ 20,self_attn.q_proj,0.16357,0.01000,0.140
145
+ 20,self_attn.o_proj,0.01051,0.01000,0.144
146
+ 20,mlp.up_proj,0.65037,0.01000,0.145
147
+ 20,mlp.gate_proj,0.71838,0.01000,0.147
148
+ 20,mlp.down_proj,0.02858,0.01000,0.766
149
+ 21,self_attn.k_proj,0.03181,0.01000,0.144
150
+ 21,self_attn.v_proj,0.10603,0.01000,0.144
151
+ 21,self_attn.q_proj,0.23383,0.01000,0.144
152
+ 21,self_attn.o_proj,0.02307,0.01000,0.144
153
+ 21,mlp.up_proj,1.75373,0.01000,0.158
154
+ 21,mlp.gate_proj,3.94375,0.01000,0.146
155
+ 21,mlp.down_proj,17.75035,0.01000,0.779
156
+ 22,self_attn.k_proj,0.02056,0.01000,0.141
157
+ 22,self_attn.v_proj,0.06087,0.01000,0.140
158
+ 22,self_attn.q_proj,0.15773,0.01000,0.148
159
+ 22,self_attn.o_proj,0.01726,0.01000,0.140
160
+ 22,mlp.up_proj,0.53556,0.01000,0.144
161
+ 22,mlp.gate_proj,0.51474,0.01000,0.144
162
+ 22,mlp.down_proj,0.04907,0.01000,0.769
163
+ 23,self_attn.k_proj,0.02615,0.01000,0.176
164
+ 23,self_attn.v_proj,0.05774,0.01000,0.209
165
+ 23,self_attn.q_proj,0.20516,0.01000,0.209
166
+ 23,self_attn.o_proj,0.03206,0.01000,0.210
167
+ 23,mlp.up_proj,0.83435,0.01000,0.214
168
+ 23,mlp.gate_proj,0.87210,0.01000,0.224
169
+ 23,mlp.down_proj,0.48671,0.01000,0.784
quantize_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 2,
3
+ "dynamic": null,
4
+ "group_size": 128,
5
+ "desc_act": false,
6
+ "sym": true,
7
+ "lm_head": false,
8
+ "quant_method": "gptq",
9
+ "checkpoint_format": "gptq",
10
+ "meta": {
11
+ "quantizer": [
12
+ "gptqmodel:1.5.1"
13
+ ],
14
+ "uri": "https://github.com/modelcloud/gptqmodel",
15
+ "damp_percent": 0.01,
16
+ "damp_auto_increment": 0.0025,
17
+ "static_groups": false,
18
+ "true_sequential": true,
19
+ "mse": 0.0
20
+ }
21
+ }