init
Browse files- .gitattributes +1 -0
- config.json +51 -0
- model.safetensors +3 -0
- quant_log.csv +169 -0
- quantize_config.json +21 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
model.safetensors filter=lfs diff=lfs merge=lfs -text
|
config.json
ADDED
@@ -0,0 +1,51 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_attn_implementation_autoset": true,
|
3 |
+
"_name_or_path": "/root/.cache/huggingface/hub/models--Qwen--Qwen2.5-0.5B-Instruct/snapshots/7ae557604adf67be50417f59c2c2f167def9a775",
|
4 |
+
"architectures": [
|
5 |
+
"Qwen2ForCausalLM"
|
6 |
+
],
|
7 |
+
"attention_dropout": 0.0,
|
8 |
+
"bos_token_id": 151643,
|
9 |
+
"eos_token_id": 151645,
|
10 |
+
"hidden_act": "silu",
|
11 |
+
"hidden_size": 896,
|
12 |
+
"initializer_range": 0.02,
|
13 |
+
"intermediate_size": 4864,
|
14 |
+
"max_position_embeddings": 32768,
|
15 |
+
"max_window_layers": 21,
|
16 |
+
"model_type": "qwen2",
|
17 |
+
"num_attention_heads": 14,
|
18 |
+
"num_hidden_layers": 24,
|
19 |
+
"num_key_value_heads": 2,
|
20 |
+
"quantization_config": {
|
21 |
+
"bits": 2,
|
22 |
+
"checkpoint_format": "gptq",
|
23 |
+
"desc_act": false,
|
24 |
+
"dynamic": null,
|
25 |
+
"group_size": 128,
|
26 |
+
"lm_head": false,
|
27 |
+
"meta": {
|
28 |
+
"damp_auto_increment": 0.0025,
|
29 |
+
"damp_percent": 0.01,
|
30 |
+
"mse": 0.0,
|
31 |
+
"quantizer": [
|
32 |
+
"gptqmodel:1.5.1"
|
33 |
+
],
|
34 |
+
"static_groups": false,
|
35 |
+
"true_sequential": true,
|
36 |
+
"uri": "https://github.com/modelcloud/gptqmodel"
|
37 |
+
},
|
38 |
+
"quant_method": "gptq",
|
39 |
+
"sym": true
|
40 |
+
},
|
41 |
+
"rms_norm_eps": 1e-06,
|
42 |
+
"rope_scaling": null,
|
43 |
+
"rope_theta": 1000000.0,
|
44 |
+
"sliding_window": null,
|
45 |
+
"tie_word_embeddings": true,
|
46 |
+
"torch_dtype": "bfloat16",
|
47 |
+
"transformers_version": "4.47.1",
|
48 |
+
"use_cache": true,
|
49 |
+
"use_sliding_window": false,
|
50 |
+
"vocab_size": 151936
|
51 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8572721112376195fbcc8595a6a204b3bfad41f90101ba5d4e36d205c718e79f
|
3 |
+
size 641496752
|
quant_log.csv
ADDED
@@ -0,0 +1,169 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
layer,module,loss,damp,time
|
2 |
+
0,self_attn.k_proj,0.00138,0.01000,0.372
|
3 |
+
0,self_attn.v_proj,0.00003,0.01000,0.143
|
4 |
+
0,self_attn.q_proj,0.01032,0.01000,0.146
|
5 |
+
0,self_attn.o_proj,0.00001,0.01000,0.141
|
6 |
+
0,mlp.up_proj,0.09639,0.01000,0.147
|
7 |
+
0,mlp.gate_proj,0.19389,0.01000,0.144
|
8 |
+
0,mlp.down_proj,0.00145,0.01000,0.760
|
9 |
+
1,self_attn.k_proj,0.00983,0.01000,0.141
|
10 |
+
1,self_attn.v_proj,0.00101,0.01000,0.140
|
11 |
+
1,self_attn.q_proj,0.03416,0.01000,0.145
|
12 |
+
1,self_attn.o_proj,0.00206,0.01000,0.142
|
13 |
+
1,mlp.up_proj,0.10123,0.01000,0.152
|
14 |
+
1,mlp.gate_proj,0.19488,0.01000,0.147
|
15 |
+
1,mlp.down_proj,0.00114,0.01000,0.763
|
16 |
+
2,self_attn.k_proj,0.01729,0.01000,0.143
|
17 |
+
2,self_attn.v_proj,0.00211,0.01000,0.141
|
18 |
+
2,self_attn.q_proj,0.05901,0.01000,0.143
|
19 |
+
2,self_attn.o_proj,0.00084,0.01000,0.144
|
20 |
+
2,mlp.up_proj,0.16170,0.01000,0.144
|
21 |
+
2,mlp.gate_proj,0.30168,0.01000,0.144
|
22 |
+
2,mlp.down_proj,3.40212,0.01000,0.777
|
23 |
+
3,self_attn.k_proj,0.02164,0.01000,0.146
|
24 |
+
3,self_attn.v_proj,0.00550,0.01000,0.144
|
25 |
+
3,self_attn.q_proj,0.08517,0.01000,0.140
|
26 |
+
3,self_attn.o_proj,0.00069,0.01000,0.145
|
27 |
+
3,mlp.up_proj,1.55705,0.01000,0.149
|
28 |
+
3,mlp.gate_proj,14.22241,0.01000,0.153
|
29 |
+
3,mlp.down_proj,10.47623,0.01000,0.771
|
30 |
+
4,self_attn.k_proj,0.01608,0.01000,0.141
|
31 |
+
4,self_attn.v_proj,0.00733,0.01000,0.140
|
32 |
+
4,self_attn.q_proj,0.07782,0.01000,0.141
|
33 |
+
4,self_attn.o_proj,0.00357,0.01000,0.142
|
34 |
+
4,mlp.up_proj,0.26308,0.01000,0.146
|
35 |
+
4,mlp.gate_proj,0.44777,0.01000,0.146
|
36 |
+
4,mlp.down_proj,0.00393,0.01000,0.773
|
37 |
+
5,self_attn.k_proj,0.01548,0.01000,0.141
|
38 |
+
5,self_attn.v_proj,0.00733,0.01000,0.143
|
39 |
+
5,self_attn.q_proj,0.07514,0.01000,0.142
|
40 |
+
5,self_attn.o_proj,0.00126,0.01000,0.142
|
41 |
+
5,mlp.up_proj,0.32005,0.01000,0.147
|
42 |
+
5,mlp.gate_proj,0.75867,0.01000,0.143
|
43 |
+
5,mlp.down_proj,0.00588,0.01000,0.782
|
44 |
+
6,self_attn.k_proj,0.02100,0.01000,0.142
|
45 |
+
6,self_attn.v_proj,0.00581,0.01000,0.142
|
46 |
+
6,self_attn.q_proj,0.08125,0.01000,0.143
|
47 |
+
6,self_attn.o_proj,0.00086,0.01000,0.142
|
48 |
+
6,mlp.up_proj,0.25930,0.01000,0.146
|
49 |
+
6,mlp.gate_proj,0.43217,0.01000,0.147
|
50 |
+
6,mlp.down_proj,0.00290,0.01000,0.776
|
51 |
+
7,self_attn.k_proj,0.02048,0.01000,0.140
|
52 |
+
7,self_attn.v_proj,0.00687,0.01000,0.139
|
53 |
+
7,self_attn.q_proj,0.08387,0.01000,0.140
|
54 |
+
7,self_attn.o_proj,0.00102,0.01000,0.143
|
55 |
+
7,mlp.up_proj,0.27253,0.01000,0.144
|
56 |
+
7,mlp.gate_proj,0.38177,0.01000,0.150
|
57 |
+
7,mlp.down_proj,0.00328,0.01000,0.786
|
58 |
+
8,self_attn.k_proj,0.01920,0.01000,0.144
|
59 |
+
8,self_attn.v_proj,0.00666,0.01000,0.139
|
60 |
+
8,self_attn.q_proj,0.08629,0.01000,0.142
|
61 |
+
8,self_attn.o_proj,0.00223,0.01000,0.145
|
62 |
+
8,mlp.up_proj,0.24209,0.01000,0.145
|
63 |
+
8,mlp.gate_proj,0.34371,0.01000,0.147
|
64 |
+
8,mlp.down_proj,0.00234,0.01000,0.770
|
65 |
+
9,self_attn.k_proj,0.03560,0.01000,0.146
|
66 |
+
9,self_attn.v_proj,0.01327,0.01000,0.146
|
67 |
+
9,self_attn.q_proj,0.15881,0.01000,0.142
|
68 |
+
9,self_attn.o_proj,0.00107,0.01000,0.141
|
69 |
+
9,mlp.up_proj,0.24913,0.01000,0.145
|
70 |
+
9,mlp.gate_proj,0.31356,0.01000,0.153
|
71 |
+
9,mlp.down_proj,0.00273,0.01000,0.770
|
72 |
+
10,self_attn.k_proj,0.01898,0.01000,0.140
|
73 |
+
10,self_attn.v_proj,0.01226,0.01000,0.142
|
74 |
+
10,self_attn.q_proj,0.09338,0.01000,0.146
|
75 |
+
10,self_attn.o_proj,0.00256,0.01000,0.139
|
76 |
+
10,mlp.up_proj,0.19837,0.01000,0.146
|
77 |
+
10,mlp.gate_proj,0.25931,0.01000,0.152
|
78 |
+
10,mlp.down_proj,0.00200,0.01000,0.781
|
79 |
+
11,self_attn.k_proj,0.03752,0.01000,0.142
|
80 |
+
11,self_attn.v_proj,0.01387,0.01000,0.139
|
81 |
+
11,self_attn.q_proj,0.17363,0.01000,0.144
|
82 |
+
11,self_attn.o_proj,0.00194,0.01000,0.142
|
83 |
+
11,mlp.up_proj,0.24216,0.01000,0.147
|
84 |
+
11,mlp.gate_proj,0.26251,0.01000,0.147
|
85 |
+
11,mlp.down_proj,0.00267,0.01000,0.776
|
86 |
+
12,self_attn.k_proj,0.01698,0.01000,0.142
|
87 |
+
12,self_attn.v_proj,0.00979,0.01000,0.142
|
88 |
+
12,self_attn.q_proj,0.08870,0.01000,0.143
|
89 |
+
12,self_attn.o_proj,0.00231,0.01000,0.141
|
90 |
+
12,mlp.up_proj,0.20748,0.01000,0.146
|
91 |
+
12,mlp.gate_proj,0.22905,0.01000,0.148
|
92 |
+
12,mlp.down_proj,0.00239,0.01000,0.772
|
93 |
+
13,self_attn.k_proj,0.02764,0.01000,0.143
|
94 |
+
13,self_attn.v_proj,0.01191,0.01000,0.145
|
95 |
+
13,self_attn.q_proj,0.14133,0.01000,0.146
|
96 |
+
13,self_attn.o_proj,0.00121,0.01000,0.144
|
97 |
+
13,mlp.up_proj,0.24341,0.01000,0.149
|
98 |
+
13,mlp.gate_proj,0.24755,0.01000,0.147
|
99 |
+
13,mlp.down_proj,0.00313,0.01000,0.776
|
100 |
+
14,self_attn.k_proj,0.02157,0.01000,0.144
|
101 |
+
14,self_attn.v_proj,0.01694,0.01000,0.143
|
102 |
+
14,self_attn.q_proj,0.12651,0.01000,0.145
|
103 |
+
14,self_attn.o_proj,0.00391,0.01000,0.143
|
104 |
+
14,mlp.up_proj,0.23182,0.01000,0.145
|
105 |
+
14,mlp.gate_proj,0.26455,0.01000,0.157
|
106 |
+
14,mlp.down_proj,0.00381,0.01000,0.783
|
107 |
+
15,self_attn.k_proj,0.01991,0.01000,0.145
|
108 |
+
15,self_attn.v_proj,0.01372,0.01000,0.144
|
109 |
+
15,self_attn.q_proj,0.10473,0.01000,0.146
|
110 |
+
15,self_attn.o_proj,0.00366,0.01000,0.145
|
111 |
+
15,mlp.up_proj,0.29403,0.01000,0.149
|
112 |
+
15,mlp.gate_proj,0.30643,0.01000,0.145
|
113 |
+
15,mlp.down_proj,0.00572,0.01000,0.765
|
114 |
+
16,self_attn.k_proj,0.03065,0.01000,0.144
|
115 |
+
16,self_attn.v_proj,0.02840,0.01000,0.140
|
116 |
+
16,self_attn.q_proj,0.21804,0.01000,0.146
|
117 |
+
16,self_attn.o_proj,0.00136,0.01000,0.143
|
118 |
+
16,mlp.up_proj,0.34904,0.01000,0.146
|
119 |
+
16,mlp.gate_proj,0.45293,0.01000,0.147
|
120 |
+
16,mlp.down_proj,0.01072,0.01000,0.766
|
121 |
+
17,self_attn.k_proj,0.02077,0.01000,0.145
|
122 |
+
17,self_attn.v_proj,0.02172,0.01000,0.145
|
123 |
+
17,self_attn.q_proj,0.13563,0.01000,0.144
|
124 |
+
17,self_attn.o_proj,0.00379,0.01000,0.142
|
125 |
+
17,mlp.up_proj,0.40654,0.01000,0.150
|
126 |
+
17,mlp.gate_proj,0.63069,0.01000,0.153
|
127 |
+
17,mlp.down_proj,0.00786,0.01000,0.771
|
128 |
+
18,self_attn.k_proj,0.02198,0.01000,0.149
|
129 |
+
18,self_attn.v_proj,0.01774,0.01000,0.141
|
130 |
+
18,self_attn.q_proj,0.10674,0.01000,0.145
|
131 |
+
18,self_attn.o_proj,0.00129,0.01000,0.143
|
132 |
+
18,mlp.up_proj,0.38269,0.01000,0.148
|
133 |
+
18,mlp.gate_proj,0.47625,0.01000,0.147
|
134 |
+
18,mlp.down_proj,0.00765,0.01000,0.777
|
135 |
+
19,self_attn.k_proj,0.01776,0.01000,0.144
|
136 |
+
19,self_attn.v_proj,0.01706,0.01000,0.142
|
137 |
+
19,self_attn.q_proj,0.11138,0.01000,0.143
|
138 |
+
19,self_attn.o_proj,0.00345,0.01000,0.143
|
139 |
+
19,mlp.up_proj,0.57152,0.01000,0.146
|
140 |
+
19,mlp.gate_proj,0.68538,0.01000,0.148
|
141 |
+
19,mlp.down_proj,0.01574,0.01000,0.783
|
142 |
+
20,self_attn.k_proj,0.02558,0.01000,0.145
|
143 |
+
20,self_attn.v_proj,0.05435,0.01000,0.146
|
144 |
+
20,self_attn.q_proj,0.16357,0.01000,0.140
|
145 |
+
20,self_attn.o_proj,0.01051,0.01000,0.144
|
146 |
+
20,mlp.up_proj,0.65037,0.01000,0.145
|
147 |
+
20,mlp.gate_proj,0.71838,0.01000,0.147
|
148 |
+
20,mlp.down_proj,0.02858,0.01000,0.766
|
149 |
+
21,self_attn.k_proj,0.03181,0.01000,0.144
|
150 |
+
21,self_attn.v_proj,0.10603,0.01000,0.144
|
151 |
+
21,self_attn.q_proj,0.23383,0.01000,0.144
|
152 |
+
21,self_attn.o_proj,0.02307,0.01000,0.144
|
153 |
+
21,mlp.up_proj,1.75373,0.01000,0.158
|
154 |
+
21,mlp.gate_proj,3.94375,0.01000,0.146
|
155 |
+
21,mlp.down_proj,17.75035,0.01000,0.779
|
156 |
+
22,self_attn.k_proj,0.02056,0.01000,0.141
|
157 |
+
22,self_attn.v_proj,0.06087,0.01000,0.140
|
158 |
+
22,self_attn.q_proj,0.15773,0.01000,0.148
|
159 |
+
22,self_attn.o_proj,0.01726,0.01000,0.140
|
160 |
+
22,mlp.up_proj,0.53556,0.01000,0.144
|
161 |
+
22,mlp.gate_proj,0.51474,0.01000,0.144
|
162 |
+
22,mlp.down_proj,0.04907,0.01000,0.769
|
163 |
+
23,self_attn.k_proj,0.02615,0.01000,0.176
|
164 |
+
23,self_attn.v_proj,0.05774,0.01000,0.209
|
165 |
+
23,self_attn.q_proj,0.20516,0.01000,0.209
|
166 |
+
23,self_attn.o_proj,0.03206,0.01000,0.210
|
167 |
+
23,mlp.up_proj,0.83435,0.01000,0.214
|
168 |
+
23,mlp.gate_proj,0.87210,0.01000,0.224
|
169 |
+
23,mlp.down_proj,0.48671,0.01000,0.784
|
quantize_config.json
ADDED
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bits": 2,
|
3 |
+
"dynamic": null,
|
4 |
+
"group_size": 128,
|
5 |
+
"desc_act": false,
|
6 |
+
"sym": true,
|
7 |
+
"lm_head": false,
|
8 |
+
"quant_method": "gptq",
|
9 |
+
"checkpoint_format": "gptq",
|
10 |
+
"meta": {
|
11 |
+
"quantizer": [
|
12 |
+
"gptqmodel:1.5.1"
|
13 |
+
],
|
14 |
+
"uri": "https://github.com/modelcloud/gptqmodel",
|
15 |
+
"damp_percent": 0.01,
|
16 |
+
"damp_auto_increment": 0.0025,
|
17 |
+
"static_groups": false,
|
18 |
+
"true_sequential": true,
|
19 |
+
"mse": 0.0
|
20 |
+
}
|
21 |
+
}
|