Upload model
Browse files- model-00001.safetensors +2 -2
- model-00002.safetensors +2 -2
- model.safetensors.index.json +4 -12
model-00001.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21ac17a8bcba281201233fd7a2d05f3d511d03cd8432dab07d3c49745f8279ff
|
3 |
+
size 2127522803
|
model-00002.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3eb83f66e29bb19568b4d81fd5efb7d59aa9ee9f744e511abfc1d33037420c13
|
3 |
+
size 1336917858
|
model.safetensors.index.json
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
|
@@ -966,9 +966,6 @@
|
|
966 |
"model.layers.31.post_attention_layernorm.weight.dtype": "model-00001.safetensors",
|
967 |
"model.norm.weight.dtype": "model-00001.safetensors",
|
968 |
"lm_head.weight.dtype": "model-00001.safetensors",
|
969 |
-
"model.embed_tokens.weight.manual_8_bit_weight_min": "model-00001.safetensors",
|
970 |
-
"model.embed_tokens.weight.manual_8_bit_weight_max": "model-00001.safetensors",
|
971 |
-
"model.embed_tokens.weight.shape": "model-00001.safetensors",
|
972 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
973 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
974 |
"model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
@@ -1389,6 +1386,7 @@
|
|
1389 |
"model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
1390 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
1391 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
|
|
1392 |
"model.layers.0.input_layernorm.weight.shape": "model-00001.safetensors",
|
1393 |
"model.layers.0.post_attention_layernorm.weight.shape": "model-00001.safetensors",
|
1394 |
"model.layers.1.input_layernorm.weight.shape": "model-00001.safetensors",
|
@@ -1483,12 +1481,6 @@
|
|
1483 |
"model.layers.31.post_attention_layernorm.weight.shape": "model-00001.safetensors",
|
1484 |
"model.norm.weight.shape": "model-00001.safetensors",
|
1485 |
"lm_head.weight.shape": "model-00001.safetensors",
|
1486 |
-
"model.embed_tokens.weight.manual_8_bit_weight_packed.shape": "model-00001.safetensors",
|
1487 |
-
"model.embed_tokens.weight.manual_8_bit_weight_min.shape": "model-00001.safetensors",
|
1488 |
-
"model.embed_tokens.weight.manual_8_bit_weight_max.shape": "model-00001.safetensors",
|
1489 |
-
"model.embed_tokens.weight.manual_8_bit_weight_packed.dtype": "model-00001.safetensors",
|
1490 |
-
"model.embed_tokens.weight.manual_8_bit_weight_min.dtype": "model-00001.safetensors",
|
1491 |
-
"model.embed_tokens.weight.manual_8_bit_weight_max.dtype": "model-00001.safetensors",
|
1492 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
|
1493 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.compressed": "model-00001.safetensors",
|
1494 |
"model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
|
@@ -1909,6 +1901,7 @@
|
|
1909 |
"model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
|
1910 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00002.safetensors",
|
1911 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
|
|
|
1912 |
"model.layers.0.input_layernorm.weight.compressed": "model-00002.safetensors",
|
1913 |
"model.layers.0.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
|
1914 |
"model.layers.1.input_layernorm.weight.compressed": "model-00002.safetensors",
|
@@ -2002,7 +1995,6 @@
|
|
2002 |
"model.layers.31.input_layernorm.weight.compressed": "model-00002.safetensors",
|
2003 |
"model.layers.31.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
|
2004 |
"model.norm.weight.compressed": "model-00002.safetensors",
|
2005 |
-
"lm_head.weight.compressed": "model-00002.safetensors"
|
2006 |
-
"model.embed_tokens.weight.manual_8_bit_weight_packed.compressed": "model-00002.safetensors"
|
2007 |
}
|
2008 |
}
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 3464197717
|
4 |
},
|
5 |
"weight_map": {
|
6 |
"model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
|
|
|
966 |
"model.layers.31.post_attention_layernorm.weight.dtype": "model-00001.safetensors",
|
967 |
"model.norm.weight.dtype": "model-00001.safetensors",
|
968 |
"lm_head.weight.dtype": "model-00001.safetensors",
|
|
|
|
|
|
|
969 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
970 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
971 |
"model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
|
|
1386 |
"model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
1387 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
|
1388 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
|
1389 |
+
"model.embed_tokens.weight.shape": "model-00001.safetensors",
|
1390 |
"model.layers.0.input_layernorm.weight.shape": "model-00001.safetensors",
|
1391 |
"model.layers.0.post_attention_layernorm.weight.shape": "model-00001.safetensors",
|
1392 |
"model.layers.1.input_layernorm.weight.shape": "model-00001.safetensors",
|
|
|
1481 |
"model.layers.31.post_attention_layernorm.weight.shape": "model-00001.safetensors",
|
1482 |
"model.norm.weight.shape": "model-00001.safetensors",
|
1483 |
"lm_head.weight.shape": "model-00001.safetensors",
|
|
|
|
|
|
|
|
|
|
|
|
|
1484 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
|
1485 |
"model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.compressed": "model-00001.safetensors",
|
1486 |
"model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
|
|
|
1901 |
"model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
|
1902 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00002.safetensors",
|
1903 |
"model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
|
1904 |
+
"model.embed_tokens.weight.compressed": "model-00002.safetensors",
|
1905 |
"model.layers.0.input_layernorm.weight.compressed": "model-00002.safetensors",
|
1906 |
"model.layers.0.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
|
1907 |
"model.layers.1.input_layernorm.weight.compressed": "model-00002.safetensors",
|
|
|
1995 |
"model.layers.31.input_layernorm.weight.compressed": "model-00002.safetensors",
|
1996 |
"model.layers.31.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
|
1997 |
"model.norm.weight.compressed": "model-00002.safetensors",
|
1998 |
+
"lm_head.weight.compressed": "model-00002.safetensors"
|
|
|
1999 |
}
|
2000 |
}
|