rd211 commited on
Commit
ff737fe
·
verified ·
1 Parent(s): 693f06d

Upload model

Browse files
model-00001.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8af1f6afbe73d73e26398e7992825d24a3f49204a3bcc65a28a1030263e43e47
3
- size 2127523756
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21ac17a8bcba281201233fd7a2d05f3d511d03cd8432dab07d3c49745f8279ff
3
+ size 2127522803
model-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5a3fe4203e0fc20dd9059b7a02a8337e72e79683fd8477a183a5abf6505631e
3
- size 1274249800
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3eb83f66e29bb19568b4d81fd5efb7d59aa9ee9f744e511abfc1d33037420c13
3
+ size 1336917858
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 3401529668
4
  },
5
  "weight_map": {
6
  "model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
@@ -966,9 +966,6 @@
966
  "model.layers.31.post_attention_layernorm.weight.dtype": "model-00001.safetensors",
967
  "model.norm.weight.dtype": "model-00001.safetensors",
968
  "lm_head.weight.dtype": "model-00001.safetensors",
969
- "model.embed_tokens.weight.manual_8_bit_weight_min": "model-00001.safetensors",
970
- "model.embed_tokens.weight.manual_8_bit_weight_max": "model-00001.safetensors",
971
- "model.embed_tokens.weight.shape": "model-00001.safetensors",
972
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
973
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
974
  "model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
@@ -1389,6 +1386,7 @@
1389
  "model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
1390
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
1391
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
 
1392
  "model.layers.0.input_layernorm.weight.shape": "model-00001.safetensors",
1393
  "model.layers.0.post_attention_layernorm.weight.shape": "model-00001.safetensors",
1394
  "model.layers.1.input_layernorm.weight.shape": "model-00001.safetensors",
@@ -1483,12 +1481,6 @@
1483
  "model.layers.31.post_attention_layernorm.weight.shape": "model-00001.safetensors",
1484
  "model.norm.weight.shape": "model-00001.safetensors",
1485
  "lm_head.weight.shape": "model-00001.safetensors",
1486
- "model.embed_tokens.weight.manual_8_bit_weight_packed.shape": "model-00001.safetensors",
1487
- "model.embed_tokens.weight.manual_8_bit_weight_min.shape": "model-00001.safetensors",
1488
- "model.embed_tokens.weight.manual_8_bit_weight_max.shape": "model-00001.safetensors",
1489
- "model.embed_tokens.weight.manual_8_bit_weight_packed.dtype": "model-00001.safetensors",
1490
- "model.embed_tokens.weight.manual_8_bit_weight_min.dtype": "model-00001.safetensors",
1491
- "model.embed_tokens.weight.manual_8_bit_weight_max.dtype": "model-00001.safetensors",
1492
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
1493
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.compressed": "model-00001.safetensors",
1494
  "model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
@@ -1909,6 +1901,7 @@
1909
  "model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
1910
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00002.safetensors",
1911
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
 
1912
  "model.layers.0.input_layernorm.weight.compressed": "model-00002.safetensors",
1913
  "model.layers.0.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
1914
  "model.layers.1.input_layernorm.weight.compressed": "model-00002.safetensors",
@@ -2002,7 +1995,6 @@
2002
  "model.layers.31.input_layernorm.weight.compressed": "model-00002.safetensors",
2003
  "model.layers.31.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
2004
  "model.norm.weight.compressed": "model-00002.safetensors",
2005
- "lm_head.weight.compressed": "model-00002.safetensors",
2006
- "model.embed_tokens.weight.manual_8_bit_weight_packed.compressed": "model-00002.safetensors"
2007
  }
2008
  }
 
1
  {
2
  "metadata": {
3
+ "total_size": 3464197717
4
  },
5
  "weight_map": {
6
  "model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
 
966
  "model.layers.31.post_attention_layernorm.weight.dtype": "model-00001.safetensors",
967
  "model.norm.weight.dtype": "model-00001.safetensors",
968
  "lm_head.weight.dtype": "model-00001.safetensors",
 
 
 
969
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
970
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
971
  "model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
 
1386
  "model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
1387
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.shape": "model-00001.safetensors",
1388
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.shape": "model-00001.safetensors",
1389
+ "model.embed_tokens.weight.shape": "model-00001.safetensors",
1390
  "model.layers.0.input_layernorm.weight.shape": "model-00001.safetensors",
1391
  "model.layers.0.post_attention_layernorm.weight.shape": "model-00001.safetensors",
1392
  "model.layers.1.input_layernorm.weight.shape": "model-00001.safetensors",
 
1481
  "model.layers.31.post_attention_layernorm.weight.shape": "model-00001.safetensors",
1482
  "model.norm.weight.shape": "model-00001.safetensors",
1483
  "lm_head.weight.shape": "model-00001.safetensors",
 
 
 
 
 
 
1484
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
1485
  "model.layers.0.self_attn.q_proj.weight.gptq_4_bit_scale.compressed": "model-00001.safetensors",
1486
  "model.layers.0.self_attn.k_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00001.safetensors",
 
1901
  "model.layers.29.mlp.up_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
1902
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_weight_packed.compressed": "model-00002.safetensors",
1903
  "model.layers.29.mlp.down_proj.weight.gptq_4_bit_scale.compressed": "model-00002.safetensors",
1904
+ "model.embed_tokens.weight.compressed": "model-00002.safetensors",
1905
  "model.layers.0.input_layernorm.weight.compressed": "model-00002.safetensors",
1906
  "model.layers.0.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
1907
  "model.layers.1.input_layernorm.weight.compressed": "model-00002.safetensors",
 
1995
  "model.layers.31.input_layernorm.weight.compressed": "model-00002.safetensors",
1996
  "model.layers.31.post_attention_layernorm.weight.compressed": "model-00002.safetensors",
1997
  "model.norm.weight.compressed": "model-00002.safetensors",
1998
+ "lm_head.weight.compressed": "model-00002.safetensors"
 
1999
  }
2000
  }