Commit
·
762466c
1
Parent(s):
8534e29
Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- 1-gpu/config.ini +12 -0
- 1-gpu/model.final_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.0.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.0.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.0.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.0.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.0.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.0.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.0.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.1.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.1.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.1.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.1.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.1.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.1.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.1.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.10.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.10.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.10.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.10.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.10.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.10.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.10.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.11.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.11.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.11.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.11.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.11.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.11.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.11.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.12.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.12.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.12.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.12.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.12.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.12.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.12.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.13.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.13.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.13.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.13.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.13.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.13.mlp.up_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.13.post_attention_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.14.attention.dense.weight.0.bin +3 -0
- 1-gpu/model.layers.14.attention.query_key_value.weight.0.bin +3 -0
- 1-gpu/model.layers.14.input_layernorm.weight.bin +3 -0
- 1-gpu/model.layers.14.mlp.down_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.14.mlp.gate_proj.weight.0.bin +3 -0
- 1-gpu/model.layers.14.mlp.up_proj.weight.0.bin +3 -0
1-gpu/config.ini
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
[llama]
|
2 |
+
model_name = llama_7b
|
3 |
+
head_num = 32
|
4 |
+
size_per_head = 128
|
5 |
+
inter_size = 11008
|
6 |
+
num_layer = 32
|
7 |
+
rotary_embedding = 128
|
8 |
+
vocab_size = 32000
|
9 |
+
start_id = 1
|
10 |
+
end_id = 2
|
11 |
+
weight_data_type = fp16
|
12 |
+
|
1-gpu/model.final_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ea571ad0c73c4eca486a89f8cca58b8b8ae3e0ce997634d6de000b7b40ac3a5
|
3 |
+
size 8192
|
1-gpu/model.layers.0.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:837e3b7e9d9e8ee6c24b7cd37e7f9f86ff53e50e1cddfd452d7f0120100388c9
|
3 |
+
size 33554432
|
1-gpu/model.layers.0.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8c8bccd35632748628bcff6b1a4974245f2637b68c394786972d5621778f45ff
|
3 |
+
size 100663296
|
1-gpu/model.layers.0.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11e1c646ebdb46f1455001ba01386d5372b5d306c353278965240bfd9a2ab226
|
3 |
+
size 8192
|
1-gpu/model.layers.0.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2d6fb4f93ad6745bae9a4d9f7e51133fd4c2327bc9c1f56a3695f98845f543fc
|
3 |
+
size 90177536
|
1-gpu/model.layers.0.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d915c77a4cf3bf1726b9cdb711ef24d7b9d9afacd82802622fb77e9ef09e5cf1
|
3 |
+
size 90177536
|
1-gpu/model.layers.0.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fcf17596900379dc84c4c361c19a8fba25a25a53fd0e7f405d139d6098662274
|
3 |
+
size 90177536
|
1-gpu/model.layers.0.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d95b916c82ce2346b74ba775f52626b698643bd996b0cdfb015299d390a5c566
|
3 |
+
size 8192
|
1-gpu/model.layers.1.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49581334ccce1d84a111efc1abdd4e5e1d81c9ffc3f5b08d1dad2c50608d28eb
|
3 |
+
size 33554432
|
1-gpu/model.layers.1.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7852d8f601ad3f02e215ea4060c12db0c96050780cd2985c8306c5e8f03b24b5
|
3 |
+
size 100663296
|
1-gpu/model.layers.1.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:382195c9a081a4d63e7485e60447ff247cdaac2c1bbad9123882c98b36b6f865
|
3 |
+
size 8192
|
1-gpu/model.layers.1.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69647835c0ac49bec6e4832f5477cc22eadbafe3ba80357e5a79fb616abfe17b
|
3 |
+
size 90177536
|
1-gpu/model.layers.1.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09f9bb83a8e50891cd7093e89e677d1a8b3523779f08a6f15c199a9f73f03502
|
3 |
+
size 90177536
|
1-gpu/model.layers.1.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bcadfe61fc55d7a0d3179a1333c9f25b1662c2c8666a735d97203c7b0b82bcca
|
3 |
+
size 90177536
|
1-gpu/model.layers.1.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dc62f72057280170edac6bf748860b3573385597b909ac6d5a38a4ff838da3cc
|
3 |
+
size 8192
|
1-gpu/model.layers.10.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80e6d972184a1825a0727f25292d299dbeda7007c3d7e070830a6100c17a9c74
|
3 |
+
size 33554432
|
1-gpu/model.layers.10.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bd0f66b2e3c454a02b4c4b20162864c92ea37457fe1d3dded5a6c4387d36e69e
|
3 |
+
size 100663296
|
1-gpu/model.layers.10.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e89276aabd344f19e5c82eeb64ac8cc683dfbfc09165d7b1ee5014280c733dea
|
3 |
+
size 8192
|
1-gpu/model.layers.10.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b6fbce13907b6ed2901e9928037e3bb443567f7545362e0c0cc534e191a714b2
|
3 |
+
size 90177536
|
1-gpu/model.layers.10.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1eb1c249b4406959dfbbfdd04b555c74a706e62f811bbfb5125684bae0873e7
|
3 |
+
size 90177536
|
1-gpu/model.layers.10.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd7e038987c45a0f0d263fc2f409a57d9f4c289d08b58a8ea3eb94500bd79366
|
3 |
+
size 90177536
|
1-gpu/model.layers.10.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f64a8ce510534a9c064a78e062b1fe0a9ee7474c86f278db66a71fe7b0751fe5
|
3 |
+
size 8192
|
1-gpu/model.layers.11.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c780ad7e11d870a182709a908bf7323a2a9d0f107b93456efe0b8da7c7d6698e
|
3 |
+
size 33554432
|
1-gpu/model.layers.11.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:25ccaa53224f9dd31b19007df68a623cfecc9ac507e8de5a2e71cb643a5fd760
|
3 |
+
size 100663296
|
1-gpu/model.layers.11.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6e01a601f8670aae63a7fe0090ff9110ac07cfed30c4b3cf077e43fef020a4cd
|
3 |
+
size 8192
|
1-gpu/model.layers.11.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2693edf13f731a36f89e737d2232a88bd525c7f1e93fa13cc2bf11d704d24c9f
|
3 |
+
size 90177536
|
1-gpu/model.layers.11.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:95a9c1074ec9f813de0a6eab281bab874f76427b3ae498a7c9edd7c5f067d431
|
3 |
+
size 90177536
|
1-gpu/model.layers.11.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f5a3db84d6908c7211d9e1c2914504a33e73d8b71338fe92bcaabc3c89cb4732
|
3 |
+
size 90177536
|
1-gpu/model.layers.11.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e07345b8240886d32e563f64f94d55d7888180d8a466ed78ed349a0d1c8d48dc
|
3 |
+
size 8192
|
1-gpu/model.layers.12.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f94694d6ca2c58f58403f3ea1e42128a1a436f4fab4654d7f08873130296105
|
3 |
+
size 33554432
|
1-gpu/model.layers.12.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e3d3705b5bdee5d8b94680fc2676fa714a79c6fb57202ec20660712caaebb6f2
|
3 |
+
size 100663296
|
1-gpu/model.layers.12.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8c141d8c5c9e73025d002b2c7cbf01576ed2876159682dd12a1f95f798abae6
|
3 |
+
size 8192
|
1-gpu/model.layers.12.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:473d54eedd2167db6fe7356b72412399b6610fffb5f6a57a68e4887ddd61454a
|
3 |
+
size 90177536
|
1-gpu/model.layers.12.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8dd967a2ec9fce42253930d134a1a51515c98a80d171a23c832752e1816ad572
|
3 |
+
size 90177536
|
1-gpu/model.layers.12.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:21103ab5aa964a116e4d6af6e941679e8a9f343a8239b4f1496e3eec66bb5043
|
3 |
+
size 90177536
|
1-gpu/model.layers.12.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:997d9661459cb3787e78476e43937b29b6eed74fef33b727e82e0cc28e8fde27
|
3 |
+
size 8192
|
1-gpu/model.layers.13.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed557a7bc0641cdbf7f5346c289392cbb83f87184a70617f5ed28a2341f4055a
|
3 |
+
size 33554432
|
1-gpu/model.layers.13.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9c5336639adddf66d1b391540bfced65384d6cfca970fc6570d4f29ca1147b1d
|
3 |
+
size 100663296
|
1-gpu/model.layers.13.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6370e77f39777eb6463a09b5e3b950f73d1af015d5c72bd3fdcb496a53c32102
|
3 |
+
size 8192
|
1-gpu/model.layers.13.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c20c3198f96dc48e0abe68fdd8e2a931f318dc94dee30e507dd9726b5ef31b0
|
3 |
+
size 90177536
|
1-gpu/model.layers.13.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:79c6206cd6c92faef5b59847e2063cc670d2f0f44b7ea25318699b37601ed6d8
|
3 |
+
size 90177536
|
1-gpu/model.layers.13.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:200b5047a5fa6d0be82d996edf89b216678ae3abb12bf58a9f3328368f70549a
|
3 |
+
size 90177536
|
1-gpu/model.layers.13.post_attention_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:501d56f776a08107289b94c570589fa48e10832af11fbb28dc2234ce031702cc
|
3 |
+
size 8192
|
1-gpu/model.layers.14.attention.dense.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0aadf257305b47ffcef8acbb32d0dabbe3f456f1b585997ca485f886dffdc331
|
3 |
+
size 33554432
|
1-gpu/model.layers.14.attention.query_key_value.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ad531200d177e92137e5053c9b74785cc61785a491f2afc2bdac4d8ca9ac4e17
|
3 |
+
size 100663296
|
1-gpu/model.layers.14.input_layernorm.weight.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:11d8cebc52794aa4ff4fa99accfe8a18b40588d34a540a07b689984e5b07de5b
|
3 |
+
size 8192
|
1-gpu/model.layers.14.mlp.down_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8dc9e846877b9b7796bcaa00e782dd0955414b88cde0ca4b4e3ce3d50e2762d6
|
3 |
+
size 90177536
|
1-gpu/model.layers.14.mlp.gate_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e47be6d117882296a90b5fbdc81b9ed26b564c8e1045929de3cc2b71b314d615
|
3 |
+
size 90177536
|
1-gpu/model.layers.14.mlp.up_proj.weight.0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b0fd1d5d7e598bd3454807304972cad9afee17e37d6cf37659f03bd2fcda230
|
3 |
+
size 90177536
|