user
commited on
Commit
·
937476a
1
Parent(s):
b018a2f
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- README.md +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.input.0.1x6.int64.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.1.1x6.int64.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.0.1x6x64.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.1.1x6x64.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.output.1x6x512.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.output.1x6x512.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000007.model.layers.0.self_attn.output.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.output.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.output.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.output.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.input.0.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.output.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.output.1x6x512.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.output.1x6x512.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000020.model.layers.1.self_attn.output.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.output.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.output.1x6x8192.bfloat16.mmap +3 -0
- meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000024.model.layers.1.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
README.md
CHANGED
@@ -1,3 +1,6 @@
|
|
1 |
---
|
2 |
license: pddl
|
3 |
---
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: pddl
|
3 |
---
|
4 |
+
|
5 |
+
This repository contains the complete hidden states of large language models
|
6 |
+
completing the text "Once upon a time," using transformers.pipeline .
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.input.0.1x6.int64.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:85f21c71560487a86b380ea852aa2caddfa0a57cd8a62285202876ddd514e631
|
3 |
+
size 48
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.1.1x6.int64.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f190072c5052f4f440d4a607c25f5bced487c420806c9aab4ca5b0653e72da61
|
3 |
+
size 48
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.0.1x6x64.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53a032642e73a762986d86dfefad074e6d2b36af427145d4006b629d53619e05
|
3 |
+
size 768
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.1.1x6x64.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8977cb35819c130544bd972211aef7b994d5af657f02579276ecfb0f40a849ce
|
3 |
+
size 768
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5146d1a828356b17161527a75ea1e04710061d64ac355dc76593288fb23a3eea
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.output.1x6x512.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:299581fa75d9e4f245a830a1f4a72db9bf183f0d13e85011f9bb4e3145443ddc
|
3 |
+
size 6144
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.output.1x6x512.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6ab709b55384fffe2b2c52931638cf9d716ece4bfa969bcd615be433c2a1ad3f
|
3 |
+
size 6144
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c3c7e25b79438927812afe3bb6ccd6bfd6a577532adffc3a8ff25f666bea68b
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c33c60ba4d4008947c1bee59519f7e0e78e0cf2aeedfc6f7dea5a435a542409a
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000007.model.layers.0.self_attn.output.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c33c60ba4d4008947c1bee59519f7e0e78e0cf2aeedfc6f7dea5a435a542409a
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26266ae73b0857ee8f45b37a9b22d3d279c125334cf129466d2ce30797d1f7ee
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.output.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9b91603c7feb05cddf562dbde96142fd8a5c70353d6e132cba15630212ac8c0
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a9b91603c7feb05cddf562dbde96142fd8a5c70353d6e132cba15630212ac8c0
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.output.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8b7b1da88c64653a5e0c713fb36b5d4bc4b5e29cac26a290a0b46be8095bd2be
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.output.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d5b75929bc7be18773ff15732beae706ad8dc413d0f18fe6538c312d218484ef
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.input.0.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7884a4430db96a6f29f5523bb2d9ff0ea2f690fcb0e7068465bd1562bb2c7f82
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1bbcb2afbd88b193ad9de8faa2d39bb665a29592bf90aa67d8559ff3adaf507b
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1bbcb2afbd88b193ad9de8faa2d39bb665a29592bf90aa67d8559ff3adaf507b
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.output.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a97b3506c1182db4069f509875c57eda36663b2f3c69955c9e0f697e0e418e28
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a97b3506c1182db4069f509875c57eda36663b2f3c69955c9e0f697e0e418e28
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:264bace9d4aab425bc2f284c8ff1a15813a458f4041501f50b8ea6f56a14943d
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.output.1x6x512.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d94b9354ffde37b0f07571160ac0414ed330b0c4072974d2983448df8607c68e
|
3 |
+
size 6144
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.output.1x6x512.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a11aad8666bdb5d0b5978e361e26c3da7fca38d6d01b09f0bcd10fc4db706ca5
|
3 |
+
size 6144
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a30f13ce8ed0af14dd1211bc382a89d6c9428d514c560b2f9984137920d4d940
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c2686f1c95249c262cd971782c7363f41b73d962348635c0138e4128afcb044
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000020.model.layers.1.self_attn.output.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c2686f1c95249c262cd971782c7363f41b73d962348635c0138e4128afcb044
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80d5346be58bf3ee9786baae7ce785ed362108acb190e01b8d7ea65c416c0347
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.output.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
|
3 |
+
size 24576
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.output.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:900b4590f619831e234f5c254e1b0471ea8c5b691c129d0d8e85729230abc64b
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:900b4590f619831e234f5c254e1b0471ea8c5b691c129d0d8e85729230abc64b
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.output.1x6x8192.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ab84316e85416844fb3deaeaf4828f03ead23d785b0b69dadb81e43163b22ec6
|
3 |
+
size 98304
|
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000024.model.layers.1.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
|
3 |
+
size 24576
|