user commited on
Commit
937476a
·
1 Parent(s): b018a2f

meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. README.md +3 -0
  2. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.input.0.1x6.int64.mmap +3 -0
  3. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.output.1x6x2048.bfloat16.mmap +3 -0
  4. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.0.1x6x2048.bfloat16.mmap +3 -0
  5. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.1.1x6.int64.mmap +3 -0
  6. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.0.1x6x64.bfloat16.mmap +3 -0
  7. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.1.1x6x64.bfloat16.mmap +3 -0
  8. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
  9. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
  10. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  11. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.output.1x6x2048.bfloat16.mmap +3 -0
  12. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  13. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.output.1x6x512.bfloat16.mmap +3 -0
  14. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  15. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.output.1x6x512.bfloat16.mmap +3 -0
  16. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  17. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.output.1x6x2048.bfloat16.mmap +3 -0
  18. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000007.model.layers.0.self_attn.output.0.1x6x2048.bfloat16.mmap +3 -0
  19. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
  20. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
  21. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  22. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.output.1x6x8192.bfloat16.mmap +3 -0
  23. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap +3 -0
  24. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.output.1x6x8192.bfloat16.mmap +3 -0
  25. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  26. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.output.1x6x8192.bfloat16.mmap +3 -0
  27. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.input.0.1x6x8192.bfloat16.mmap +3 -0
  28. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.output.1x6x2048.bfloat16.mmap +3 -0
  29. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.input.0.1x6x2048.bfloat16.mmap +3 -0
  30. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.output.1x6x2048.bfloat16.mmap +3 -0
  31. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.input.0.1x6x2048.bfloat16.mmap +3 -0
  32. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.output.0.1x6x2048.bfloat16.mmap +3 -0
  33. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
  34. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
  35. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  36. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.output.1x6x2048.bfloat16.mmap +3 -0
  37. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  38. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.output.1x6x512.bfloat16.mmap +3 -0
  39. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  40. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.output.1x6x512.bfloat16.mmap +3 -0
  41. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  42. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.output.1x6x2048.bfloat16.mmap +3 -0
  43. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000020.model.layers.1.self_attn.output.0.1x6x2048.bfloat16.mmap +3 -0
  44. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap +3 -0
  45. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.output.1x6x2048.bfloat16.mmap +3 -0
  46. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
  47. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.output.1x6x8192.bfloat16.mmap +3 -0
  48. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap +3 -0
  49. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.output.1x6x8192.bfloat16.mmap +3 -0
  50. meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000024.model.layers.1.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap +3 -0
README.md CHANGED
@@ -1,3 +1,6 @@
1
  ---
2
  license: pddl
3
  ---
 
 
 
 
1
  ---
2
  license: pddl
3
  ---
4
+
5
+ This repository contains the complete hidden states of large language models
6
+ completing the text "Once upon a time," using transformers.pipeline .
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.input.0.1x6.int64.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85f21c71560487a86b380ea852aa2caddfa0a57cd8a62285202876ddd514e631
3
+ size 48
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000000.model.embed_tokens.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.input.1.1x6.int64.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f190072c5052f4f440d4a607c25f5bced487c420806c9aab4ca5b0653e72da61
3
+ size 48
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.0.1x6x64.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53a032642e73a762986d86dfefad074e6d2b36af427145d4006b629d53619e05
3
+ size 768
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000001.model.rotary_emb.output.1.1x6x64.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8977cb35819c130544bd972211aef7b994d5af657f02579276ecfb0f40a849ce
3
+ size 768
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000002.model.layers.0.input_layernorm.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000003.model.layers.0.self_attn.q_proj.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5146d1a828356b17161527a75ea1e04710061d64ac355dc76593288fb23a3eea
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000004.model.layers.0.self_attn.k_proj.output.1x6x512.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299581fa75d9e4f245a830a1f4a72db9bf183f0d13e85011f9bb4e3145443ddc
3
+ size 6144
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97983665225acc9951f3aecf404db7178de6531935b1fb38ec8c4747b7ad5de4
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000005.model.layers.0.self_attn.v_proj.output.1x6x512.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ab709b55384fffe2b2c52931638cf9d716ece4bfa969bcd615be433c2a1ad3f
3
+ size 6144
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c3c7e25b79438927812afe3bb6ccd6bfd6a577532adffc3a8ff25f666bea68b
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000006.model.layers.0.self_attn.o_proj.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c33c60ba4d4008947c1bee59519f7e0e78e0cf2aeedfc6f7dea5a435a542409a
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000007.model.layers.0.self_attn.output.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c33c60ba4d4008947c1bee59519f7e0e78e0cf2aeedfc6f7dea5a435a542409a
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26266ae73b0857ee8f45b37a9b22d3d279c125334cf129466d2ce30797d1f7ee
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000008.model.layers.0.post_attention_layernorm.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000009.model.layers.0.mlp.gate_proj.output.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b91603c7feb05cddf562dbde96142fd8a5c70353d6e132cba15630212ac8c0
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9b91603c7feb05cddf562dbde96142fd8a5c70353d6e132cba15630212ac8c0
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000010.model.layers.0.mlp.act_fn.output.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b7b1da88c64653a5e0c713fb36b5d4bc4b5e29cac26a290a0b46be8095bd2be
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000011.model.layers.0.mlp.up_proj.output.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5b75929bc7be18773ff15732beae706ad8dc413d0f18fe6538c312d218484ef
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.input.0.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7884a4430db96a6f29f5523bb2d9ff0ea2f690fcb0e7068465bd1562bb2c7f82
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000012.model.layers.0.mlp.down_proj.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bbcb2afbd88b193ad9de8faa2d39bb665a29592bf90aa67d8559ff3adaf507b
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df245f870cb3208fab67737b6d355523d22732374a9fcf7262e14559e1d2b134
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000013.model.layers.0.mlp.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bbcb2afbd88b193ad9de8faa2d39bb665a29592bf90aa67d8559ff3adaf507b
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83a89fe5677c4e1de735ab2c833dc390407412f50a5e6be2a48c51f5af04a429
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000014.model.layers.0.output.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a97b3506c1182db4069f509875c57eda36663b2f3c69955c9e0f697e0e418e28
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a97b3506c1182db4069f509875c57eda36663b2f3c69955c9e0f697e0e418e28
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000015.model.layers.1.input_layernorm.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000016.model.layers.1.self_attn.q_proj.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:264bace9d4aab425bc2f284c8ff1a15813a458f4041501f50b8ea6f56a14943d
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000017.model.layers.1.self_attn.k_proj.output.1x6x512.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d94b9354ffde37b0f07571160ac0414ed330b0c4072974d2983448df8607c68e
3
+ size 6144
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edc25a7ee5287691c3905273c599c893976e9795aa5aa6f766ce7719d7660864
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000018.model.layers.1.self_attn.v_proj.output.1x6x512.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a11aad8666bdb5d0b5978e361e26c3da7fca38d6d01b09f0bcd10fc4db706ca5
3
+ size 6144
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a30f13ce8ed0af14dd1211bc382a89d6c9428d514c560b2f9984137920d4d940
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000019.model.layers.1.self_attn.o_proj.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c2686f1c95249c262cd971782c7363f41b73d962348635c0138e4128afcb044
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000020.model.layers.1.self_attn.output.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c2686f1c95249c262cd971782c7363f41b73d962348635c0138e4128afcb044
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80d5346be58bf3ee9786baae7ce785ed362108acb190e01b8d7ea65c416c0347
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000021.model.layers.1.post_attention_layernorm.output.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
3
+ size 24576
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000022.model.layers.1.mlp.gate_proj.output.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:900b4590f619831e234f5c254e1b0471ea8c5b691c129d0d8e85729230abc64b
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.input.0.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:900b4590f619831e234f5c254e1b0471ea8c5b691c129d0d8e85729230abc64b
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000023.model.layers.1.mlp.act_fn.output.1x6x8192.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab84316e85416844fb3deaeaf4828f03ead23d785b0b69dadb81e43163b22ec6
3
+ size 98304
meta-llama_Llama-3.2-1B_4e20de362430cd3b72f300e6b0f18e50e7166e08/0000024.model.layers.1.mlp.up_proj.input.0.1x6x2048.bfloat16.mmap ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:299b07c26a8302cf9e70111a95ff72e53f19b07a9aa8f7ddfefc07ba93ecb3c1
3
+ size 24576