TinyPixel commited on
Commit
9e85ad3
·
verified ·
1 Parent(s): 5a8658a

Upload LlamaForCausalLM

Browse files
config.json CHANGED
@@ -1,8 +1,10 @@
1
  {
2
- "_name_or_path": "meta-llama/Llama-2-7b-hf",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
 
6
  "bos_token_id": 1,
7
  "eos_token_id": 2,
8
  "hidden_act": "silu",
@@ -18,9 +20,10 @@
18
  "pretraining_tp": 1,
19
  "rms_norm_eps": 1e-05,
20
  "rope_scaling": null,
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "float16",
23
  "transformers_version": "4.31.0",
24
  "use_cache": true,
25
- "vocab_size": 32000
26
  }
 
1
  {
2
+ "_name_or_path": "TinyPixel/l2-chatml",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
  "bos_token_id": 1,
9
  "eos_token_id": 2,
10
  "hidden_act": "silu",
 
20
  "pretraining_tp": 1,
21
  "rms_norm_eps": 1e-05,
22
  "rope_scaling": null,
23
+ "rope_theta": 10000.0,
24
  "tie_word_embeddings": false,
25
  "torch_dtype": "float16",
26
  "transformers_version": "4.31.0",
27
  "use_cache": true,
28
+ "vocab_size": 32002
29
  }
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5441103ee26d87bd3e4eaa745a117fc958be0c871f5b77708916c995cea75ec2
3
- size 9976638373
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a6620091cc4106c02e3dfbbfd52fb77ed1b37f79850553e1e0ba830a48b8bd4
3
+ size 9976654757
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6637a30b9250c0c9d06b939323e570b77450bc955c8503a2a77645c0d84f2ba2
3
- size 3500317102
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99bceb4d5e500224d47c0f384d257dc04443715247917a21723d710c91b54b9e
3
+ size 3500333486
pytorch_model.bin.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 13476839424
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00002-of-00002.bin",
 
1
  {
2
  "metadata": {
3
+ "total_size": 13476872192
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "pytorch_model-00002-of-00002.bin",