rd211 commited on
Commit
540a14b
·
verified ·
1 Parent(s): e07cd4f

Upload model

Browse files
config.json CHANGED
@@ -5,8 +5,8 @@
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
- "bos_token_id": 90000,
9
- "eos_token_id": 90001,
10
  "head_dim": 128,
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
@@ -19,35 +19,6 @@
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
21
  "pretraining_tp": 1,
22
- "quantization_config": {
23
- "config_groups": {
24
- "group_0": {
25
- "input_activations": null,
26
- "output_activations": null,
27
- "targets": [
28
- "Linear"
29
- ],
30
- "weights": {
31
- "actorder": null,
32
- "block_structure": null,
33
- "dynamic": false,
34
- "group_size": null,
35
- "num_bits": 8,
36
- "observer": "minmax",
37
- "observer_kwargs": {},
38
- "strategy": "channel",
39
- "symmetric": true,
40
- "type": "int"
41
- }
42
- }
43
- },
44
- "format": "pack-quantized",
45
- "global_compression_ratio": 1.4619490646162168,
46
- "ignore": [],
47
- "kv_cache_scheme": null,
48
- "quant_method": "compressed-tensors",
49
- "quantization_status": "compressed"
50
- },
51
  "rms_norm_eps": 1e-05,
52
  "rope_scaling": {
53
  "factor": 8.0,
@@ -61,5 +32,5 @@
61
  "torch_dtype": "bfloat16",
62
  "transformers_version": "4.46.0.dev0",
63
  "use_cache": true,
64
- "vocab_size": 90002
65
  }
 
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
+ "bos_token_id": 16000,
9
+ "eos_token_id": 16001,
10
  "head_dim": 128,
11
  "hidden_act": "silu",
12
  "hidden_size": 4096,
 
19
  "num_hidden_layers": 32,
20
  "num_key_value_heads": 8,
21
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
22
  "rms_norm_eps": 1e-05,
23
  "rope_scaling": {
24
  "factor": 8.0,
 
32
  "torch_dtype": "bfloat16",
33
  "transformers_version": "4.46.0.dev0",
34
  "use_cache": true,
35
+ "vocab_size": 16002
36
  }
model-00001.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ad7bfbbc4af2329771ae80b392f84f56a2135eb6034140b2885dfcc58a753dd0
3
  size 2127524202
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52ba4109a3588dceeeaa6a0e95afbce89cfce1959f8fb0f26c2a2c9aaace711d
3
  size 2127524202
model-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c1f2decb4df6334cf086aa124ebbf54055753674021c9626a7597fe7b9fc76e
3
- size 1592416566
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f3444171638be798a157c39429f4c481aa06cb2741274cfe97e6ebd2be73f79
3
+ size 1134584970
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 3719696208
4
  },
5
  "weight_map": {
6
  "model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 3261864772
4
  },
5
  "weight_map": {
6
  "model.layers.0.self_attn.q_proj.weight.shape": "model-00001.safetensors",
special_tokens_map.json CHANGED
@@ -12,6 +12,5 @@
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
- },
16
- "pad_token": "<|end_of_text|>"
17
  }
 
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
+ }
 
16
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a26f4c94b5df1ea07103f220db22a49948724ee7eaabcf1e57cf124c715c28a
3
- size 5559269
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1815b9fadc431c29dfcb0bb41b6f488388f0d23f7ab41270ce7ea08ca9000b42
3
+ size 839048
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"added_tokens_decoder": {"90000": {"content": "<|begin_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true}, "90001": {"content": "<|end_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true}}, "bos_token": "<|begin_of_text|>", "clean_up_tokenization_spaces": true, "eos_token": "<|end_of_text|>", "model_input_names": ["input_ids", "attention_mask"], "model_max_length": 131072, "pad_token": "<|end_of_text|>", "tokenizer_class": "PreTrainedTokenizerFast"}
 
1
+ {"added_tokens_decoder": {"16000": {"content": "<|begin_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true}, "16001": {"content": "<|end_of_text|>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false, "special": true}}, "bos_token": "<|begin_of_text|>", "clean_up_tokenization_spaces": true, "eos_token": "<|end_of_text|>", "model_input_names": ["input_ids", "attention_mask"], "model_max_length": 131072, "tokenizer_class": "PreTrainedTokenizerFast"}