georgehalal commited on
Commit
f659e16
·
verified ·
1 Parent(s): c1439cc

Initial Private Upload

Browse files
.gitattributes CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ main_benchmark.png filter=lfs diff=lfs merge=lfs -text
37
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,181 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: cc-by-nc-sa-4.0
4
+ pipeline_tag: text-ranking
5
+ ---
6
+
7
+ # Contextual AI Reranker v2 6B
8
+
9
+ ## Highlights
10
+
11
+ Our reranker is on the cost/performance Pareto frontier across 5 key areas:
12
+ - Instruction following (including capability to rank more recent information higher)
13
+ - Question answering
14
+ - Multilinguality
15
+ - Product search / recommendation systems
16
+ - Real-world use cases
17
+
18
+ <p align="center">
19
+ <img src="main_benchmark.png" width="1200"/>
20
+ <p>
21
+
22
+ For more details on these and other benchmarks, please refer to our [blogpost](https://contextual.ai/blog/rerank-v2).
23
+
24
+ ## Overview
25
+
26
+ - Model Type: Text Reranking
27
+ - Supported Languages: 100+
28
+ - Number of Paramaters: 6B
29
+ - Context Length: up to 32K
30
+ - Blogpost: https://contextual.ai/blog/rerank-v2
31
+
32
+ ## Quickstart
33
+
34
+ ### vLLM usage
35
+
36
+ Requires vllm==0.10.0 for NVFP4 or vllm>=0.8.5 for BF16.
37
+
38
+ ```python
39
+ import os
40
+ os.environ['VLLM_USE_V1'] = '0' # v1 engine doesn’t support logits processor yet
41
+
42
+ import torch
43
+ from vllm import LLM, SamplingParams
44
+
45
+
46
+ def logits_processor(_, scores):
47
+ """Custom logits processor for vLLM reranking."""
48
+ index = scores[0].view(torch.uint16)
49
+ scores = torch.full_like(scores, float("-inf"))
50
+ scores[index] = 1
51
+ return scores
52
+
53
+
54
+ def format_prompts(query: str, instruction: str, documents: list[str]) -> list[str]:
55
+ """Format query and documents into prompts for reranking."""
56
+ if instruction:
57
+ instruction = f" {instruction}"
58
+ prompts = []
59
+ for doc in documents:
60
+ prompt = f"Check whether a given document contains information helpful to answer the query.\n<Document> {doc}\n<Query> {query}{instruction} ??"
61
+ prompts.append(prompt)
62
+ return prompts
63
+
64
+
65
+ def infer_w_vllm(model_path: str, query: str, instruction: str, documents: list[str]):
66
+ model = LLM(
67
+ model=model_path,
68
+ gpu_memory_utilization=0.85,
69
+ max_model_len=8192,
70
+ dtype="bfloat16",
71
+ max_logprobs=2,
72
+ max_num_batched_tokens=262144,
73
+ )
74
+ sampling_params = SamplingParams(
75
+ temperature=0,
76
+ max_tokens=1,
77
+ logits_processors=[logits_processor]
78
+ )
79
+ prompts = format_prompts(query, instruction, documents)
80
+
81
+ outputs = model.generate(prompts, sampling_params, use_tqdm=False)
82
+
83
+ # Extract scores and create results
84
+ results = []
85
+ for i, output in enumerate(outputs):
86
+ score = (
87
+ torch.tensor([output.outputs[0].token_ids[0]], dtype=torch.uint16)
88
+ .view(torch.bfloat16)
89
+ .item()
90
+ )
91
+ results.append((score, i, documents[i]))
92
+
93
+ # Sort by score (descending)
94
+ results = sorted(results, key=lambda x: x[0], reverse=True)
95
+
96
+ print(f"Query: {query}")
97
+ print(f"Instruction: {instruction}")
98
+ for score, doc_id, doc in results:
99
+ print(f"Score: {score:.4f} | Doc: {doc}")
100
+ ```
101
+
102
+
103
+ ### Transformers Usage
104
+
105
+ Requires transformers>=4.51.0 for BF16. Not supported for NVFP4.
106
+
107
+ ```python
108
+ import torch
109
+ from transformers import AutoTokenizer, AutoModelForCausalLM
110
+
111
+
112
+ def format_prompts(query: str, instruction: str, documents: list[str]) -> list[str]:
113
+ """Format query and documents into prompts for reranking."""
114
+ if instruction:
115
+ instruction = f" {instruction}"
116
+ prompts = []
117
+ for doc in documents:
118
+ prompt = f"Check whether a given document contains information helpful to answer the query.\n<Document> {doc}\n<Query> {query}{instruction} ??"
119
+ prompts.append(prompt)
120
+ return prompts
121
+
122
+
123
+ def infer_w_hf(model_path: str, query: str, instruction: str, documents: list[str]):
124
+ device = "cuda" if torch.cuda.is_available() else "cpu"
125
+ dtype = torch.bfloat16 if torch.cuda.is_available() else torch.float32
126
+
127
+ tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=True)
128
+ if tokenizer.pad_token is None:
129
+ tokenizer.pad_token = tokenizer.eos_token
130
+ tokenizer.padding_side = "left" # so -1 is the real last token for all prompts
131
+
132
+ model = AutoModelForCausalLM.from_pretrained(model_path, torch_dtype=dtype).to(device)
133
+ model.eval()
134
+
135
+ prompts = format_prompts(query, instruction, documents)
136
+ enc = tokenizer(
137
+ prompts,
138
+ return_tensors="pt",
139
+ padding=True,
140
+ truncation=True,
141
+ )
142
+ input_ids = enc["input_ids"].to(device)
143
+ attention_mask = enc["attention_mask"].to(device)
144
+
145
+ with torch.no_grad():
146
+ out = model(input_ids=input_ids, attention_mask=attention_mask)
147
+
148
+ next_logits = out.logits[:, -1, :] # [batch, vocab]
149
+
150
+ scores_bf16 = next_logits[:, 0].to(torch.bfloat16)
151
+ scores = scores_bf16.float().tolist()
152
+
153
+ # Sort by score (descending)
154
+ results = sorted([(s, i, documents[i]) for i, s in enumerate(scores)], key=lambda x: x[0], reverse=True)
155
+
156
+ print(f"Query: {query}")
157
+ print(f"Instruction: {instruction}")
158
+ for score, doc_id, doc in results:
159
+ print(f"Score: {score:.4f} | Doc: {doc}")
160
+ ```
161
+
162
+ ## Citation
163
+
164
+ If you use this model, please cite:
165
+
166
+ ```bibtex
167
+ @misc{ctxl_rerank_v2_instruct_multilingual,
168
+ title={Contextual AI Reranker v2},
169
+ author={George Halal, Sheshansh Agrawal, Bo Han, Arnav Palkhiwala},
170
+ year={2025},
171
+ url={https://contextual.ai/blog/rerank-v2},
172
+ }
173
+ ```
174
+
175
+ ## License
176
+
177
+ Creative Commons Attribution Non Commercial Share Alike 4.0 (cc-by-nc-sa-4.0)
178
+
179
+ ## Contact
180
+
181
+ For questions or issues, please open an issue on the model repository or contact [email protected].
config.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "head_dim": 128,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 5120,
11
+ "id2label": {
12
+ "0": "LABEL_0"
13
+ },
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 14336,
16
+ "label2id": {
17
+ "LABEL_0": 0
18
+ },
19
+ "max_position_embeddings": 1024000,
20
+ "model_type": "mistral",
21
+ "num_attention_heads": 32,
22
+ "num_hidden_layers": 20,
23
+ "num_key_value_heads": 8,
24
+ "rms_norm_eps": 1e-05,
25
+ "rope_theta": 1000000.0,
26
+ "sliding_window": null,
27
+ "tie_word_embeddings": false,
28
+ "torch_dtype": "bfloat16",
29
+ "transformers_version": "4.51.3",
30
+ "use_cache": true,
31
+ "vocab_size": 131072
32
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.51.3"
6
+ }
main_benchmark.png ADDED

Git LFS Details

  • SHA256: 73b74cd012b57c5f8e8abfac5135359af7e2758ddaa76f3c05e8ff47de202c84
  • Pointer size: 131 Bytes
  • Size of remote file: 154 kB
model-00001-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2465871762bc01184085f9109bc43502bce735e9a778d9b08951f309291db69c
3
+ size 4865522496
model-00002-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:accf737a92b0504ffa7032f372f0a0b11135be28284131ddc9006feee4e2f11d
3
+ size 4907529424
model-00003-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d4d775b666ac606e92f121d4dcfded4bba72628374f72abbaacf72efc6eaee9
3
+ size 2474756752
model-00004-of-00004.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fff2ef6c64625c518b8388f86c6c3cee57aad1fdad4f7f33878f7b0218f1473f
3
+ size 2684354688
model.safetensors.index.json ADDED
@@ -0,0 +1,190 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "metadata": {
3
+ "total_size": 14932142080
4
+ },
5
+ "weight_map": {
6
+ "lm_head.weight": "model-00004-of-00004.safetensors",
7
+ "model.embed_tokens.weight": "model-00001-of-00004.safetensors",
8
+ "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
9
+ "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
10
+ "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
11
+ "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
12
+ "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
13
+ "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
14
+ "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
15
+ "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
16
+ "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
17
+ "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors",
18
+ "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
19
+ "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
20
+ "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
21
+ "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
22
+ "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
23
+ "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
24
+ "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
25
+ "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
26
+ "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors",
27
+ "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
28
+ "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
29
+ "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
30
+ "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
31
+ "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
32
+ "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
33
+ "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
34
+ "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
35
+ "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors",
36
+ "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
37
+ "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
38
+ "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
39
+ "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
40
+ "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
41
+ "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
42
+ "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
43
+ "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
44
+ "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors",
45
+ "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
46
+ "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
47
+ "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
48
+ "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
49
+ "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
50
+ "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
51
+ "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
52
+ "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
53
+ "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors",
54
+ "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
55
+ "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
56
+ "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
57
+ "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
58
+ "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
59
+ "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
60
+ "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
61
+ "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
62
+ "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors",
63
+ "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
64
+ "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
65
+ "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
66
+ "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
67
+ "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
68
+ "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
69
+ "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
70
+ "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
71
+ "model.layers.15.input_layernorm.weight": "model-00003-of-00004.safetensors",
72
+ "model.layers.15.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
73
+ "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
74
+ "model.layers.15.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
75
+ "model.layers.15.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
76
+ "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
77
+ "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
78
+ "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
79
+ "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
80
+ "model.layers.16.input_layernorm.weight": "model-00003-of-00004.safetensors",
81
+ "model.layers.16.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
82
+ "model.layers.16.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
83
+ "model.layers.16.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
84
+ "model.layers.16.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
85
+ "model.layers.16.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
86
+ "model.layers.16.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
87
+ "model.layers.16.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
88
+ "model.layers.16.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
89
+ "model.layers.17.input_layernorm.weight": "model-00003-of-00004.safetensors",
90
+ "model.layers.17.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
91
+ "model.layers.17.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
92
+ "model.layers.17.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
93
+ "model.layers.17.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
94
+ "model.layers.17.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
95
+ "model.layers.17.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
96
+ "model.layers.17.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
97
+ "model.layers.17.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
98
+ "model.layers.18.input_layernorm.weight": "model-00003-of-00004.safetensors",
99
+ "model.layers.18.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
100
+ "model.layers.18.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
101
+ "model.layers.18.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
102
+ "model.layers.18.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
103
+ "model.layers.18.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
104
+ "model.layers.18.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
105
+ "model.layers.18.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
106
+ "model.layers.18.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
107
+ "model.layers.19.input_layernorm.weight": "model-00003-of-00004.safetensors",
108
+ "model.layers.19.mlp.down_proj.weight": "model-00003-of-00004.safetensors",
109
+ "model.layers.19.mlp.gate_proj.weight": "model-00003-of-00004.safetensors",
110
+ "model.layers.19.mlp.up_proj.weight": "model-00003-of-00004.safetensors",
111
+ "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00004.safetensors",
112
+ "model.layers.19.self_attn.k_proj.weight": "model-00003-of-00004.safetensors",
113
+ "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00004.safetensors",
114
+ "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00004.safetensors",
115
+ "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00004.safetensors",
116
+ "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors",
117
+ "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
118
+ "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
119
+ "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
120
+ "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
121
+ "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
122
+ "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
123
+ "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
124
+ "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
125
+ "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors",
126
+ "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
127
+ "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
128
+ "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
129
+ "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
130
+ "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
131
+ "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
132
+ "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
133
+ "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
134
+ "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors",
135
+ "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
136
+ "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
137
+ "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
138
+ "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
139
+ "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
140
+ "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
141
+ "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
142
+ "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
143
+ "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors",
144
+ "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
145
+ "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
146
+ "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors",
147
+ "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors",
148
+ "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
149
+ "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
150
+ "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
151
+ "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
152
+ "model.layers.6.input_layernorm.weight": "model-00002-of-00004.safetensors",
153
+ "model.layers.6.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
154
+ "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors",
155
+ "model.layers.6.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
156
+ "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
157
+ "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors",
158
+ "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors",
159
+ "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors",
160
+ "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors",
161
+ "model.layers.7.input_layernorm.weight": "model-00002-of-00004.safetensors",
162
+ "model.layers.7.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
163
+ "model.layers.7.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
164
+ "model.layers.7.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
165
+ "model.layers.7.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
166
+ "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
167
+ "model.layers.7.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
168
+ "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
169
+ "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
170
+ "model.layers.8.input_layernorm.weight": "model-00002-of-00004.safetensors",
171
+ "model.layers.8.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
172
+ "model.layers.8.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
173
+ "model.layers.8.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
174
+ "model.layers.8.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
175
+ "model.layers.8.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
176
+ "model.layers.8.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
177
+ "model.layers.8.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
178
+ "model.layers.8.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
179
+ "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors",
180
+ "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors",
181
+ "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors",
182
+ "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors",
183
+ "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors",
184
+ "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors",
185
+ "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors",
186
+ "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors",
187
+ "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors",
188
+ "model.norm.weight": "model-00003-of-00004.safetensors"
189
+ }
190
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "<pad>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0240ce510f08e6c2041724e9043e33be9d251d1e4a4d94eb68cd47b954b61d2
3
+ size 17078292
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff