File size: 1,196 Bytes
696de72 c214471 696de72 5b7d584 696de72 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 |
{
"model_lib": "zephyr-7b-beta-chatRDM-q4f32_1",
"local_id": "zephyr-7b-beta-q4f32_1",
"conv_template": "mistral_default",
"temperature": 0.7,
"repetition_penalty": 1.0,
"top_p": 0.95,
"mean_gen_len": 128,
"max_gen_len": 512,
"num_shards": 1,
"use_presharded_weights": false,
"shift_fill_factor": 0.3,
"conv_config": {
"seps": [
" "
],
"stop_tokens": [
2
],
"offset": 0,
"messages": [
],
"stop_str": "</s>",
"roles": [
"<|user|>",
"<|assistant|>"
],
"role_msg_sep": " ",
"role_empty_sep": " ",
"system": "<|system|> A chat between a curious user and an artificial intelligence research data management assistant. The assistant gives helpful, detailed, and polite answers to the user's questions. [/INST]",
"add_bos": true,
"name": "Llama2"
},
"tokenizer_files": [
"tokenizer.json",
"added_tokens.json",
"tokenizer.model"
],
"model_category": "mistral",
"model_name": "zephyr-7b-beta",
"vocab_size": 32000,
"prefill_chunk_size": 4096,
"sliding_window": 4096
}
|