Update model.yml
Browse files
model.yml
CHANGED
@@ -16,7 +16,7 @@ stream: true # true | false
|
|
16 |
# Engine / Model Settings
|
17 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
18 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
19 |
-
engine:
|
20 |
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
21 |
# Prompt template: Can only be retrieved from instruct model
|
22 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|
|
|
16 |
# Engine / Model Settings
|
17 |
ngl: 33 # Infer from base config.json -> num_attention_heads
|
18 |
ctx_len: 8192 # Infer from base config.json -> max_position_embeddings
|
19 |
+
engine: llama-cpp
|
20 |
prompt_template: "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\n{system_message}<|eot_id|><|start_header_id|>user<|end_header_id|>\n\n{prompt}<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n"
|
21 |
# Prompt template: Can only be retrieved from instruct model
|
22 |
# - https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct/blob/main/tokenizer_config.json#L2053
|