--- base_model: - PocketDoc/Dans-PersonalityEngine-V1.2.0-24b - SicariusSicariiStuff/Redemption_Wind_24B tags: - merge - mergekit - lazymergekit - PocketDoc/Dans-PersonalityEngine-V1.2.0-24b - SicariusSicariiStuff/Redemption_Wind_24B --- # WindEngine-24B-Instruct WindEngine-24B-Instruct is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [PocketDoc/Dans-PersonalityEngine-V1.2.0-24b](https://huggingface.co/PocketDoc/Dans-PersonalityEngine-V1.2.0-24b) * [SicariusSicariiStuff/Redemption_Wind_24B](https://huggingface.co/SicariusSicariiStuff/Redemption_Wind_24B)
WinterEngine-24B-Instruct

A versatile and powerful model designed for general-purpose text generation, roleplay, storywriting, scientific exploration, and more.

Key Details

BASE MODEL: mistralai/Mistral-Small-24B-Base-2501
LICENSE: apache-2.0
LANGUAGE: English
CONTEXT LENGTH: 32768 tokens
        

Recommended Settings

TEMPERATURE: 1.0
TOP_P: 0.95
MIN_P: 0.05
        

Prompting Format

<|im_start|>system
system prompt<|im_end|>
<|im_start|>user
Hi there!<|im_end|>
<|im_start|>assistant
Nice to meet you!<|im_end|>
        
## 🧩 Configuration ```yaml slices: - sources: - model: PocketDoc/Dans-PersonalityEngine-V1.2.0-24b layer_range: [0, 40] - model: SicariusSicariiStuff/Redemption_Wind_24B layer_range: [0, 40] merge_method: slerp base_model: PocketDoc/Dans-PersonalityEngine-V1.2.0-24b parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ```