Darkknight535's picture
Update README.md
4bf9f91 verified
|
raw
history blame
3.36 kB
metadata
base_model:
  - PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
  - SicariusSicariiStuff/Redemption_Wind_24B
tags:
  - merge
  - mergekit
  - lazymergekit
  - PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
  - SicariusSicariiStuff/Redemption_Wind_24B

WindEngine-24B-Instruct

WindEngine-24B-Instruct is a merge of the following models using LazyMergekit:

WinterEngine-24B-Instruct

A versatile and powerful model designed for general-purpose text generation, roleplay, storywriting, scientific exploration, and more.

<div class="details">
    <h3>Key Details</h3>
    <pre>

BASE MODEL: mistralai/Mistral-Small-24B-Base-2501 LICENSE: apache-2.0 LANGUAGE: English CONTEXT LENGTH: 32768 tokens

<div class="settings">
    <h3>Recommended Settings</h3>
    <pre>

TEMPERATURE: 1.0 TOP_P: 0.95 MIN_P: 0.05

<div class="prompt-format">
    <h3>Prompting Format</h3>
    <pre>

<|im_start|>system system prompt<|im_end|> <|im_start|>user Hi there!<|im_end|> <|im_start|>assistant Nice to meet you!<|im_end|>

🧩 Configuration

slices:
  - sources:
      - model: PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
        layer_range: [0, 40]
      - model: SicariusSicariiStuff/Redemption_Wind_24B
        layer_range: [0, 40]
merge_method: slerp
base_model: PocketDoc/Dans-PersonalityEngine-V1.2.0-24b
parameters:
  t:
    - filter: self_attn
      value: [0, 0.5, 0.3, 0.7, 1]
    - filter: mlp
      value: [1, 0.5, 0.7, 0.3, 0]
    - value: 0.5
dtype: bfloat16