--- license: apache-2.0 datasets: - ToastyPigeon/roselily-v0 - PocketDoc/Dans-Systemmaxx - allenai/tulu-3-sft-personas-instruction-following - ZeusLabs/WizardLM_evol_instruct_fuzzy_dedup_sharegpt base_model: - mistralai/Mistral-Small-24B-Base-2501 --- This is a double fine-tuned version of Mistral Small 24B Base 2501. Stage 1 was shoving 30M tokens of human-writen story content into it using completion training ([ToastyPigeon/ms3-base-roselily](https://huggingface.co/ToastyPigeon/ms3-base-roselily)), which is about half of my WIP Roselily dataset (~60M tokens total). Stage 2 was teaching it instruct (this model) using a mix of public instruction following data and a private instruct dataset from ZeusLabs. This model should accept (in theory) any of the following instruct formats: **Tekken v7** ``` [SYSTEM_PROMPT]{system prompt}[/SYSTEM_PROMPT][INST]{user message}[/INST]{assistant response} ``` **ChatML** ``` <|im_start|>system {system prompt}<|im_end|> <|im_start|>user {user message}<|im_end|> <|im_start|>assistant {assistant response}<|im_end|> ``` **Fizzpaca** ``` ### System: {system prompt} ### Instruction: {user message} ### Response: {assistant response} ``` The Tekken tokens were already in the tokenizer. unused special tokens #20 and 21 were repurposed for the ChatML tokens. Fizzpaca did not add any. You may need to add both `` and `<|im_end|>` as stop tokens for it to work properly with all formats.