--- base_model: - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3 - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3-alpaca_gpt4_zh - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4 - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5 - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-reasoning - Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-roleplay library_name: transformers tags: - mergekit - merge - chat pipeline_tag: text-generation --- # NQLSG-Qwen2.5-14B-MegaFusion-v6 This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using NQLSG-Qwen2.5-14B-Base2 as a base. ### Models Merged The following models were included in the merge: * NQLSG-Qwen2.5-14B-Base1 * NQLSG-Qwen2.5-14B-Base2 * NQLSG-Qwen2.5-14B-Base3 * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3-alpaca_gpt4_zh](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3-alpaca_gpt4_zh) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-reasoning](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-reasoning) * [Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-roleplay](https://huggingface.co/Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-roleplay) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: NQLSG-Qwen2.5-14B-Base2 chat_template: auto dtype: bfloat16 merge_method: sce parameters: int8_mask: 1.0 slices: - sources: - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3 - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v3-alpaca_gpt4_zh - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4 - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v4-reasoning - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5 - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-reasoning - layer_range: [0, 48] model: Lunzima/NQLSG-Qwen2.5-14B-MegaFusion-v5-roleplay - layer_range: [0, 48] model: NQLSG-Qwen2.5-14B-Base1 - layer_range: [0, 48] model: NQLSG-Qwen2.5-14B-Base2 - layer_range: [0, 48] model: NQLSG-Qwen2.5-14B-Base3 tokenizer: {} ```