--- base_model: - TheDrummer/Anubis-70B-v1 - SicariusSicariiStuff/Negative_LLAMA_70B - LatitudeGames/Wayfarer-Large-70B-Llama-3.3 - KaraKaraWitch/Llama-MiraiFanfare-3.3-70B - Black-Ink-Guild/Pernicious_Prophecy_70B - KaraKaraWitch/Llama-3.X-Workout-70B library_name: transformers tags: - mergekit - merge ---
New merge. This an experiment to increase the "Madness" in a model. Merge is based on top UGI-Bench models (So yeah, I would think this would be benchmaxxing.) This is the second time I'm using SCE. The previous MagicalGirl model seems to be quite happy with it. Added `KaraKaraWitch/Llama-MiraiFanfare-3.3-70B` based on feedback I got from others (People generally seem to remember this rather than other models). So I'm not sure how this would play into the merge. This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## UGI-Results Pretty interesting. As of 05/03/25, it's in the top 10th: | Bench | Results | |---------------|--------------------| | UGI-Score | 52.48 / 100 | | Unruly | 3.8 / 10 | | Internet | 5.1 / 10 | | Society | 5.4 / 10 | | Willing | 7 / 10 | | NatInt | 41.86 / 100 | | Coding | 22 | | Politial Lean | −3.9% (Liberalism) | ## Merge Details ### Merge Method This model was merged using the [SCE](https://arxiv.org/abs/2408.07990) merge method using [KaraKaraWitch/Llama-3.X-Workout-70B](https://huggingface.co/KaraKaraWitch/Llama-3.X-Workout-70B) as a base. ### Models Merged The following models were included in the merge: * [TheDrummer/Anubis-70B-v1](https://huggingface.co/TheDrummer/Anubis-70B-v1) * [SicariusSicariiStuff/Negative_LLAMA_70B](https://huggingface.co/SicariusSicariiStuff/Negative_LLAMA_70B) * [LatitudeGames/Wayfarer-Large-70B-Llama-3.3](https://huggingface.co/LatitudeGames/Wayfarer-Large-70B-Llama-3.3) * [KaraKaraWitch/Llama-MiraiFanfare-3.3-70B](https://huggingface.co/KaraKaraWitch/Llama-MiraiFanfare-3.3-70B) * [Black-Ink-Guild/Pernicious_Prophecy_70B](https://huggingface.co/Black-Ink-Guild/Pernicious_Prophecy_70B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: SicariusSicariiStuff/Negative_LLAMA_70B - model: TheDrummer/Anubis-70B-v1 - model: KaraKaraWitch/Llama-MiraiFanfare-3.3-70B - model: Black-Ink-Guild/Pernicious_Prophecy_70B - model: LatitudeGames/Wayfarer-Large-70B-Llama-3.3 merge_method: sce base_model: KaraKaraWitch/Llama-3.X-Workout-70B parameters: select_topk: 1.0 dtype: bfloat16 ```