--- base_model: - TheDrummer/Llama-3SOME-8B-v2 - cgato/L3-TheSpice-8b-v0.8.3 - Sao10K/L3-8B-Stheno-v3.2 - SicariusSicariiStuff/Wingless_Imp_8B - deepseek-ai/DeepSeek-R1-Distill-Llama-8B - NeverSleep/Lumimaid-v0.2-8B library_name: transformers tags: - mergekit - merge license: cc-by-nc-4.0 --- # What is this? A Llama3 model with Deepseek R1 Distill merge. Maybe it's not suit for RP? ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6796de7e92228cec6c379f82/KUXEH72dnhgO-_R3iAqMR.png) Overall, this merge model is the best and smartest RP, ERP model. But the IFEval score is lower than other model, so I think it's wont follow well your instructions? I didn't test yet, will have a test later
## Merge Detail

### Models Merged The following models were included in the merge: * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2) * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3) * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) * [SicariusSicariiStuff/Wingless_Imp_8B](https://huggingface.co/SicariusSicariiStuff/Wingless_Imp_8B) * [deepseek-ai/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: NeverSleep/Lumimaid-v0.2-8B merge_method: model_stock dtype: bfloat16 models: - model: cgato/L3-TheSpice-8b-v0.8.3 - model: Sao10K/L3-8B-Stheno-v3.2 - model: TheDrummer/Llama-3SOME-8B-v2 - model: SicariusSicariiStuff/Wingless_Imp_8B - model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B ```