| | --- |
| | base_model: |
| | - TheDrummer/Llama-3SOME-8B-v2 |
| | - cgato/L3-TheSpice-8b-v0.8.3 |
| | - Sao10K/L3-8B-Stheno-v3.2 |
| | - SicariusSicariiStuff/Wingless_Imp_8B |
| | - deepseek-ai/DeepSeek-R1-Distill-Llama-8B |
| | - NeverSleep/Lumimaid-v0.2-8B |
| | library_name: transformers |
| | tags: |
| | - mergekit |
| | - merge |
| | license: cc-by-nc-4.0 |
| | --- |
| | # What is this? |
| |
|
| | A Llama3 model with Deepseek R1 Distill merge. Maybe it's not suit for RP? |
| |
|
| |  |
| |
|
| | Overall, this merge model is the best and smartest RP, ERP model. But the IFEval score is lower than other model, so I think it's wont follow well your instructions? I didn't test yet, will have a test later |
| |
|
| | <details> |
| | <summary>## Merge Detail</summary> |
| | <p> |
| | ### Models Merged |
| | |
| | The following models were included in the merge: |
| | * [TheDrummer/Llama-3SOME-8B-v2](https://huggingface.co/TheDrummer/Llama-3SOME-8B-v2) |
| | * [cgato/L3-TheSpice-8b-v0.8.3](https://huggingface.co/cgato/L3-TheSpice-8b-v0.8.3) |
| | * [Sao10K/L3-8B-Stheno-v3.2](https://huggingface.co/Sao10K/L3-8B-Stheno-v3.2) |
| | * [SicariusSicariiStuff/Wingless_Imp_8B](https://huggingface.co/SicariusSicariiStuff/Wingless_Imp_8B) |
| | * [deepseek-ai/DeepSeek-R1-Distill-Llama-8B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B) |
| |
|
| | ### Configuration |
| |
|
| | The following YAML configuration was used to produce this model: |
| |
|
| | ```yaml |
| | base_model: NeverSleep/Lumimaid-v0.2-8B |
| | merge_method: model_stock |
| | dtype: bfloat16 |
| | models: |
| | - model: cgato/L3-TheSpice-8b-v0.8.3 |
| | - model: Sao10K/L3-8B-Stheno-v3.2 |
| | - model: TheDrummer/Llama-3SOME-8B-v2 |
| | - model: SicariusSicariiStuff/Wingless_Imp_8B |
| | - model: deepseek-ai/DeepSeek-R1-Distill-Llama-8B |
| | ``` |
| |
|
| | </p> |
| | </details> |