Cheng-1 / README.md
marcuscedricridia's picture
Update README.md
b3fb741 verified
---
library_name: transformers
tags:
- mergekit
- merge
license: mit
---
# Cheng-1: Multi-Specialty Merged Language Model
## Model Overview
**Cheng-1** is a high-performance language model created through strategic merging of top-tier, pre-existing fine-tuned models. It excels in **coding, math, translation, and roleplay** without requiring additional fine-tuning. The final model was built using the **model_stock** method with a restore model to maintain strong instruction-following and mathematical abilities.
## Development Process
### 1. Foundation Model - "Yell-Qwen2.5-7B-1M"
- **Base Merge:** Combined `Qwen2.5-7B-Instruct-1M` with `Qwen2.5-7B` using **SCE merging**.
- **Purpose:** Established a strong general-purpose foundation for later merges.
#### **Merge Code:**
```yaml
merge_method: sce
models:
- model: Qwen/Qwen2.5-7B-Instruct-1M
- model: Qwen/Qwen2.5-7B
base_model: Qwen/Qwen2.5-7B-Instruct-1M
parameters:
select_topk: 1
dtype: bfloat16
tokenizer_source: base
normalize: true
int8_mask: true
name: Yell-Qwen2.5-7B-1M
```
### 2. Domain-Specific Merges
- **Coding:** Merged `AceCoder-Qwen2.5-7B-Ins-Rule` with Yell-Qwen2.5-7B-1M.
- **Translation:** Merged `DRT-7B` with Yell-Qwen2.5-7B-1M.
- **Math:** Merged `AceMath-7B-Instruct` with Yell-Qwen2.5-7B-1M.
- **Method:** All three were merged using **della merging**, producing three intermediate models.
#### **Merge Code:**
```yaml
merge_method: della
base_model: marcuscedricridia/Yell-Qwen2.5-7B-1M
models:
- model: TIGER-Lab/AceCoder-Qwen2.5-7B-Ins-Rule
parameters:
density: 1
weight: 1
lambda: 0.9
- model: Krystalan/DRT-7B
parameters:
density: 1
weight: 1
lambda: 0.9
- model: nvidia/AceMath-7B-Instruct
parameters:
density: 1
weight: 1
lambda: 0.9
parameters:
density: 1
weight: 1
lambda: 0.9
normalize: true
int8_mask: true
dtype: bfloat16
tokenizer_source: base
name: Cheng-1
```
### 3. Final Model Stock Merge
- **Models Combined:**
- `mergekit-della-wpunuct`
- `mergekit-della-phphmhr`
- `mergekit-della-qejrhsk`
- `Hush-Qwen2.5-7B-RP-v1.2-1M` (Roleplay model)
- **Base Model:** `YOYO-AI/Qwen2.5-7B-it-restore`
- **Final Method:** Used **model_stock merging** to integrate all models into Cheng-1.
#### **Merge Code:**
```yaml
merge_method: model_stock
base_model: YOYO-AI/Qwen2.5-7B-it-restore
models:
- model: marcuscedricridia/mergekit-della-wpunuct
- model: marcuscedricridia/mergekit-della-phphmhr
- model: marcuscedricridia/mergekit-della-qejrhsk
- model: marcuscedricridia/Hush-Qwen2.5-7B-RP-v1.2-1M
dtype: bfloat16
tokenizer_source: base
int8_mask: true
normalize: true
name: Cheng-1
```
## Benchmarks
```
Model: marcuscedricridia/Cheng-1
Precision: torch.bfloat16
Revision: cd8c9dd37c67c2e1b7c683fdd5e72b7f08c074b9
Average: 36.06
IFEval: 77.89
BBH: 36.54
MATH: 48.94
GPQA: 6.15
MUSR: 9.62
MMLU-PRO: 37.21
```
## Conclusion
Cheng-1 is a versatile model optimized for multiple domains. By merging top-performing models in coding, math, translation, and roleplay, it achieves balanced and strong benchmark results without direct fine-tuning.