File size: 468 Bytes
541bd3e c4b3afb 541bd3e c4b3afb 541bd3e c4b3afb 541bd3e 024c1c9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 |
---
license: apache-2.0
base_model:
- BAAI/Emu3-Gen
library_name: transformers
tags:
- merge
---
This is an interpolated upscale of [BAAI/Emu3-Gen](https://huggingface.co/BAAI/Emu3-Gen) from 8B to 11.5B.
For each layer in [7,8,9,10,11,12,13,14,15,16,22,23,24], the weights were lerp'd between the previous layer and the current and inserted between the two.
Expansion script is [here](https://huggingface.co/lodrick-the-lafted/Emu3-Gen-12B/blob/main/emu3_expand.py). |