File size: 961 Bytes
6682ea4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 |
---
license: mit
library_name: diffusers
pipeline_tag: text-to-image
---
# LightGen: Efficient Image Generation through Knowledge Distillation and Direct Preference Optimization
<p align="center">
<img src="https://github.com/XianfengWu01/LightGen/blob/main/demo/demo.png" width="720">
</p>
## About
This model (LightGen) introduces a novel pre-train pipeline for text-to-image models. It uses knowledge distillation (KD) and Direct Preference Optimization (DPO) to achieve efficient image generation. Drawing inspiration from data KD techniques, LightGen distills knowledge from state-of-the-art text-to-image models into a compact Masked Autoregressive (MAR) architecture with only $0.7B$ parameters.
It is based on [this paper](https://arxiv.org/abs/2503.08619), code release on [this github repo](https://github.com/XianfengWu01/LightGen).
Currently, we just release some checkpoint without DPO
## 🦉 ToDo List
- [ ] Release Complete Checkpoint. |