File size: 961 Bytes
6682ea4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
---
license: mit
library_name: diffusers
pipeline_tag: text-to-image
---

# LightGen: Efficient Image Generation through Knowledge Distillation and Direct Preference Optimization

<p align="center">
  <img src="https://github.com/XianfengWu01/LightGen/blob/main/demo/demo.png" width="720">
</p>

## About

This model (LightGen) introduces a novel pre-train pipeline for text-to-image models. It uses knowledge distillation (KD) and Direct Preference Optimization (DPO) to achieve efficient image generation. Drawing inspiration from data KD techniques, LightGen distills knowledge from state-of-the-art text-to-image models into a compact Masked Autoregressive (MAR) architecture with only $0.7B$ parameters.

It is based on [this paper](https://arxiv.org/abs/2503.08619), code release on [this github repo](https://github.com/XianfengWu01/LightGen).

Currently, we just release some checkpoint without DPO

## 🦉 ToDo List

- [ ] Release Complete Checkpoint.