USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning
Abstract
USO, a unified model, achieves state-of-the-art performance in both style similarity and subject consistency by disentangling and re-composing content and style through a disentangled learning scheme and style reward-learning paradigm.
Existing literature typically treats style-driven and subject-driven generation as two disjoint tasks: the former prioritizes stylistic similarity, whereas the latter insists on subject consistency, resulting in an apparent antagonism. We argue that both objectives can be unified under a single framework because they ultimately concern the disentanglement and re-composition of content and style, a long-standing theme in style-driven research. To this end, we present USO, a Unified Style-Subject Optimized customization model. First, we construct a large-scale triplet dataset consisting of content images, style images, and their corresponding stylized content images. Second, we introduce a disentangled learning scheme that simultaneously aligns style features and disentangles content from style through two complementary objectives, style-alignment training and content-style disentanglement training. Third, we incorporate a style reward-learning paradigm denoted as SRL to further enhance the model's performance. Finally, we release USO-Bench, the first benchmark that jointly evaluates style similarity and subject fidelity across multiple metrics. Extensive experiments demonstrate that USO achieves state-of-the-art performance among open-source models along both dimensions of subject consistency and style similarity. Code and model: https://github.com/bytedance/USO
Community
๐ฅ๐ฅ We introduce USO, an open-sourced unified customization model supports freely combine any subjects with any styles in any scenarios, delivering outputs with high subject/identity consistency and strong style fidelity while ensuring natural, non-plastic portraits.
๐ code link: https://github.com/bytedance/USO
๐ project page: https://bytedance.github.io/USO/
๐ huggingface space: https://huggingface.co/spaces/bytedance-research/USO
๐ model checkpoint: https://huggingface.co/bytedance-research/USO
Open-sourced unified customization model
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- IC-Custom: Diverse Image Customization via In-Context Learning (2025)
- FreeCus: Free Lunch Subject-driven Customization in Diffusion Transformers (2025)
- StyDeco: Unsupervised Style Transfer with Distilling Priors and Semantic Decoupling (2025)
- CSD-VAR: Content-Style Decomposition in Visual Autoregressive Models (2025)
- DreamPoster: A Unified Framework for Image-Conditioned Generative Poster Design (2025)
- A Training-Free Style-Personalization via Scale-wise Autoregressive Model (2025)
- SPG: Style-Prompting Guidance for Style-Specific Content Creation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper