Add paper abstract and BibTeX citation
#2
by
nielsr
HF Staff
- opened
README.md
CHANGED
@@ -1,17 +1,16 @@
|
|
1 |
---
|
|
|
|
|
|
|
|
|
2 |
license: other
|
3 |
license_name: community-license-agreement
|
4 |
license_link: LICENSE
|
5 |
-
language:
|
6 |
-
- en
|
7 |
-
base_model:
|
8 |
-
- black-forest-labs/FLUX.1-Fill-dev
|
9 |
pipeline_tag: image-to-image
|
10 |
tags:
|
11 |
- image customization
|
12 |
---
|
13 |
|
14 |
-
|
15 |
<div align="center">
|
16 |
<a href="https://github.com/TencentARC/IC-Custom">
|
17 |
<img src='https://github.com/TencentARC/IC-Custom/blob/main/assets/IC-Custom-logo.png?raw=true' width='120px'>
|
@@ -42,6 +41,9 @@ tags:
|
|
42 |
</a>
|
43 |
</div>
|
44 |
|
|
|
|
|
|
|
45 |
<p align="center">
|
46 |
IC-Custom is designed for diverse image customization scenarios, including:
|
47 |
</p>
|
@@ -52,7 +54,14 @@ tags:
|
|
52 |
- **Position-free**: Input a reference image and a target description to generate a new image with the reference image's ID
|
53 |
*Examples*: IP customization, character creation.
|
54 |
|
55 |
-
|
56 |
### Citation
|
57 |
|
58 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- black-forest-labs/FLUX.1-Fill-dev
|
4 |
+
language:
|
5 |
+
- en
|
6 |
license: other
|
7 |
license_name: community-license-agreement
|
8 |
license_link: LICENSE
|
|
|
|
|
|
|
|
|
9 |
pipeline_tag: image-to-image
|
10 |
tags:
|
11 |
- image customization
|
12 |
---
|
13 |
|
|
|
14 |
<div align="center">
|
15 |
<a href="https://github.com/TencentARC/IC-Custom">
|
16 |
<img src='https://github.com/TencentARC/IC-Custom/blob/main/assets/IC-Custom-logo.png?raw=true' width='120px'>
|
|
|
41 |
</a>
|
42 |
</div>
|
43 |
|
44 |
+
### Abstract
|
45 |
+
Image customization, a crucial technique for industrial media production, aims to generate content that is consistent with reference images. However, current approaches conventionally separate image customization into position-aware and position-free customization paradigms and lack a universal framework for diverse customization, limiting their applications across various scenarios. To overcome these limitations, we propose IC-Custom, a unified framework that seamlessly integrates position-aware and position-free image customization through in-context learning. IC-Custom concatenates reference images with target images to a polyptych, leveraging DiT's multi-modal attention mechanism for fine-grained token-level interactions. We introduce the In-context Multi-Modal Attention (ICMA) mechanism with learnable task-oriented register tokens and boundary-aware positional embeddings to enable the model to correctly handle different task types and distinguish various inputs in polyptych configurations. To bridge the data gap, we carefully curated a high-quality dataset of 12k identity-consistent samples with 8k from real-world sources and 4k from high-quality synthetic data, avoiding the overly glossy and over-saturated synthetic appearance. IC-Custom supports various industrial applications, including try-on, accessory placement, furniture arrangement, and creative IP customization. Extensive evaluations on our proposed ProductBench and the publicly available DreamBench demonstrate that IC-Custom significantly outperforms community workflows, closed-source models, and state-of-the-art open-source approaches. IC-Custom achieves approximately 73% higher human preference across identity consistency, harmonicity, and text alignment metrics, while training only 0.4% of the original model parameters.
|
46 |
+
|
47 |
<p align="center">
|
48 |
IC-Custom is designed for diverse image customization scenarios, including:
|
49 |
</p>
|
|
|
54 |
- **Position-free**: Input a reference image and a target description to generate a new image with the reference image's ID
|
55 |
*Examples*: IP customization, character creation.
|
56 |
|
|
|
57 |
### Citation
|
58 |
|
59 |
+
```bibtex
|
60 |
+
@article{li2025iccustom,
|
61 |
+
title={IC-Custom: Diverse Image Customization via In-Context Learning},
|
62 |
+
author={Li, Yaowei and Zhu, Yu and Wu, Xu and Liu, Bo and Li, Jia and Lu, Yong and Zhang, Song and Luo, Yujun},
|
63 |
+
journal={arXiv preprint arXiv:2507.01926},
|
64 |
+
year={2025},
|
65 |
+
url={https://arxiv.org/abs/2507.01926}
|
66 |
+
}
|
67 |
+
```
|