Image-to-Image
English
custom_model
image customization

Add paper abstract and BibTeX citation

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +16 -7
README.md CHANGED
@@ -1,17 +1,16 @@
1
  ---
 
 
 
 
2
  license: other
3
  license_name: community-license-agreement
4
  license_link: LICENSE
5
- language:
6
- - en
7
- base_model:
8
- - black-forest-labs/FLUX.1-Fill-dev
9
  pipeline_tag: image-to-image
10
  tags:
11
  - image customization
12
  ---
13
 
14
-
15
  <div align="center">
16
  <a href="https://github.com/TencentARC/IC-Custom">
17
  <img src='https://github.com/TencentARC/IC-Custom/blob/main/assets/IC-Custom-logo.png?raw=true' width='120px'>
@@ -42,6 +41,9 @@ tags:
42
  </a>
43
  </div>
44
 
 
 
 
45
  <p align="center">
46
  IC-Custom is designed for diverse image customization scenarios, including:
47
  </p>
@@ -52,7 +54,14 @@ tags:
52
  - **Position-free**: Input a reference image and a target description to generate a new image with the reference image's ID
53
  *Examples*: IP customization, character creation.
54
 
55
-
56
  ### Citation
57
 
58
- If you find IC-Custom useful, please consider giving it a ⭐ on [GitHub](https://github.com/TencentARC/IC-Custom).
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - black-forest-labs/FLUX.1-Fill-dev
4
+ language:
5
+ - en
6
  license: other
7
  license_name: community-license-agreement
8
  license_link: LICENSE
 
 
 
 
9
  pipeline_tag: image-to-image
10
  tags:
11
  - image customization
12
  ---
13
 
 
14
  <div align="center">
15
  <a href="https://github.com/TencentARC/IC-Custom">
16
  <img src='https://github.com/TencentARC/IC-Custom/blob/main/assets/IC-Custom-logo.png?raw=true' width='120px'>
 
41
  </a>
42
  </div>
43
 
44
+ ### Abstract
45
+ Image customization, a crucial technique for industrial media production, aims to generate content that is consistent with reference images. However, current approaches conventionally separate image customization into position-aware and position-free customization paradigms and lack a universal framework for diverse customization, limiting their applications across various scenarios. To overcome these limitations, we propose IC-Custom, a unified framework that seamlessly integrates position-aware and position-free image customization through in-context learning. IC-Custom concatenates reference images with target images to a polyptych, leveraging DiT's multi-modal attention mechanism for fine-grained token-level interactions. We introduce the In-context Multi-Modal Attention (ICMA) mechanism with learnable task-oriented register tokens and boundary-aware positional embeddings to enable the model to correctly handle different task types and distinguish various inputs in polyptych configurations. To bridge the data gap, we carefully curated a high-quality dataset of 12k identity-consistent samples with 8k from real-world sources and 4k from high-quality synthetic data, avoiding the overly glossy and over-saturated synthetic appearance. IC-Custom supports various industrial applications, including try-on, accessory placement, furniture arrangement, and creative IP customization. Extensive evaluations on our proposed ProductBench and the publicly available DreamBench demonstrate that IC-Custom significantly outperforms community workflows, closed-source models, and state-of-the-art open-source approaches. IC-Custom achieves approximately 73% higher human preference across identity consistency, harmonicity, and text alignment metrics, while training only 0.4% of the original model parameters.
46
+
47
  <p align="center">
48
  IC-Custom is designed for diverse image customization scenarios, including:
49
  </p>
 
54
  - **Position-free**: Input a reference image and a target description to generate a new image with the reference image's ID
55
  *Examples*: IP customization, character creation.
56
 
 
57
  ### Citation
58
 
59
+ ```bibtex
60
+ @article{li2025iccustom,
61
+ title={IC-Custom: Diverse Image Customization via In-Context Learning},
62
+ author={Li, Yaowei and Zhu, Yu and Wu, Xu and Liu, Bo and Li, Jia and Lu, Yong and Zhang, Song and Luo, Yujun},
63
+ journal={arXiv preprint arXiv:2507.01926},
64
+ year={2025},
65
+ url={https://arxiv.org/abs/2507.01926}
66
+ }
67
+ ```