--- license: openrail++ library_name: diffusers tags: - text-to-image - stable-diffusion --- # Conceptrol: Concept Control of Zero-shot Personalized Image Generation ## Model Card This model implements Conceptrol, a training-free method that boosts zero-shot personalized image generation across Stable Diffusion, SDXL, and FLUX. It works without additional training, data, or models.

[Conceptrol: Concept Control of Zero-shot Personalized Image Generation](https://huggingface.co/papers/2503.06568) **Abstract:** Personalized image generation with text-to-image diffusion models generates unseen images based on reference image content. Zero-shot adapter methods such as IP-Adapter and OminiControl are especially interesting because they do not require test-time fine-tuning. However, they struggle to balance preserving personalized content and adherence to the text prompt. We identify a critical design flaw resulting in this performance gap: current adapters inadequately integrate personalization images with the textual descriptions. The generated images, therefore, replicate the personalized content rather than adhere to the text prompt instructions. Yet the base text-to-image has strong conceptual understanding capabilities that can be leveraged. We propose Conceptrol, a simple yet effective framework that enhances zero-shot adapters without adding computational overhead. Conceptrol constrains the attention of visual specification with a textual concept mask that improves subject-driven generation capabilities. It achieves as much as 89% improvement on personalization benchmarks over the vanilla IP-Adapter and can even outperform fine-tuning approaches such as Dreambooth LoRA. ## Quick Start #### 1. Environment Setup ``` bash conda create -n conceptrol python=3.10 conda activate conceptrol pip install -r requirements.txt ``` #### 2. Go to `demo_sd.ipynb` / `demo_sdxl.ipynb` / `demo_flux.py` for fun! ## Local Setup using Gradio #### 1. Start Gradio Interface ``` bash pip install gradio gradio gradio_src/app.py ``` #### 2. Use the GUI! ## Supporting Models | Model Name | Link | |-----------------------|-------------------------------------------------------------| | Stable Diffusion 1.5 | [stable-diffusion-v1-5/stable-diffusion-v1-5](https://huggingface.co/stable-diffusion-v1-5/stable-diffusion-v1-5) | | Realistic Vision V5.1 | [SG161222/Realistic_Vision_V5.1_noVAE](https://huggingface.co/SG161222/Realistic_Vision_V5.1_noVAE) | | Stable Diffusion XL-1024 | [stabilityai/stable-diffusion-xl-base-1.0](https://huggingface.co/stabilityai/stable-diffusion-xl-base-1.0) | | Animagine XL v4.0 | [cagliostrolab/animagine-xl-4.0](https://huggingface.co/cagliostrolab/animagine-xl-4.0)| | Realistic Vision XL V5.0 | [SG161222/RealVisXL_V5.0](https://huggingface.co/SG161222/RealVisXL_V5.0) | | FLUX-schnell | [black-forest-labs/FLUX.1-schnell](https://huggingface.co/black-forest-labs/FLUX.1-schnell) | | Adapter Name | Link | |-----------------------|-------------------------------------------------------------| | IP-Adapter | [h94/IP-Adapter](https://huggingface.co/h94/IP-Adapter/tree/main) | | OminiControl | [Yuanshi/OminiControl](https://huggingface.co/Yuanshi/OminiControl) | ## Source Code https://github.com/QY-H00/Conceptrol ## Citation ``` bibtex @article{he2025conceptrol, title={Conceptrol: Concept Control of Zero-shot Personalized Image Generation}, author={Qiyuan He and Angela Yao}, journal={arXiv preprint arXiv:2503.06568}, year={2025} } ``` ## Acknowledgement We thank the following repositories for their great work: [diffusers](https://github.com/huggingface/diffusers), [transformers](https://github.com/huggingface/transformers), [IP-Adapter](https://github.com/tencent-ailab/IP-Adapter), [OminiControl](https://github.com/Yuanshi9815/OminiControl)