title | emoji | colorFrom | colorTo | sdk | sdk_version | app_file | pinned | license | models | ||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Artistic Portrait Generation |
🎨 |
yellow |
gray |
gradio |
5.22.0 |
app.py |
true |
apache-2.0 |
|
IP Adapter Art is a specialized version that uses a professional style encoder. Its goal is to achieve style control through reference images in the text-to-image diffusion model and solve the problems of instability and incomplete stylization of existing methods. This is a preprint version, and more models and training data coming soon.
can be used to conduct experiments directly.
For local experiments, please refer to a demo.
Local experiments require a basic torch environment and dependencies:
conda create -n artadapter python=3.10
conda activate artadapter
pip install -r requirements.txt
pip install git+https://github.com/openai/CLIP.git
pip install -e .
Evaluation using StyleBench style images. Image quality is evaluated using improved aesthetic predictor
CLIP Style Similarity | CSD Style Similarity | CLIP Text Alignment | Image Quality | Average | |
---|---|---|---|---|---|
DEADiff | 61.99 | 43.54 | 20.82 | 60.76 | 46.78 |
StyleShot | 63.01 | 52.40 | 18.93 | 55.54 | 47.47 |
Instant Style | 65.39 | 58.39 | 21.09 | 60.62 | 51.37 |
Art-Adapter(ours) | 67.03 | 65.02 | 20.25 | 62.23 | 53.63 |
We built an artistic portrait generation pipeline using Art-Adapter, PuLID, and ControlNet. The structure is shown in the figure below.
@misc{ipadapterart,
author = {Hao Ai, Xiaosai Zhang},
title = {IP Adapter Art},
year = {2024},
publisher = {GitHub},
journal = {GitHub repository},
howpublished = {\url{https://github.com/aihao2000/IP-Adapter-Art}}
}