File size: 1,312 Bytes
653bf59
df5be5b
 
 
 
 
653bf59
4230602
df5be5b
4230602
df5be5b
 
6794c0f
645e5bd
 
df5be5b
 
 
05b0c43
df5be5b
 
 
 
61e61fd
 
 
 
 
6794c0f
61e61fd
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
tags:
  - text-to-image
  - stable-diffusion
  - dress
  - try on
license: apache-2.0
language:
  - en
library_name: diffusers
---

# IMAGDressing: Interactive Modular Apparel Generation for Dressing

## IMAGDressing-v1: Customizable Virtual Dressing

<div align="center">

[**Project Page**](https://imagdressing.github.io/) **|** [**Paper**](https://arxiv.org/pdf/2407.12705) **|** [**Code**](https://github.com/muzishen/IMAGDressing)**|** [**Data**](https://huggingface.co/datasets/IMAGDressing/IGPair)

</div>

---

## Introduction

To address the need for flexible and controllable customizations in virtual try-on systems, we propose IMAGDressing-v1. Specifically, we introduce a garment UNet that captures semantic features from CLIP and texture features from VAE. Our hybrid attention module includes a frozen self-attention and a trainable cross-attention, integrating these features into a frozen denoising UNet to ensure user-controlled editing. We will release a comprehensive dataset, IGv1, with over 200,000 pairs of clothing and dressed images, and establish a standard data assembly pipeline. Furthermore, IMAGDressing-v1 can be combined with extensions like ControlNet, IP-Adapter, T2I-Adapter, and AnimateDiff to enhance diversity and controllability. 

![framework](assets/pipeline.png)