- Mar-11-24: DiffuseMix paper is released arxiv link. 🔥🔥
Setup anaconda environment using environment.yml
file.
conda env create --name DiffuseMix --file=environment.yml
conda remove -n DiffuseMix --all # In case environment installation faileds
Below is the list of prompts, if your accuracy is low then you can use all prompts to increase the performance. Remember that each prompt takes a time to generate images, so the best way is to start from two prompts then increase the number of prompts.
prompts = ["Autumn", "snowy", "watercolor art","sunset", "rainbow", "aurora",
"mosaic", "ukiyo-e", "a sketch with crayon"]
train
└─── class 1
└───── n04355338_22023.jpg
└─── class 2
└───── n03786901_5410.jpg
└─── ...
To introduce the structural complexity, you can download fractal image dataset from here Fractal Dataset
`python3 main.py --train_dir PATH --fractal_dir PATH --prompts "sunset,Autumn"
If you find our work useful in your research please consider citing our paper:
@article{diffuseMix2024,
title={DIFFUSEMIX: Label-Preserving Data Augmentation with Diffusion Models},
author={Khawar Islam, Muhammad Zaigham Zaheer, Arif Mahmood, Karthik Nandakumar},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2024}
}
I am grateful to Adversarial-AutoMixup (@JinXins) for providing the source and target images, which significantly saved me a lot of time. Thank you once again. I am also exceptionally thankful to the author of IPMix, (@hzlsaber), for presenting their method's figures clearly, which greatly aided my paper. Additionally, their timely responses to my concerns saved me considerable time. Lastly, my thanks again go to the author of GuidedMixup, (@3neutronstar), for their insights on datasets and method outputs.