StyleDrop
The code of StyleDrop-PyTorch has been moved to aim-uofa/StyleDrop-PyTorch. Please try it and have fun!
This is an unofficial PyTorch implementation of StyleDrop: Text-to-Image Generation in Any Style.
Unlike the parameters in the paper in (Round 1), we set d_prj=32
, is_shared=False
, which we found work better, these hyperparameters can be seen in configs/custom.py
.
we release them to facilitate community research.
News
- [07/11/2023] The code of StyleDrop-PyTorch has been moved to aim-uofa/StyleDrop-PyTorch. Please try it and have fun!
- [07/06/2023] Online Gradio Demo is available here
Todo List
- Release the code.
- Add gradio inference demo (runs in local).
- Add iterative training (Round 2).
Data & Weights Preparation
First, download VQGAN from this link (from MAGE, thanks!), and put the downloaded VQGAN in assets/vqgan_jax_strongaug.ckpt
.
Then, download the pre-trained checkpoints from this link to assets/ckpts
for evaluation or to continue training for more iterations.
finally, prepare empty_feature by runnig command python extract_empty_feature.py
And the final directory structure is as follows:
.
βββ assets
β βββ ckpts
β β βββ cc3m-285000.ckpt
β β β βββ lr_scheduler.pth
β β β βββ nnet_ema.pth
β β β βββ nnet.pth
β β β βββ optimizer.pth
β β β βββ step.pth
β β βββ imagenet256-450000.ckpt
β β βββ lr_scheduler.pth
β β βββ nnet_ema.pth
β β βββ nnet.pth
β β βββ optimizer.pth
β β βββ step.pth
β βββ fid_stats
β β βββ fid_stats_cc3m_val.npz
β β βββ fid_stats_imagenet256_guided_diffusion.npz
β βββ pipeline.png
| βββ contexts
β β βββ empty_context.npy
βββ βββ vqgan_jax_strongaug.ckpt
Dependencies
Same as MUSE-PyTorch.
conda install pytorch torchvision torchaudio cudatoolkit=11.3
pip install accelerate==0.12.0 absl-py ml_collections einops wandb ftfy==6.1.1 transformers==4.23.1 loguru webdataset==0.2.5 gradio
Train
All style data in the paper are placed in the data directory
- Modify
data/one_style.json
(It should be noted thatone_style.json
andstyle data
must be in the same directory), The format isfile_name:[object,style]
{"image_03_05.jpg":["A bear","in kid crayon drawing style"]}
- Training script as follows.
#!/bin/bash
unset EVAL_CKPT
unset ADAPTER
export OUTPUT_DIR="output_dir/for/this/experiment"
accelerate launch --num_processes 8 --mixed_precision fp16 train_t2i_custom_v2.py --config=configs/custom.py
Inference
The pretrained style_adapter weights can be downloaded from
#!/bin/bash
export EVAL_CKPT="assets/ckpts/cc3m-285000.ckpt"
export ADAPTER="path/to/your/style_adapter"
export OUTPUT_DIR="output/for/this/experiment"
accelerate launch --num_processes 8 --mixed_precision fp16 train_t2i_custom_v2.py --config=configs/custom.py
Gradio Demo
Put the style_adapter weights in ./style_adapter
folder and run the following command will launch the demo:
python gradio_demo.py
The demo is also hosted on HuggingFace.
Citation
@article{sohn2023styledrop,
title={StyleDrop: Text-to-Image Generation in Any Style},
author={Sohn, Kihyuk and Ruiz, Nataniel and Lee, Kimin and Chin, Daniel Castro and Blok, Irina and Chang, Huiwen and Barber, Jarred and Jiang, Lu and Entis, Glenn and Li, Yuanzhen and others},
journal={arXiv preprint arXiv:2306.00983},
year={2023}
}
Acknowlegment
- The implementation is based on MUSE-PyTorch
- Many thanks for the generous help from Zanlin Ni