Fast, lightweight, architecture-independent, low-level frequency control for diffusion-based Image-to-Image translations.
Filtered-Guided Diffusion for Controllable Image Generation
Zeqi Gu*,
Ethan Yang*,
Abe Davis
* denotes equal Contribution
GitHub | Paper | Project Page
We provide a lightweight implementation of FGD which contains all the core functionality described in our paper. Our code is based on the 🤗 diffusers library and taichi lang for efficient computation of the cross bilateral matrix.
A full explanation of how to use our code is described in the jupyter notebook demo.ipynb.
For questions regarding the code, or access to a more comprehensive set of fuctions (although much less user friendly) including experimental features, debugging, and evaluation, please contact both authors at zg45@cornell.edu and eey8@cornell.edu.
We provide a requirements.txt file which contains the packages our implementation of FGD was tested on. Note running our code requires a GPU.
diffusers==0.30.0
numpy==2.0.1
Pillow==10.4.0
pytorch_lightning==2.4.0
taichi==1.7.1
torch==2.4.0+cu118
tqdm==4.66.5
transformers==4.44.0
Note: jupyter notebook is also required in order to run our demo as we do not provide a command line interface.
For those wishing to use our work, please use the following citation:
@inproceedings{gu2024filter,
title={Filter-Guided Diffusion for Controllable Image Generation},
author={Gu, Zeqi and Yang, Ethan and Davis, Abe},
booktitle={ACM SIGGRAPH 2024 Conference Papers},
pages={1--10},
year={2024}
}