Skip to content

Code for CCS 2025 Paper 'On the Feasibility of Poisoning Text-to-Image AI Models via Adversarial Mislabeling'

License

Notifications You must be signed in to change notification settings

stanleykywu/amp

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

AMP: Adversarial Mislabeling for Poisoning

This repository contains the main research code for our paper "On the Feasibility of Poisoning Text-to-Image AI Models via Adversarial Mislabeling" - S. Wu, R. Bhaskar, A. Ha, S. Shan, H. Zheng, B. Zhao (accepted to CCS 2025)

Overview

plot The figure above (Figure 1 in the paper) illustrates the full pipeline of our poisoning attack on text-to-image diffusion models. This repository contains code we used for parts A and B as followed:

  1. [Code for Part A.] Generating adversarial images that fool VLMs (./adversarial_mislabeling_attack)

    • We include our targeted white-box attack (section 5 in the paper) against all three VLMs we evaluated against (CogVLM, xGen-MM, and LLaVA)
    • Setup and usage can be found in its own README
  2. [Code for Part B.] Fine-tuning text-to-image models (./fine_tuning)

    • We include our fine-tuning scripts for all three text-to-image models we evaluated against (SD21, SDXL, and FLUX)
    • Setup and usage can be found in its own README
  3. Miscelaneous implementations of our project (./misc)

    • Evaluation metrics (./misc/metrics)
    • Concept selection from section 4.2 in the paper (./misc/concept_selection)
    • Setup and usage can be found in its own README

Citation

@inproceedings{wu2025amp,
  title={On the Feasibility of Poisoning Text-to-Image AI Models via Adversarial Mislabeling},
  author={Wu, Stanley and Bhaskar, Ronik and Ha, Anna Yoo Jeong and Shan, Shawn and Zheng, Haitao and Zhao, Ben Y},
  booktitle={ACM SIGSAC Conference on Computer and Communications Security},
  year={2025},
}

For any questions, please email stanleywu@cs.uchicago.edu.

About

Code for CCS 2025 Paper 'On the Feasibility of Poisoning Text-to-Image AI Models via Adversarial Mislabeling'

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published