Skip to content

95-ag/dl-model-extraction

 
 

Repository files navigation

Model Extraction Attacks

door.py contains the main experiment code. The final project report is available here.

About

This project analyzes neural network model extraction techniques. It further investigates whether an extracted model can be used to conduct membership inference attacks and adversarial attacks on the original model.

Experimental Setup

  • Used victim models for CIFAR-10 (from zenodo.org) and custom-trained CIFAR-100 models for extraction attack analysis.
  • Varied attacker model architectures to test extraction effectiveness.
  • Applied extraction techniques to an out-of-distribution dataset, assembled from downsampled 32x32 ImageNet data. A mapping between ImageNet and CIFAR-10 classes was prepared (note: Deer and Horse classes were sourced online and downsampled).

Resources

References

Acknowledgments

About

Model extraction attack — exploratory implementation and analysis for learning purposes

Topics

Resources

License

Stars

Watchers

Forks

Languages

  • Jupyter Notebook 85.3%
  • Python 14.7%