ml-model-remember
ml-model-remember copied to clipboard
Code for Machine Learning Models that Remember Too Much (in CCS 2017)
Machine Learning Models that Remember Too Much
This repo contains an example for attacks in the paper Machine Learning that Remember Too Much (https://arxiv.org/pdf/1709.07886.pdf). The example is based on CIFAR10 dataset.
Train a malicious model
python train.py --attack ATTACK
Available ATTACK are cap (capacity abuse attack), cor (correlate value encoding attack) and sgn (sign encoding attack).
Test attack quality
python test_model --attack ATTACK