advex-uar icon indicating copy to clipboard operation
advex-uar copied to clipboard

Frank-Wolfe-L1 Attack

Open tanimutomo opened this issue 6 years ago • 3 comments

Hi, thanks for releasing the codes. When I used FrankWolfeAttack in advex-uar/advex_uar/attacks/fw_attack.py for evaluating the UAR for the model trained by PGD-Linf AT, the accuracy for all eps are lower than 10%. In addition to this evaluation problem, in the training with FrankWolfeAttack, the accuracy for the natural training data doesn't increase (stay in lower than 10%).

Could you tell me some advices for dealing with this attack method?

tanimutomo avatar Oct 08 '19 16:10 tanimutomo

Hi Tomoki, thanks for the interest in our work.

Could you tell us the exact model, command line parameters, and environment that you are using?

ddkang avatar Oct 09 '19 05:10 ddkang

Thanks for replying.

I re-implemented the training and testing code that calculates UAR scores heavily based on your code by myself.

So, the used codes for my experiment are not exactly the same as this repo.

The experimental results based on my training and testing codes almost achieved the same UAR score as your paper except for the FW-L1 Attack. The accuracies for the FW-L1 attack for various eps are around 3% - 10%. The details about the experiment are below.

In addition, when I trained the model (ResNet56) using the same settings as your paper, all accuracies (both training and validation set ) stayed in 10% during all epochs.

Are there specific techniques for dealing with FW-L1 Attack?

Thanks.

Codes

But, the following codes is exactly the same as your codes.

  • dataset
  • model (ResNet56) Other codes (e.g. Trainer, testing code) are implemented by myself.

Params

  • Dataset: CIFAR10
  • model : ResNet56
  • Epochs: 200
  • Optimizer: SGD (lr = 0.1, scheduler([100, 150], gamma=0.1))
  • Attack: PGD-Linf (eps = 32.0, step_size = eps / sqrt(num_of_steps) , num_of_steps = 10)

Environment

  • python 3.7
  • torch 1.1.0
  • torchvision 0.3.0

tanimutomo avatar Oct 09 '19 06:10 tanimutomo

There can be many subtle issues that occur with adversarial training and attacks. It's difficult to debug without seeing code.

Have you tried using our PGD-Linf trained model with our FW attack code?

ddkang avatar Oct 09 '19 15:10 ddkang