/Fast-Gradient-Sign-Adversarial-Attack

Fast Gradient Sign Adversarial Attack(FGSM) examples creation using FashionMnist dataset

Primary LanguageJupyter Notebook

Fast-Gradient-Sign-Adversarial-Attack

Fast Gradient Sign Adversarial Attack(FGSM) examples creation using FashionMnist dataset.
inspiration: https://pytorch.org/tutorials/beginner/fgsm_tutorial.html

About Adversarial Attacks:

Fast Gradient Sign Attack:

General goal: Add the least amount of pertubations to the input that causes desired misclassification.

Assumptions on attacker's knowledge:

  1. White-Box: Attacker has full knowledge and access to the model, architecture, inputs, outputs and weights.
  2. Black-Box: Attacher has knowledge only about the inputs and outputs of the model and no information about the underlynig model architecture or weights.

Goals:

  1. misclassification: Attacker only wants the output classification to be wrong and does not care about what the new classification is.
  2. Source/Target misclassification: pertubations to the input that belongs to a specific source class so that it is classified as a specific target class.

FGSM : White-box attack with the goal of misclassification.

Fast Gradient Sign Attack: Use gradient of the loss w.r.t input data, then adjust the inputs to maximize the loss


In this repo we implement FGSM on the FashionMNIST dataset