adversarial robustness paper

Attack

black-box

white-box

Defense

Currently, the defenses against the adversarial attacks are being developed along three main directions: (for details,read this paper)

  1. Using modified training during learning or modified input during testing.
  2. Modifying networks, e.g. by adding more layers/subnetworks, changing loss/activation functions etc.
  3. Using external models as network add-on when classifying unseen examples.

Figure from "Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey"

Modified training / input

Modified networks

Adversarial Detecting

Network add-on

Analysis of Adversarial Examples

Model Compression And Adversarial Robustness

Others

blogs