Adversarial Input Detection Using Image Processing Techniques (IPT) Authors: • Kishor Datta Gupta • Dipankar Dasgupta • Zahid Akhtar Presented by : Kishor Datta Gupta
Adversarial Attack (AA) on AI/ML Types: • Poisoning Attack : Manipulate training data • Evasion Attack: Manipulate input data • Trojan AI : Manipulate AI Architecture (example: Changes weights value) “Manipulation of training data, Machine Learning (ML) model architecture, or manipulate testing data in a way that will result in wrong output from ML”
Different Adversarial Traits Different Attack method has different types of noise/manipulation style
Detection of Adversarial Traits(1) Clean and adversarial images have quantifiable noise difference
Detection of Adversarial Traits(2)
Detection of Adversarial Traits(3)
Methodology
Methodology(2)
Comparison
Summary Work against adaptive attack Don’t reduce ML efficiency Can identify attack type Applicable for cross-platform Work for both blackbox and whitebox attack Our defense has below properties:
Q/A

Adversarial Input Detection Using Image Processing Techniques (IPT)