Adversarial Input Detection Using Image
Processing Techniques (IPT)
Authors:
• Kishor Datta Gupta
• Dipankar Dasgupta
• Zahid Akhtar
Presented by :
Kishor Datta Gupta
Adversarial Attack (AA) on AI/ML
Types:
• Poisoning Attack : Manipulate
training data
• Evasion Attack: Manipulate
input data
• Trojan AI : Manipulate AI
Architecture (example:
Changes weights value)
“Manipulation of training
data, Machine Learning
(ML) model architecture,
or manipulate testing
data in a way that will
result in wrong output
from ML”
Different Adversarial Traits
Different Attack method has different types of noise/manipulation style
Detection of Adversarial Traits(1)
Clean and adversarial images have quantifiable noise difference
Detection of Adversarial Traits(2)
Detection of Adversarial Traits(3)
Methodology
Methodology(2)
Comparison
Summary
Work against adaptive attack
Don’t reduce ML efficiency
Can identify attack type
Applicable for cross-platform
Work for both blackbox and whitebox attack
Our defense has below properties:
Q/A

Adversarial Input Detection Using Image Processing Techniques (IPT)