http://pralab.diee.unica.it @biggiobattista 2
The iCub is the humanoid robot developed at the
Italian Institute of Technology as part of the EU
project RobotCub and adopted by more than 20
laboratories worldwide.
It has 53 motors that move the head, arms and hands,
waist, and legs. It can see and hear, it has the sense of
proprioception (body configuration)
and movement (using accelerometers and gyroscopes).
[http://www.icub.org]
The object recognition system of iCub uses visual features
extracted with CNN models trained on the ImageNet dataset
[G. Pasquale et al. MLIS 2015]
The iCub Humanoid
http://pralab.diee.unica.it @biggiobattista
Crafting the Adversarial Examples
• Key idea: shift the attack sample towards the decision boundary
– under a maximum input perturbation (Euclidean distance)
• Multiclass boundaries are obtained as the difference between
the competing classes (e.g., one-vs-all multiclass classification)
5
f1
f2
f3
f1-f3
http://pralab.diee.unica.it @biggiobattista
Error-generic Evasion
• Error-generic evasion
– k is the true class (blue)
– l is the competing (closest) class in feature space (red)
• The attack minimizes the objective to have the sample
misclassified as the closest class (could be any!)
6
1 0 1
1
0
1
Indiscriminate evasion
http://pralab.diee.unica.it @biggiobattista
Error-specific Evasion
• Error-specific evasion
– k is the target class (green)
– l is the competing class (initially, the blue class)
• The attack maximizes the objective to have the sample
misclassified as the target class
7
max
1 0 1
1
0
1
Targeted evasion
http://pralab.diee.unica.it @biggiobattista 10
Adversarial example generated
by manipulating only a
specific region, to simulate a
sticker that could be applied to
the real-world object
This image is classified as cup
The ‘Sticker’ Attack against iCub
http://pralab.diee.unica.it @biggiobattista
Why ML is Vulnerable to Evasion?
• Attack samples far from training data are anyway assigned to
‘legitimate’ classes
• Rejecting such blind-spot evasion points should improve security!
11
1 0 1
1
0
1
SVM-RBF (higher rejection rate)
1 0 1
1
0
1
SVM-RBF (no reject)
http://pralab.diee.unica.it @biggiobattista 12
Countering Adversarial Examples
maximum input perturbation (Euclidean distance)
visually-indistinguishable perturbations
Error-specific evasion (similar results for error-generic attacks)
http://pralab.diee.unica.it @biggiobattista
Conclusions and Future Work
• Adversarial Examples against iCub
• Countermeasure based on rejecting blind-spot evasion attacks
• Main open issue: instability of deep features
13
small changes in input space (pixels)
aligned with the gradient direction...
... correspond to large changes in
deep feature space!