Biometric Presentation Attack
Detection: Beyond the Visible
Spectrum
TAKEN FROM IEEE TRANSACTIONS ON
INFORMATION FORENSICS AND SECURITY,
VOL 15, 2020
1
CONTENT
 INTRODUCTION
 DEFINATION
 RELATED WORKS
 PRESENTATION ATTACK DETECTION METHODOLOGY: HARDWARE
 PRESENTATION ATTACK DETECTION METHODOLOGY: SOFTWARE
 EXPERIMENTAL FRAMEWORK
 EXPERIMENTAL RESULTS
 CONCLUSION
 REFERENCE
2
INTRODUCTION
Biometric recognition systems is based on the individuals’ biological
(e.g., iris or fingerprint) or behavioural (e.g., signature or voice)
characteristics.
In spite of their numerous advantages, biometric systems are
vulnerable to external attacks as any other security-related
technology. The biometric capture device is probably the most
exposed one.
One can simply present the capture device with a presentation attack
instrument (PAI), such as a gummy finger or a fingerprint overlay, in
order to interfere with its intended behaviour. These attacks are
known as presentation attacks (PAs).
3
The initial approaches to PAD were based on the so-called
handcrafted features, such as texture descriptors or motion
analysis.
However, deep learning (DL) has become a thriving topic in the
biometric recognition in general.
More specifically, convolutional neural networks (CNNs) and deep
belief networks (DBNs) have been used for fingerprint PAD
purposes.
Selected SWIR wavelengths are used to discriminate skin from
other materials.
Human skin shows characteristic remission properties for
multispectral SWIR wavelengths, which are independent of a
person’s age, gender or skin type.
4
DEFINITION
 Bona fide Presentation: A normal or genuine presentation.
 Presentation attack (PA): An attack carried out on the capture device to
either conceal your identity or impersonate someone else.
 Presentation attack instrument (PAI): biometric characteristic or object
used in a presentation attack.
 Attack Presentation Classification Error Rate (APCER): Proportion of attack
presentations using the same PAI species incorrectly classified as bona fide
presentations in a specific scenario.
 Bona fide Presentation Classification Error Rate (BPCER):Proportion of bona
fide presentations incorrectly classified as presentation attacks in a
specific scenario.
5
RELATED WORKS
Key works on fingerprint PAD for both non-conventional sensors and conventional sensors.
Non-Conventional Fingerprint Sensor
6
Deep Learning for conventional sensors
7
PRESENTATION ATTACK DETECTION METHODOLOGY:
HARDWARE
 The finger SWIR capture device
contain a camera and SWIR sensor.
 Captures 64 × 64 px images, with a
25 mm fixed focal length lens
optimised for wavelengths within
900 – 1700 nm.
 Region of interest(ROI) is extracted
from background of size 18 × 58 px
8
9
Example of bonafides
and PAs acquired by
SWIR sensors
PRESENTATION ATTACK DETECTION METHODOLOGY:
SOFTWARE
Two approaches are: i) handcrafted features
ii) deep learning features
A) Handcrafted features
This method is builds upon the raw spectral signature(ss) of the pixels
across all four wavelengths in order to capture bona fide presentation and
PAI materials.
SWIR sensor provides raw spectral signature(ss) as:
𝒔𝒔 𝒙, 𝒚 = {i1(x,y), … … … iN(x,y)}
iN(x,y) represents the intensity value of pixels for n-th wavelength.
10
 However, this original representation of the sensor is vulnerable to
illumination changes.
 Only the differences among wavelengths will be used as our set of
handcrafted features. Therefore, for each pixel, the final normalised
difference feature vector d(x, y) is computed as follows:
d(x, y) ={d [ia,ib](x, y)} 1≤a<b≤N
 For each pixel with coordinates (x,y), the normalised difference feature
vector d(x,y) is used to classify as skin vs. non-skin with a Support
Vector Machine (SVM) classifier.
11
B) Deep Learning Features
1. Training CNN Models From Scratch:
 The first approach is focused on training residual CNNs from scratch.
 The characteristics of this network is the insertion of shortcut
connections every few stacked layers, converting the plain network into
its residual version.
 The residual connections allow the use of deeper neural network
architectures and at the same time decrease their training time
significantly.
12
2. Adapting Pre-Trained CNN Models
 The second approach evaluates the potential of state-of-the-art pre-trained
models for fingerprint PAD.
 Here the classifier is replaced and retrained and the weights of the last
convolutional layers is adapted.
 MobileNet and VGG19 network architectures pre-trained using the
ImageNet database are proposed to use.
 MobileNet network is based on depthwise separable convolutions, which
factorize a standard convolution into: i) a depthwise convolution, and ii) a
1×1 convolution called pointwise convolution.
 VGG19 network is one of the most popular network architecture, providing
very good results due to its simplicity.
13
EXPERIMENTAL FRAMEWORK
A. DATABASE
 Data were collected in two different stages and comprises both bona
fide and PA samples.
 For the bona fide samples, a total of 163 subjects participated during
the first stage. For each of them, all 5 fingers of the right hand were
captured. For the second stage, there were a total of 399 subjects.
 For the PA samples, there are a total of 35 different PAI species.
14
B. EXPERIMENTAL PROTOCOL
 The main goal behind the experimental protocol design is to
analyze and prove the soundness of our proposed fingerprint PAD
approach in a realistic scenario.
 For the development of our proposed fingerprint PAD methods,
both training and validation datasets are used in order to train the
weights of the systems and select the optimal network
architectures.
15
EXPERIMENTAL RESULTS16
 Spectral signature pixel-wise approach has achieved a 12.61% D-EER.
 It is not possible to get an APCER ≤ 2%, and for APCER ≈ 5%, the BPCER is
over 20% .
 For the case of training end-to-end residual CNN models from scratch, the
best result obtained is a 2.25% D-EER. This result outperforms the
handcrafted feature approach by an 82%.
 Low APCERs below 1% can be achieved for BPCERs below 8%, thereby
overcoming the main drawback of the handcrafted feature approach.
 Three CNN approaches have been fused in two by two basis.
 That yields convenient systems (i.e., low BPCER) even for highly secure (i.e.,
very low APCER) scenarios.
 Our proposed fingerprint PAD system has achieved a final 1.35% D-EER.
Furthermore, other operating points yield a BPCER of 2% for APCER ≤ 0.5%,
and an APCER ≈ 7% for BPCER=0.1%.
17
CONCLUSION
 A fingerprint PAD scheme based on i) a new capture device able to
acquire images within the short wave infrared (SWIR) spectrum, and
ii) state-of-the-art deep learning techniques.
 The best performance was reached for the fusion of two end-to-end
CNNs: the residual CNN trained from scratch and the adapted VGG19
pre-trained model. A D-EER of 1.35% was obtained.
 These results clearly outperform those achieved with the handcrafted
features, which yielded a D-EER over 12%.
 The use of SWIR images in combination with state-of-the-art CNNs
offers a reliable and efficient solution to the threat posed by presentation
attacks.
18
Reference
 J. Kolberg, M. Gomez-Barrero, S. Venkatesh, R. Raghavendra, and C. Busch,
“Presentation attack detection for finger recognition,” in Handbook of Vascular
Biometrics, S. Marcel, A. Uhl, R. Veldhuis, and C. Busch, Eds. Springer, 2019.
 E. Park, W. Kim, Q. Li, J. Kim, and H. Kim, “Fingerprint liveness detection using
CNN features of random sample patches,” in Proc. Int. Conf. Biometrics Special
Interest Group (BIOSIG), Sep. 2016.
 R. Tolosana, M. Gomez-Barrero, J. Kolberg, A. Morales, C. Busch, and J. Ortega,
“Towards fingerprint presentation attack detection based on convolutional
neural networks and short wave infrared imaging,” in Proc. Int. Conf.
Biometrics Special Interest Group (BIOSIG), Sep. 2018.
19
THANK YOU20

Biometric presentation attack detection

  • 1.
    Biometric Presentation Attack Detection:Beyond the Visible Spectrum TAKEN FROM IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, VOL 15, 2020 1
  • 2.
    CONTENT  INTRODUCTION  DEFINATION RELATED WORKS  PRESENTATION ATTACK DETECTION METHODOLOGY: HARDWARE  PRESENTATION ATTACK DETECTION METHODOLOGY: SOFTWARE  EXPERIMENTAL FRAMEWORK  EXPERIMENTAL RESULTS  CONCLUSION  REFERENCE 2
  • 3.
    INTRODUCTION Biometric recognition systemsis based on the individuals’ biological (e.g., iris or fingerprint) or behavioural (e.g., signature or voice) characteristics. In spite of their numerous advantages, biometric systems are vulnerable to external attacks as any other security-related technology. The biometric capture device is probably the most exposed one. One can simply present the capture device with a presentation attack instrument (PAI), such as a gummy finger or a fingerprint overlay, in order to interfere with its intended behaviour. These attacks are known as presentation attacks (PAs). 3
  • 4.
    The initial approachesto PAD were based on the so-called handcrafted features, such as texture descriptors or motion analysis. However, deep learning (DL) has become a thriving topic in the biometric recognition in general. More specifically, convolutional neural networks (CNNs) and deep belief networks (DBNs) have been used for fingerprint PAD purposes. Selected SWIR wavelengths are used to discriminate skin from other materials. Human skin shows characteristic remission properties for multispectral SWIR wavelengths, which are independent of a person’s age, gender or skin type. 4
  • 5.
    DEFINITION  Bona fidePresentation: A normal or genuine presentation.  Presentation attack (PA): An attack carried out on the capture device to either conceal your identity or impersonate someone else.  Presentation attack instrument (PAI): biometric characteristic or object used in a presentation attack.  Attack Presentation Classification Error Rate (APCER): Proportion of attack presentations using the same PAI species incorrectly classified as bona fide presentations in a specific scenario.  Bona fide Presentation Classification Error Rate (BPCER):Proportion of bona fide presentations incorrectly classified as presentation attacks in a specific scenario. 5
  • 6.
    RELATED WORKS Key workson fingerprint PAD for both non-conventional sensors and conventional sensors. Non-Conventional Fingerprint Sensor 6
  • 7.
    Deep Learning forconventional sensors 7
  • 8.
    PRESENTATION ATTACK DETECTIONMETHODOLOGY: HARDWARE  The finger SWIR capture device contain a camera and SWIR sensor.  Captures 64 × 64 px images, with a 25 mm fixed focal length lens optimised for wavelengths within 900 – 1700 nm.  Region of interest(ROI) is extracted from background of size 18 × 58 px 8
  • 9.
    9 Example of bonafides andPAs acquired by SWIR sensors
  • 10.
    PRESENTATION ATTACK DETECTIONMETHODOLOGY: SOFTWARE Two approaches are: i) handcrafted features ii) deep learning features A) Handcrafted features This method is builds upon the raw spectral signature(ss) of the pixels across all four wavelengths in order to capture bona fide presentation and PAI materials. SWIR sensor provides raw spectral signature(ss) as: 𝒔𝒔 𝒙, 𝒚 = {i1(x,y), … … … iN(x,y)} iN(x,y) represents the intensity value of pixels for n-th wavelength. 10
  • 11.
     However, thisoriginal representation of the sensor is vulnerable to illumination changes.  Only the differences among wavelengths will be used as our set of handcrafted features. Therefore, for each pixel, the final normalised difference feature vector d(x, y) is computed as follows: d(x, y) ={d [ia,ib](x, y)} 1≤a<b≤N  For each pixel with coordinates (x,y), the normalised difference feature vector d(x,y) is used to classify as skin vs. non-skin with a Support Vector Machine (SVM) classifier. 11
  • 12.
    B) Deep LearningFeatures 1. Training CNN Models From Scratch:  The first approach is focused on training residual CNNs from scratch.  The characteristics of this network is the insertion of shortcut connections every few stacked layers, converting the plain network into its residual version.  The residual connections allow the use of deeper neural network architectures and at the same time decrease their training time significantly. 12
  • 13.
    2. Adapting Pre-TrainedCNN Models  The second approach evaluates the potential of state-of-the-art pre-trained models for fingerprint PAD.  Here the classifier is replaced and retrained and the weights of the last convolutional layers is adapted.  MobileNet and VGG19 network architectures pre-trained using the ImageNet database are proposed to use.  MobileNet network is based on depthwise separable convolutions, which factorize a standard convolution into: i) a depthwise convolution, and ii) a 1×1 convolution called pointwise convolution.  VGG19 network is one of the most popular network architecture, providing very good results due to its simplicity. 13
  • 14.
    EXPERIMENTAL FRAMEWORK A. DATABASE Data were collected in two different stages and comprises both bona fide and PA samples.  For the bona fide samples, a total of 163 subjects participated during the first stage. For each of them, all 5 fingers of the right hand were captured. For the second stage, there were a total of 399 subjects.  For the PA samples, there are a total of 35 different PAI species. 14
  • 15.
    B. EXPERIMENTAL PROTOCOL The main goal behind the experimental protocol design is to analyze and prove the soundness of our proposed fingerprint PAD approach in a realistic scenario.  For the development of our proposed fingerprint PAD methods, both training and validation datasets are used in order to train the weights of the systems and select the optimal network architectures. 15
  • 16.
  • 17.
     Spectral signaturepixel-wise approach has achieved a 12.61% D-EER.  It is not possible to get an APCER ≤ 2%, and for APCER ≈ 5%, the BPCER is over 20% .  For the case of training end-to-end residual CNN models from scratch, the best result obtained is a 2.25% D-EER. This result outperforms the handcrafted feature approach by an 82%.  Low APCERs below 1% can be achieved for BPCERs below 8%, thereby overcoming the main drawback of the handcrafted feature approach.  Three CNN approaches have been fused in two by two basis.  That yields convenient systems (i.e., low BPCER) even for highly secure (i.e., very low APCER) scenarios.  Our proposed fingerprint PAD system has achieved a final 1.35% D-EER. Furthermore, other operating points yield a BPCER of 2% for APCER ≤ 0.5%, and an APCER ≈ 7% for BPCER=0.1%. 17
  • 18.
    CONCLUSION  A fingerprintPAD scheme based on i) a new capture device able to acquire images within the short wave infrared (SWIR) spectrum, and ii) state-of-the-art deep learning techniques.  The best performance was reached for the fusion of two end-to-end CNNs: the residual CNN trained from scratch and the adapted VGG19 pre-trained model. A D-EER of 1.35% was obtained.  These results clearly outperform those achieved with the handcrafted features, which yielded a D-EER over 12%.  The use of SWIR images in combination with state-of-the-art CNNs offers a reliable and efficient solution to the threat posed by presentation attacks. 18
  • 19.
    Reference  J. Kolberg,M. Gomez-Barrero, S. Venkatesh, R. Raghavendra, and C. Busch, “Presentation attack detection for finger recognition,” in Handbook of Vascular Biometrics, S. Marcel, A. Uhl, R. Veldhuis, and C. Busch, Eds. Springer, 2019.  E. Park, W. Kim, Q. Li, J. Kim, and H. Kim, “Fingerprint liveness detection using CNN features of random sample patches,” in Proc. Int. Conf. Biometrics Special Interest Group (BIOSIG), Sep. 2016.  R. Tolosana, M. Gomez-Barrero, J. Kolberg, A. Morales, C. Busch, and J. Ortega, “Towards fingerprint presentation attack detection based on convolutional neural networks and short wave infrared imaging,” in Proc. Int. Conf. Biometrics Special Interest Group (BIOSIG), Sep. 2018. 19
  • 20.

Editor's Notes

  • #4  no need to carry tokens or memorise passwords
  • #5 The "handcrafted features" were commonly used with "traditional" machine learning approaches for object recognition and computer vision like Support Vector Machines. understanding of how an object moves within an environment.
  • #7 1.Not many details about the acquired database or the experimental protocol are available. 2. the spectroscopic properties of living against the cadaver fingers were analyzed using four wavelengths. However, no PAIs were analysed in their work. 3. Finger vein image over same database was analyzed with help of gaussian pyramid and LBP.
  • #8 1.CNN optimization was used , best accuracy was achieved 2. 2 Pre trained CNN cafenet and googlenet, performance was compared with Siamese ntw. ROI: Different approach to optimize CNN model, output were classified by SVM, no data agumantation required Patch: deep boltzman machine which learn more complex samples. Instead of using deep n/ws 10 different handcrafted descriptors were used, which were fed to self developed deep ntw for final fusion.
  • #9  Hamamatsu G11097-0606S InGaAs area image sensor, 4 wavelength 1200nm ,1300nm 1450nm,1550nm Fingerprint verification can be done with 1.3 MP camera and VIS NIS lens.
  • #10 4 wavelength sample acquired from 2 bonafide and 3 PAIs Playdoh finger show some similarities wrt BP which makes PAD task harder. Change are different making it easier to discriminate from BP
  • #11 each individual score si generated by the individual PAD algorithms
  • #12 Due to having the finger slot open to outer world
  • #13 These ntw have outperformed the traditional ntw in many different datasets such as ImageNet for both image classification and object detection. Batch normalization is applied r8 after each convolution
  • #14  first layers of the CNN extract more general features related to directional edges and colours, whereas the last layers of the network are in charge of extracting more abstract features related to the specific task
  • #15 People from different gender, ethinicity,and age were considerd PAI species were categorized into 8 groups.
  • #17 1.for APCER ≤ 0.5%, the corresponding BPCER values for the fused systems (solid lines) are significantly lower than those of the individual networks (dashed lines): close to 2% 2.for low BPCER ≤1%, the best APCER (≤10%) is achieved for either the residual CNN alone (dashed orange) or its fusion with the VGG19 (solid dark blue).