Iris is a powerful tool for reliable human identification. It has the potential to identify individuals with a
high degree of assurance. Extracting good features is the most significant step in the iris recognition
system. In the past, different features have been used to implement iris recognition system. Most of them are
depend on hand-crafted features designed by biometrics specialists. Due to the success of deep learning in
computer vision problems, the features learned by the Convolutional Neural Network (CNN) have gained
much attention to be applied for iris recognition system. In this paper, we evaluate the extracted learned
features from a pre-trained Convolutional Neural Network (Alex-Net Model) followed by a multi-class
Support Vector Machine (SVM) algorithm to perform classification. The performance of the proposed
system is investigated when extracting features from the segmented iris image and from the normalized iris
image. The proposed iris recognition system is tested on four public datasets IITD, iris databases CASIAIris-V1,
CASIA-Iris-thousand and, CASIA-Iris- V3 Interval. The system achieved excellent results with the
very high accuracy rate.
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITIONijaia
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize persons with a high degree of assurance. Extracting effective features is the most important stage in the iris recognition system. Different features have been used to perform iris recognition system. A lot of them are based on hand-crafted features designed by biometrics experts. According to the achievement of deep learning in object recognition problems, the features learned by the Convolutional Neural Network (CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for features extracting and classification. The performance of the iris recognition system is tested on four public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The results show that the proposed system is achieved a very high accuracy rate.
With the technology development of medical industry, processing data is expanding rapidly and computation time
also increases due to many factors like 3D, 4D treatment planning, the increasing sophistication of MRI pulse
sequences and the growing complexity of algorithms. Graphics processing unit (GPU) addresses these problems
and gives the solutions for using their features such as, high computation throughput, high memory bandwidth,
support for floating-point arithmetic and low cost. Compute unified device architecture (CUDA) is a popular GPU
programming model introduced by NVIDIA for parallel computing. This review paper briefly discusses the need of
GPU CUDA computing in the medical image analysis. The GPU performances of existing algorithms are analyzed
and the computational gain is discussed. A few open issues, hardware configurations and optimization principles
of existing methods are discussed. This survey concludes the few optimization techniques with the medical imaging
algorithms on GPU. Finally, limitation and future scope of GPU programming are discussed.
SELF-LEARNING AI FRAMEWORK FOR SKIN LESION IMAGE SEGMENTATION AND CLASSIFICATIONijcsit
Image segmentation and classification are the two main fundamental steps in pattern recognition. To perform medical image segmentation or classification with deep learning models, it requires training on large image dataset with annotation. The dermoscopy images (ISIC archive) considered for this work does not have ground truth information for lesion segmentation. Performing manual labelling on this dataset is time-consuming. To overcome this issue, self-learning annotation scheme was proposed in the two-stage deep learning algorithm. The two-stage deep learning algorithm consists of U-Net segmentation model with the annotation scheme and CNN classifier model. The annotation scheme uses a K-means clustering algorithm along with merging conditions to achieve initial labelling information for training the U-Net model. The classifier models namely ResNet-50 and LeNet-5 were trained and tested on the image dataset without segmentation for comparison and with the U-Net segmentation for implementing the proposed self-learning Artificial Intelligence (AI) framework. The classification results of the proposed AI framework achieved training accuracy of 93.8% and testing accuracy of 82.42% when compared with the two classifier models directly trained on the input images.
Multilayer extreme learning machine for hand movement prediction based on ele...journalBEEI
Brain computer interface (BCI) technology connects humans with machines via electroencephalography (EEG). The mechanism of BCI is pattern recognition, which proceeds by feature extraction and classification. Various feature extraction and classification methods can differentiate human motor movements, especially those of the hand. Combinations of these methods can greatly improve the accuracy of the results. This article explores the performances of nine feature-extraction types computed by a multilayer extreme learning machine (ML-ELM). The proposed method was tested on different numbers of EEG channels and different ML-ELM structures. Moreover, the performance of ML-ELM was compared with those of ELM, Support Vector Machine and Naive Bayes in classifying real and imaginary hand movements in offline mode. The ML-ELM with discrete wavelet transform (DWT) as feature extraction outperformed the other classification methods with highest accuracy 0.98. So, the authors also found that the structures influenced the accuracy of ML-ELM for different task, feature extraction used and channel used.
Face Recognition Based Intelligent Door Control Systemijtsrd
This paper presents the intelligent door control system based on face detection and recognition. This system can avoid the need to control by persons with the use of keys, security cards, password or pattern to open the door. The main objective is to develop a simple and fast recognition system for personal identification and face recognition to provide the security system. Face is a complex multidimensional structure and needs good computing techniques for recognition. The system is composed of two main parts face recognition and automatic door access control. It needs to detect the face before recognizing the face of the person. In face detection step, Viola Jones face detection algorithm is applied to detect the human face. Face recognition is implemented by using the Principal Component Analysis PCA and Neural Network. Image processing toolbox which is in MATLAB 2013a is used for the recognition process in this research. The PIC microcontroller is used to automatic door access control system by programming MikroC language. The door is opened automatically for the known person according to the result of verification in the MATLAB. On the other hand, the door remains closed for the unknown person. San San Naing | Thiri Oo Kywe | Ni Ni San Hlaing ""Face Recognition Based Intelligent Door Control System"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-4 , June 2019, URL: https://www.ijtsrd.com/papers/ijtsrd23893.pdf
Paper URL: https://www.ijtsrd.com/engineering/electrical-engineering/23893/face-recognition-based-intelligent-door-control-system/san-san-naing
Neural Network based Vehicle Classification for Intelligent Traffic Controlijseajournal
Nowadays, number of vehicles has been increased and traditional systems of traffic controlling couldn’t be
able to meet the needs that cause to emergence of Intelligent Traffic Controlling Systems. They improve
controlling and urban management and increase confidence index in roads and highways. The goal of this
article is vehicles classification base on neural networks. In this research, it has been used a immovable
camera which is located in nearly close height of the road surface to detect and classify the vehicles. The
algorithm that used is included two general phases; at first, we are obtaining mobile vehicles in the traffic
situations by using some techniques included image processing and remove background of the images and
performing edge detection and morphology operations. In the second phase, vehicles near the camera are
selected and the specific features are processed and extracted. These features apply to the neural networks
as a vector so the outputs determine type of vehicle. This presented model is able to classify the vehicles in
three classes; heavy vehicles, light vehicles and motorcycles. Results demonstrate accuracy of the
algorithm and its highly functional level.
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITIONijaia
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize persons with a high degree of assurance. Extracting effective features is the most important stage in the iris recognition system. Different features have been used to perform iris recognition system. A lot of them are based on hand-crafted features designed by biometrics experts. According to the achievement of deep learning in object recognition problems, the features learned by the Convolutional Neural Network (CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for features extracting and classification. The performance of the iris recognition system is tested on four public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The results show that the proposed system is achieved a very high accuracy rate.
With the technology development of medical industry, processing data is expanding rapidly and computation time
also increases due to many factors like 3D, 4D treatment planning, the increasing sophistication of MRI pulse
sequences and the growing complexity of algorithms. Graphics processing unit (GPU) addresses these problems
and gives the solutions for using their features such as, high computation throughput, high memory bandwidth,
support for floating-point arithmetic and low cost. Compute unified device architecture (CUDA) is a popular GPU
programming model introduced by NVIDIA for parallel computing. This review paper briefly discusses the need of
GPU CUDA computing in the medical image analysis. The GPU performances of existing algorithms are analyzed
and the computational gain is discussed. A few open issues, hardware configurations and optimization principles
of existing methods are discussed. This survey concludes the few optimization techniques with the medical imaging
algorithms on GPU. Finally, limitation and future scope of GPU programming are discussed.
SELF-LEARNING AI FRAMEWORK FOR SKIN LESION IMAGE SEGMENTATION AND CLASSIFICATIONijcsit
Image segmentation and classification are the two main fundamental steps in pattern recognition. To perform medical image segmentation or classification with deep learning models, it requires training on large image dataset with annotation. The dermoscopy images (ISIC archive) considered for this work does not have ground truth information for lesion segmentation. Performing manual labelling on this dataset is time-consuming. To overcome this issue, self-learning annotation scheme was proposed in the two-stage deep learning algorithm. The two-stage deep learning algorithm consists of U-Net segmentation model with the annotation scheme and CNN classifier model. The annotation scheme uses a K-means clustering algorithm along with merging conditions to achieve initial labelling information for training the U-Net model. The classifier models namely ResNet-50 and LeNet-5 were trained and tested on the image dataset without segmentation for comparison and with the U-Net segmentation for implementing the proposed self-learning Artificial Intelligence (AI) framework. The classification results of the proposed AI framework achieved training accuracy of 93.8% and testing accuracy of 82.42% when compared with the two classifier models directly trained on the input images.
Multilayer extreme learning machine for hand movement prediction based on ele...journalBEEI
Brain computer interface (BCI) technology connects humans with machines via electroencephalography (EEG). The mechanism of BCI is pattern recognition, which proceeds by feature extraction and classification. Various feature extraction and classification methods can differentiate human motor movements, especially those of the hand. Combinations of these methods can greatly improve the accuracy of the results. This article explores the performances of nine feature-extraction types computed by a multilayer extreme learning machine (ML-ELM). The proposed method was tested on different numbers of EEG channels and different ML-ELM structures. Moreover, the performance of ML-ELM was compared with those of ELM, Support Vector Machine and Naive Bayes in classifying real and imaginary hand movements in offline mode. The ML-ELM with discrete wavelet transform (DWT) as feature extraction outperformed the other classification methods with highest accuracy 0.98. So, the authors also found that the structures influenced the accuracy of ML-ELM for different task, feature extraction used and channel used.
Face Recognition Based Intelligent Door Control Systemijtsrd
This paper presents the intelligent door control system based on face detection and recognition. This system can avoid the need to control by persons with the use of keys, security cards, password or pattern to open the door. The main objective is to develop a simple and fast recognition system for personal identification and face recognition to provide the security system. Face is a complex multidimensional structure and needs good computing techniques for recognition. The system is composed of two main parts face recognition and automatic door access control. It needs to detect the face before recognizing the face of the person. In face detection step, Viola Jones face detection algorithm is applied to detect the human face. Face recognition is implemented by using the Principal Component Analysis PCA and Neural Network. Image processing toolbox which is in MATLAB 2013a is used for the recognition process in this research. The PIC microcontroller is used to automatic door access control system by programming MikroC language. The door is opened automatically for the known person according to the result of verification in the MATLAB. On the other hand, the door remains closed for the unknown person. San San Naing | Thiri Oo Kywe | Ni Ni San Hlaing ""Face Recognition Based Intelligent Door Control System"" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-4 , June 2019, URL: https://www.ijtsrd.com/papers/ijtsrd23893.pdf
Paper URL: https://www.ijtsrd.com/engineering/electrical-engineering/23893/face-recognition-based-intelligent-door-control-system/san-san-naing
Neural Network based Vehicle Classification for Intelligent Traffic Controlijseajournal
Nowadays, number of vehicles has been increased and traditional systems of traffic controlling couldn’t be
able to meet the needs that cause to emergence of Intelligent Traffic Controlling Systems. They improve
controlling and urban management and increase confidence index in roads and highways. The goal of this
article is vehicles classification base on neural networks. In this research, it has been used a immovable
camera which is located in nearly close height of the road surface to detect and classify the vehicles. The
algorithm that used is included two general phases; at first, we are obtaining mobile vehicles in the traffic
situations by using some techniques included image processing and remove background of the images and
performing edge detection and morphology operations. In the second phase, vehicles near the camera are
selected and the specific features are processed and extracted. These features apply to the neural networks
as a vector so the outputs determine type of vehicle. This presented model is able to classify the vehicles in
three classes; heavy vehicles, light vehicles and motorcycles. Results demonstrate accuracy of the
algorithm and its highly functional level.
A Pattern Classification Based approach for Blur Classificationijeei-iaes
Blur type identification is one of the most crucial step of image restoration. In case of blind restoration of such images, it is generally assumed that the blur type is known prior to restoration of such images. However, it is not practical in real applications. So, blur type identification is extremely desirable before application of blind restoration technique to restore a blurred image. An approach to categorize blur in three classes namely motion, defocus, and combined blur is presented in this paper. Curvelet transform based energy features are utilized as features of blur patterns and a neural network is designed for classification. The simulation results show preciseness of proposed approach.
Microscopy images segmentation algorithm based on shearlet neural networkjournalBEEI
Microscopic images are becoming important and need to be studied to know the details and how-to quantitatively evaluate decellularization. Most of the existing research focuses on deep learning-based techniques that lack simplification for decellularization. A new computational method for the segmentation microscopy images based on the shearlet neural network (SNN) has been introduced. The proposal is to link the concept of shearlets transform and neural networks into a single unit. The method contains a feed-forward neural network and uses a single hidden layer. The activation functions are depending on the standard shearlet transform. The proposed SNN is a powerful technology for segmenting an electron microscopic image that is trained without relying on the pre-information of the data. The shearlet neural networks capture the features of full accuracy and contextual information, respectively. The expected value for specific inputs is estimated by learning the functional configuration of a network for the sequence of observed value. Experimental results on the segmentation of two-dimensional microscopy images are promising and confirm the benefits of the proposed approach. Lastly, we investigate on a challenging datasets ISBI 2012 that our method (SNN) achieves superior outcomes when compared to classical and deep learning-based methods.
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
Deep learning (DL) has achieved a significant performance in computer vision problems, mainly in automatic feature extraction and representation. However, it is not easy to determine the best pooling method in a different case study. For instance, experts can implement the best types of pooling in image processing cases, which might not be optimal for various tasks. Thus, it is
required to keep in line with the philosophy of DL. In dynamic neural network architecture, it is not practically possible to find
a proper pooling technique for the layers. It is the primary reason why various pooling cannot be applied in the dynamic and multidimensional dataset. To deal with the limitations, it needs to construct an optimal pooling method as a better option than max pooling and average pooling. Therefore, we introduce a dynamic pooling layer called RunPool to train the convolutional
neuralnetwork(CNN)architecture.RunPoolpoolingisproposedtoregularizetheneuralnetworkthatreplacesthedeterministic
pooling functions. In the final section, we test the proposed pooling layer to address classification problems with online social network (OSN) dataset
A presentation on Image Recognition, the basic definition and working of Image Recognition, Edge Detection, Neural Networks, use of Convolutional Neural Network in Image Recognition, Applications, Future Scope and Conclusion
An Image Based PCB Fault Detection and Its Classificationrahulmonikasharma
The field of electronics is skyrocketing like never before. The habitat for the electronic components is a printed circuit board (PCB). With the advent of newer and finer technologies it has almost become impossible to detect the faults in a printed circuit board manually which consumes lot of manpower and time. This paper proposes a simple and cost effective method of fault diagnosis in a PCB using image processing techniques. In addition to fault detection and its classification this paper addresses various problems faced during the pre-processing phase. This paper overcomes the drawbacks of the previous works such as improper orientations of the image and size variations of the image. Basically image subtraction algorithm is used for fault detection. The most commonly occurring faults are concentrated in this work and the same are implemented using MATLAB tool.
A Methodology for Extracting Standing Human Bodies from Single Imagesjournal ijrtem
Abstract: Extraction of the image of human body in unconstrained still images is challenging due to several factors, including shading, image noise, occlusions, background clutter, the high degree of human body deformability, and the unrestricted positions due to in and out of the image plane rotations. we propose a bottom-up approach for human body segmentation in static images. We decompose the problem into three sequential problems: Face detection, upper body extraction, and lower body extraction, since there is a direct pair wise correlation among them. Index Terms: Skin segmentation, Torso, Face recognition, Thresholding, Ethnicity, Morphology.
Object Recogniton Based on Undecimated Wavelet TransformIJCOAiir
Object Recognition (OR) is the mission of finding a specified object in an image or video sequence
in computer vision. An efficient method for recognizing object in an image based on Undecimated Wavelet
Transform (UWT) is proposed. In this system, the undecimated coefficients are used as features to recognize the
objects. The given original image is decomposed by using the UWT. All coefficients are taken as features for
the classification process. This method is applied to all the training images and the extracted features of
unknown object are used as an input to the K-Nearest Neighbor (K-NN) classifier to recognize the object. The
assessment of the system is agreed on using Columbia Object Image Library Dataset (COIL-100) database.
Optimized Neural Network for Classification of Multispectral ImagesIDES Editor
The proposed work involves the multiobjective PSO
based optimization of artificial neural network structure for
the classification of multispectral satellite images. The neural
network is used to classify each image pixel in various land
cove types like vegetations, waterways, man-made structures
and road network. It is per pixel supervised classification using
spectral bands (original feature space). Use of neural network
for classification requires selection of most discriminative
spectral bands and determination of optimal number of nodes
in hidden layer. We propose new methodology based on
multiobjective particle swarm optimization (MOPSO) to
determine discriminative spectral bands and the number of
hidden layer node simultaneously. The result obtained using
such optimized neural network is compared with that of
traditional classifiers like MLC and Euclidean classifier. The
performance of all classifiers is evaluated quantitatively using
Xie-Beni and â indexes. The result shows the superiority of
the proposed method.
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION gerogepatton
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize persons with a high degree of assurance. Extracting effective features is the most important stage in the iris recognition system. Different features have been used to perform iris recognition system. A lot of them are based on hand-crafted features designed by biometrics experts. According to the achievement of deep learning in object recognition problems, the features learned by the Convolutional Neural Network (CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for features extracting and classification. The performance of the iris recognition system is tested on four public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The results show that the proposed system is achieved a very high accuracy rate.
Transfer Learning with Convolutional Neural Networks for IRIS Recognitiongerogepatton
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize
persons with a high degree of assurance. Extracting effective features is the most important stage in the
iris recognition system. Different features have been used to perform iris recognition system. A lot of
them are based on hand-crafted features designed by biometrics experts. According to the achievement of
deep learning in object recognition problems, the features learned by the Convolutional Neural Network
(CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed
an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The
proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for
features extracting and classification. The performance of the iris recognition system is tested on four
public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The
results show that the proposed system is achieved a very high accuracy rate.
A Pattern Classification Based approach for Blur Classificationijeei-iaes
Blur type identification is one of the most crucial step of image restoration. In case of blind restoration of such images, it is generally assumed that the blur type is known prior to restoration of such images. However, it is not practical in real applications. So, blur type identification is extremely desirable before application of blind restoration technique to restore a blurred image. An approach to categorize blur in three classes namely motion, defocus, and combined blur is presented in this paper. Curvelet transform based energy features are utilized as features of blur patterns and a neural network is designed for classification. The simulation results show preciseness of proposed approach.
Microscopy images segmentation algorithm based on shearlet neural networkjournalBEEI
Microscopic images are becoming important and need to be studied to know the details and how-to quantitatively evaluate decellularization. Most of the existing research focuses on deep learning-based techniques that lack simplification for decellularization. A new computational method for the segmentation microscopy images based on the shearlet neural network (SNN) has been introduced. The proposal is to link the concept of shearlets transform and neural networks into a single unit. The method contains a feed-forward neural network and uses a single hidden layer. The activation functions are depending on the standard shearlet transform. The proposed SNN is a powerful technology for segmenting an electron microscopic image that is trained without relying on the pre-information of the data. The shearlet neural networks capture the features of full accuracy and contextual information, respectively. The expected value for specific inputs is estimated by learning the functional configuration of a network for the sequence of observed value. Experimental results on the segmentation of two-dimensional microscopy images are promising and confirm the benefits of the proposed approach. Lastly, we investigate on a challenging datasets ISBI 2012 that our method (SNN) achieves superior outcomes when compared to classical and deep learning-based methods.
RunPool: A Dynamic Pooling Layer for Convolution Neural NetworkPutra Wanda
Deep learning (DL) has achieved a significant performance in computer vision problems, mainly in automatic feature extraction and representation. However, it is not easy to determine the best pooling method in a different case study. For instance, experts can implement the best types of pooling in image processing cases, which might not be optimal for various tasks. Thus, it is
required to keep in line with the philosophy of DL. In dynamic neural network architecture, it is not practically possible to find
a proper pooling technique for the layers. It is the primary reason why various pooling cannot be applied in the dynamic and multidimensional dataset. To deal with the limitations, it needs to construct an optimal pooling method as a better option than max pooling and average pooling. Therefore, we introduce a dynamic pooling layer called RunPool to train the convolutional
neuralnetwork(CNN)architecture.RunPoolpoolingisproposedtoregularizetheneuralnetworkthatreplacesthedeterministic
pooling functions. In the final section, we test the proposed pooling layer to address classification problems with online social network (OSN) dataset
A presentation on Image Recognition, the basic definition and working of Image Recognition, Edge Detection, Neural Networks, use of Convolutional Neural Network in Image Recognition, Applications, Future Scope and Conclusion
An Image Based PCB Fault Detection and Its Classificationrahulmonikasharma
The field of electronics is skyrocketing like never before. The habitat for the electronic components is a printed circuit board (PCB). With the advent of newer and finer technologies it has almost become impossible to detect the faults in a printed circuit board manually which consumes lot of manpower and time. This paper proposes a simple and cost effective method of fault diagnosis in a PCB using image processing techniques. In addition to fault detection and its classification this paper addresses various problems faced during the pre-processing phase. This paper overcomes the drawbacks of the previous works such as improper orientations of the image and size variations of the image. Basically image subtraction algorithm is used for fault detection. The most commonly occurring faults are concentrated in this work and the same are implemented using MATLAB tool.
A Methodology for Extracting Standing Human Bodies from Single Imagesjournal ijrtem
Abstract: Extraction of the image of human body in unconstrained still images is challenging due to several factors, including shading, image noise, occlusions, background clutter, the high degree of human body deformability, and the unrestricted positions due to in and out of the image plane rotations. we propose a bottom-up approach for human body segmentation in static images. We decompose the problem into three sequential problems: Face detection, upper body extraction, and lower body extraction, since there is a direct pair wise correlation among them. Index Terms: Skin segmentation, Torso, Face recognition, Thresholding, Ethnicity, Morphology.
Object Recogniton Based on Undecimated Wavelet TransformIJCOAiir
Object Recognition (OR) is the mission of finding a specified object in an image or video sequence
in computer vision. An efficient method for recognizing object in an image based on Undecimated Wavelet
Transform (UWT) is proposed. In this system, the undecimated coefficients are used as features to recognize the
objects. The given original image is decomposed by using the UWT. All coefficients are taken as features for
the classification process. This method is applied to all the training images and the extracted features of
unknown object are used as an input to the K-Nearest Neighbor (K-NN) classifier to recognize the object. The
assessment of the system is agreed on using Columbia Object Image Library Dataset (COIL-100) database.
Optimized Neural Network for Classification of Multispectral ImagesIDES Editor
The proposed work involves the multiobjective PSO
based optimization of artificial neural network structure for
the classification of multispectral satellite images. The neural
network is used to classify each image pixel in various land
cove types like vegetations, waterways, man-made structures
and road network. It is per pixel supervised classification using
spectral bands (original feature space). Use of neural network
for classification requires selection of most discriminative
spectral bands and determination of optimal number of nodes
in hidden layer. We propose new methodology based on
multiobjective particle swarm optimization (MOPSO) to
determine discriminative spectral bands and the number of
hidden layer node simultaneously. The result obtained using
such optimized neural network is compared with that of
traditional classifiers like MLC and Euclidean classifier. The
performance of all classifiers is evaluated quantitatively using
Xie-Beni and â indexes. The result shows the superiority of
the proposed method.
TRANSFER LEARNING WITH CONVOLUTIONAL NEURAL NETWORKS FOR IRIS RECOGNITION gerogepatton
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize persons with a high degree of assurance. Extracting effective features is the most important stage in the iris recognition system. Different features have been used to perform iris recognition system. A lot of them are based on hand-crafted features designed by biometrics experts. According to the achievement of deep learning in object recognition problems, the features learned by the Convolutional Neural Network (CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for features extracting and classification. The performance of the iris recognition system is tested on four public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The results show that the proposed system is achieved a very high accuracy rate.
Transfer Learning with Convolutional Neural Networks for IRIS Recognitiongerogepatton
Iris is one of the common biometrics used for identity authentication. It has the potential to recognize
persons with a high degree of assurance. Extracting effective features is the most important stage in the
iris recognition system. Different features have been used to perform iris recognition system. A lot of
them are based on hand-crafted features designed by biometrics experts. According to the achievement of
deep learning in object recognition problems, the features learned by the Convolutional Neural Network
(CNN) have gained great attention to be used in the iris recognition system. In this paper, we proposed
an effective iris recognition system by using transfer learning with Convolutional Neural Networks. The
proposed system is implemented by fine-tuning a pre-trained convolutional neural network (VGG-16) for
features extracting and classification. The performance of the iris recognition system is tested on four
public databases IITD, iris databases CASIA-Iris-V1, CASIA-Iris-thousand and, CASIA-Iris-Interval. The
results show that the proposed system is achieved a very high accuracy rate.
Machine learning based augmented reality for improved learning application th...IJECEIAES
Detection of objects and their location in an image are important elements of current research in computer vision. In May 2020, Meta released its state-ofthe-art object-detection model based on a transformer architecture called detection transformer (DETR). There are several object-detection models such as region-based convolutional neural network (R-CNN), you only look once (YOLO) and single shot detectors (SSD), but none have used a transformer to accomplish this task. These models mentioned earlier, use all sorts of hyperparameters and layers. However, the advantages of using a transformer pattern make the architecture simple and easy to implement. In this paper, we determine the name of a chemical experiment through two steps: firstly, by building a DETR model, trained on a customized dataset, and then integrate it into an augmented reality mobile application. By detecting the objects used during the realization of an experiment, we can predict the name of the experiment using a multi-class classification approach. The combination of various computer vision techniques with augmented reality is indeed promising and offers a better user experience.
The automotive industry requires an automated system to sort different sizes and shapes
objects, images which are the mainly used component in the industry, to improve the overall
productivity. There are things at which humans are still way ahead of the machines in terms of
efficiency one of such thing is the recognition especially pattern recognition. There are several
methods which are tested for giving the machines the intelligence in efficient way for pattern
recognition purpose. The artificial neural network is one of the most optimization techniques used
for training the networks for efficient recognition. Computer vision is the science and technology of
machines that can see. The machine is made by integration of many parts to extract information from
an image in order to solve some task. Principle component analysis is a technique that will be
suitably used for the application purpose for sorting, inspection, fault diagnosis in various field.
A Survey of Convolutional Neural Network Architectures for Deep Learning via ...ijtsrd
Convolutional Neural Network CNN designs can successfully classify, predict and cluster in many artificial intelligence applications. In the health sector, intensive studies continue for disease classification. When the literature in this field is examined, it is seen that the studies are concentrated on the health sector. Thanks to these studies, doctors can make an accurate diagnosis by examining radiological images more consistently. In addition, doctors can save time to do other patient work by using CNN. In this study, related current manuscripts in the health sector were examined. The contributions of these publications to the literature were explained and evaluated. Complementary and contradictory arguments of the presented perspectives were revealed. It has been stated that the current status of the studies carried out and in which direction the future studies should evolve and that they can make an important contribution to the literature. Suggestions have been made for the guidance for future studies. Ahmet Özcan | Mahmut Ünver | Atilla Ergüzen "A Survey of Convolutional Neural Network Architectures for Deep Learning via Health Images" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-2 , February 2022, URL: https://www.ijtsrd.com/papers/ijtsrd49156.pdf Paper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/49156/a-survey-of-convolutional-neural-network-architectures-for-deep-learning-via-health-images/ahmet-özcan
I developed a Convolutional Neural Network using Python. This particular CNN is able to identify the correct individual based solely off of a photo with the knowledge of facial recognition.
Embracing GenAI - A Strategic ImperativePeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
Synthetic Fiber Construction in lab .pptxPavel ( NSTU)
Synthetic fiber production is a fascinating and complex field that blends chemistry, engineering, and environmental science. By understanding these aspects, students can gain a comprehensive view of synthetic fiber production, its impact on society and the environment, and the potential for future innovations. Synthetic fibers play a crucial role in modern society, impacting various aspects of daily life, industry, and the environment. ynthetic fibers are integral to modern life, offering a range of benefits from cost-effectiveness and versatility to innovative applications and performance characteristics. While they pose environmental challenges, ongoing research and development aim to create more sustainable and eco-friendly alternatives. Understanding the importance of synthetic fibers helps in appreciating their role in the economy, industry, and daily life, while also emphasizing the need for sustainable practices and innovation.
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
A Survey of Techniques for Maximizing LLM Performance.pptx
CONVOLUTIONAL NEURAL NETWORK BASED FEATURE EXTRACTION FOR IRIS RECOGNITION
1. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
DOI:10.5121/ijcsit.2018.10206 65
CONVOLUTIONAL NEURAL NETWORK BASED
FEATURE EXTRACTION FOR IRIS
RECOGNITION
Maram.G Alaslani1
and Lamiaa A. Elrefaei1,2
1
Computer Science Department, Faculty of Computing and Information Technology,
King Abdulaziz University, Jeddah, Saudi Arabia
2
Electrical Engineering Department, Faculty of Engineering at Shoubra, Benha
University, Cairo, Egypt
ABSTRACT
Iris is a powerful tool for reliable human identification. It has the potential to identify individuals with a
high degree of assurance. Extracting good features is the most significant step in the iris recognition
system. In the past, different features have been used to implement iris recognition system. Most of them are
depend on hand-crafted features designed by biometrics specialists. Due to the success of deep learning in
computer vision problems, the features learned by the Convolutional Neural Network (CNN) have gained
much attention to be applied for iris recognition system. In this paper, we evaluate the extracted learned
features from a pre-trained Convolutional Neural Network (Alex-Net Model) followed by a multi-class
Support Vector Machine (SVM) algorithm to perform classification. The performance of the proposed
system is investigated when extracting features from the segmented iris image and from the normalized iris
image. The proposed iris recognition system is tested on four public datasets IITD, iris databases CASIA-
Iris-V1, CASIA-Iris-thousand and, CASIA-Iris- V3 Interval. The system achieved excellent results with the
very high accuracy rate.
KEYWORDS
Biometrics, Iris, Recognition, Deep learning, Convolutional Neural Network (CNN), Feature extraction
(FE).
1. INTRODUCTION
In recent years, the concept of personal identity becomes critical, and the biometrics is a popular
way for authentication, which has been considered as the most secure and hardest way for
authentication purpose[1]. Biometric systems are developing technologies that can be used in
automatic systems for identifying individuals uniquely and effectively and it becomes a good
alternative to the traditional methods such as passwords. According to a study published in [1],
users choose to use smartphone biometrics as an alternate for passwords because it provides
additional safety for new technologies like Apple Pay [1].
Biometrics is an automated technique for authenticating an individual depends on a physical or
behavioral characteristic. Physical characteristics like fingerprints, voice, face, and iris. Behavior
characteristics are traits which can be learned or acquired such as, speaker verification, keystroke
dynamics and dynamic signature verification [2].
2. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
66
Among all unique physical features, iris biometrics is known as the most accurate and impossible
to reproduce or replicate [3]. The iris is a visible but protected structure. Also, iris does not
significantly change over time, so it is perfect and reliable for identity authentication [3, 4].
Iris recognition system has main stages. First, segmenting the required region of iris, after that it
is normalized to be a fixed pattern in polar coordinates. Then, the features extracted from that
pattern to be recognized at the last stage [5].
Extracting effective features is the major important stage in a lot of object recognition and
computer vision tasks. Therefore, several researchers have focused on designing robust features
for a variety of image classification tasks [6].
Nowadays, much attention is given to feature learning algorithms and Convolutional Neural
Networks (CCN). In this algorithm, the image is fed directly to the convolutional neural
networks, then the algorithm extracts the best features of this image [6, 7].
In this paper, we propose an iris recognition system where the features are extracted from the pre-
trained CNN Alex-Net model, and for the classification task, the multi-class Support Vector
Machine (SVM) is used. The performance of the proposed system is investigated when 1)
extracting features directly from the segmented iris image, and 2) extracting features from the
normalized iris pattern. The experimental study is tested on four public datasets collected under
different conditions IITD iris databases[8], CASIA-Iris-V1[9], CASIA-Iris-thousand [10], and
CASIA-Iris- V3 Interval [11].
The rest of this paper is organized as follows: Section 2, provides a brief background about the
basic idea of Convolutional Neural Networks. In section 3, some related works are presented.
Section 4, introduces a description of the proposed iris recognition system. The experimental
results and analysis are presented in section 5. Finally, the Conclusion is given in section 6.
2. CONVOLUTIONAL NEURAL NETWORK BACKGROUND
Convolutional Neural Networks (CNN) belong to a specific category of Neural networks
methods. CNN has not only been able to learn image feature representations automatically, but
they have also outperformed many conventional hand-crafted feature techniques [12].
Neural networks models have a hierarchical representation of data and depend on the computation
of layers that have a sequential implementation, the previous layers output will be the next layers
input. Every layer gives one representation level. And, there are a set of weights that
parameterized the layers. Also, the input units linking to output units through the weights in
addition to a group of biases [13].The weights in the Convolutional Neural Networks (CNN), are
shared locally, which means that each location of the input has the similar weights. The filter
form by the weights linked to the similar output [13].
A Convolutional Neural Network (CNN) comprises of alternating layers of locally connected
convolutional layers where every layer has the same number of filters. downsampling layers, and
the fully connected layers that work as a classifier [14]. Figure 1, shows the overall architecture
of a CNN.
Convolutional Neural Networks architecture has three concepts that make it effective: local
receptive fields, weights sharing, and downsampling operations [15]. The local receptive field
means every neuron accepts input from a small portion of the preceding layer. Also, it has the
same size of the convolution filter. Local receptive fields are used in convolutional and
3. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
67
downsampling layers. The weights sharing is applied to the convolutional layer to control the
capacity and to decrease the complexity of the model. Finally, the nonlinear downsampling which
used in the downsampling layers to decrease the spatial size of the image as well as decrease the
number of the free parameters of the model. These concepts help the CNN to be strong and
effective in recognition tasks [15]. In more detail, the Convolutional Neural Networks layers are:
The Convolutional Layer: the weights in this layer are made of a set of learnable filters produced
randomly and learned through the back-propagation algorithm. The feature map is the outcome of
every filter that convolved through the entire image. Also, the feature maps have the same
number of the applied filters in that layer [15].
As shown in Figure 1, the first convolutional layer containing 6 filters that produced 6 feature
maps which arranged together. Every feature map represents specific features from the image, for
example, represented points, or represented vertical edges [16]. The convolution operation is
described in (1).
Here j is the specific convolution feature map, Mj is a selection of input maps, kij is the filter, bj is
the feature map bias, l is the layer in the CNN, and f is the activation function [14]. The ReLU is
the common activation function which used to add non-linearity to the network [15].
The Pooling Layer: which implements a downsampling operation to decrease the spatial size of
the convolutional layers. First, the size of pooling mask and pooling operation type must be
determined and after that applied at the pooling layer [17].
The pooling operation implemented on the pixel values captured by the pooling mask, multiply it
by a trainable coefficient, after that added to a trainable bias [14]. The pooling operation is
described in (2).
Where is the result of the downsampling operation applied on the jth
region in the input, is
the jth
region of interest captured by the pooling mask in the previous layer, pool is the specific
operation done on the region (max or average), is a trainable coefficient, is a trainable bias,
and f is an activation function [18]. The max pooling is the most common pooling operation
which is used in this paper.
Fully connected layers: which used the extracted features in the preceding layers to do the
classification task [19]. The result of the last convolutional or pooling layer is fed to the fully
connected layers like in an original neural network [15].
Figure 1. An illustration of the Convolutional Neural Networks architecture. The gray squares mention the
feature maps and the green squares mention the convolution filter. The cross-lines among the last two
layers mention the fully connected neurons[15].
4. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
68
3. RELATED WORK
Several works have been addressed the use of the Convolutional Neural Network (CNN) for
biometric recognition. Some of these works are presented here.
The authors in [5] proposed an iris recognition system where they used the pertained VGG-Net to
extract the deep features. Then they used a multi-class SVM algorithm for classification. They
tested their system on two iris databases, IIT iris dataset, and CASIA 1000 Iris dataset. Their
experiment achieved high results with the accuracy recognition of 99.4%.
The authors in [20] proposed a system for face recognition with strong 4-layer CNN architecture;
the proposed system can handle facial images which have facial expressions, poses, occlusions
and changing illumination. The results of the experiment showed a high accuracy rate of 99.5%
on AR database. Their experiment on the 35-individuals from FERET database shown
recognition accuracy of 85.13%.
The authors in [14] proposed a system for the diagnosis of iris nevus depending on a
Convolutional Neural Network in addition to deep belief network. Iris nevus is defined as a
pigmented growth located around the pupil or in the front of the eye. They used a pertained
LeNet-5 architecture for CNN. Their accuracy rates are 93.35% for CNN and 93.67% for the
deep belief network.
The authors in [7] proposed an algorithm which uses a convolutional network named scattering
transform/network which delivers a multi-layer representation of the signal and is invariant to
translation, small deformation, and rotation. After extraction of scattering features, they used the
Principal component analysis to decrease the data dimensionality and then recognition is
performed using a multi-class support vector machine. Their proposed algorithm has been tested
on three face datasets and gives a very high recognition rate.
The authors in [21] proposed a MATLAB-based finger-vein recognition system based on CNN
with Graphical User Interface as the user input. To retrain the network for new incoming subjects,
they used two layers of CNN out of the proposed four-layer CNN. The pre-processing stages for
finger-vein images and CNN design have been conducted on different platforms. They tested the
system on 50 images that are developed in-house. The experiment achieved an accuracy rate of
96%.
Table 1. summarize the related work presented in this section.
4. THE PROPOSED IRIS RECOGNITION SYSTEM
The proposed iris recognition system using the Convolutional Neural Network (Alex Net) for
feature extraction is shown in Figure 2, The development of the proposed iris recognition system
is discussed in three parts: the pre-processing stage, feature extraction stage, and classification
stage.
4.1. The preprocessing stage
In this stage, iris segmentation and normalization are performed. In the proposed system, the
primary goal of selecting the iris region as an input to Convolutional Neural Network (Alex-net)
model as an alternative to the entire eye image, as proposed in [15], is to decrease the
computational complexity of the model. An additional goal is to avoid the decline of matching
performance as well as the extraction of feature causing by eyelids and eyelashes appearance.
5. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
69
For the iris segmentation process, the circular Hough transform is applied for detecting the
boundaries of the iris and the pupil which involve first perform Canny Edge detection to generate
an edge map [22]. The Canny Edge detecting contains five stages: Smoothing, Finding,
Gradients, Non-maximum suppression, Double thresholding, Edge tracking through hysteresis
[23]. Algorithm 1 lists all the required steps for iris segmentation.
For iris normalization process, the rubber sheet model is used. As shown in Figure 3, In this
model every pixel within the localized iris area is remapped from Cartesian coordinate (x,y) to
polar coordinate (r,θ) where r is on the interval [0,1], and θ is angle [0,2π]. The representation of
mapping the iris region to the normalized polar coordinates can be modeled as in (3) [15, 24].
Where I(x,y) is the iris area, (x,y) is the original Cartesian coordinates, (r,θ) is the corresponding
normalized polar coordinates xp, yp and xl, yl are the coordinates of the pupil and iris boundaries
along the θ direction [15, 24].
The results of the segmentation and the normalization steps are shown in Figure 4.
Table 1. Summary of Using Convolutional Neural Network for Biometric Recognition
Reference#
(Year)
Biometric
CNN
Model
Classification Dataset Accuracy
[5]
(2016)
iris VGG-Net
multi-class
SVM
IIT iris dataset, and
CASIA 1000 Iris dataset
99.4% on IIT
iris dataset
90% on CASIA
1000 Iris dataset
[20]
(2014)
face LeNet-5
classification
depends on the
winner-takes-
all rule
AR database and FERET
database
99.5% on AR
database
85.13% on
FERET
database
[14]
(2017)
iris LeNet-5
multi-class
SVM
A free database (The
Eye Cancer Foundation,
Eye Cancer Network)
93.35%
[7]
(2016)
face
scattering
transform/
network
multi-class
SVM
Yale Face Database
Georgia Tech Face
Database
Extended Yale Face
Database
93.1%
[21]
(2016)
Finger-
vein
LeNet-5
classification
depends on the
winner-takes-
all rule
The database collected
in-house using 6
different fingers. There
are 60 participants from
the Universiti Teknikal
Malaysia Melaka. The
total is 600 samples.
96%
6. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
70
Figure 2. The proposed iris recognition system using Convolutional Neural Network (Alex Net) for feature
extraction
Figure 3. Normalized iris region using the rubber sheet model [15].
(a) (b) (c) (d)
Figure 4. The result of segmentation and normalization stage for (Image 01-L) from IIT Iris Database (a)
Original input image. (b) Detected iris and pupil boundary. (c) segmented iris image. (d) Normalized iris
image.
Algorithm 1: Automatic iris region segmentation from an eye image.
//Input: The input eye image
//Output: center and radius of iris circle, and center and radius of pupil circle
1: Define the range of pupil and iris radius, manually set the range of radius, according to the
database used
2: Find the iris boundaries
• Perform canny edge detection to generate an edge map.
a) Apply Gaussian filter
b) Apply gamma function
c) Apply non-maximum suppression
d) Apply hysteresis thresholding
• Apply circular Hough transform
• Draw circles of different radius for all edge point.
• Find the maximum in the Hough space, and it will be the parameters of the circle
3: Find pupil boundary using the same steps but just using the region within the previously detected
iris boundary
4: Return the center and radius of iris and pupil circle
4.2. The feature extraction stage
The pre-trained Convolutional Neural Network model (Alex-Net) is used for feature extraction
process. This model was designed by the Super Vision group[25]. Alex-Net is a scaled version of
the conventional Le-Net [12].
7. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
71
Alex-Net is trained on the Image-Net Large-Scale Visual Recognition Challenge (ILSVRC).
Alex-Net is trained to classify the 1.2 million images in the image-Net database into 1000
different classes. Figure 5 shows the overall architecture of Alex-Net model [25]. It contains set
of layers; the input layer is the first layer which defines the input dimensions. The Alex-Net
model needs the input image to be 227-by-227-by-3. The middle layers make up the bulk of the
AlexNet. These layers consist of series of five convolutional layers, followed by rectified linear
units (ReLU) and max-pooling layers. Next to these layers, three fully-connected layers. The
classification layer is the final layer [25].
The first convolutional layers perform 11x11 convolutions with stride 4 and with no padding, the
second convolutional layers perform 5x5 convolutions with stride 1 and pad 2, The other
convolutional layers perform 3x3 convolutions with stride 1 and pad 1, and 2x2 pooling (with no
padding). The detailed explanation of Alex-Net layers is shown in Table 2.
Figure 5. The architecture of Alex-Net model [25]
Table 2. The Detailed Explanation of The Alex-Net Layer
Type of Layer No. of
Filter
Feature Map Size Kernel
Size
No. of
Stride
No. of Padding
Image input layer 227 x 227 x3
(height x width x
channel)
(1st convolutional layer)
Relu-1
Cross-channel normalization
Max pooling1
96
1
55 x 55 x 96
27x27x96
11x11
3x3
4x4
2x2
0x0
0x0
(2nd convolutional layer)
Relu-2
Cross-channel normalization
Max pooling2
256
1
27 x 27 x 256
13x13x256
5x5
3x3
1x1
2x2
2x2
0x0
(3rd convolutional layer)
Relu-3
384 13 x 13 x 384 3x3 1x1 1x1
(4th convolutional layer)
Relu-4
384 13 x 13 x 384 3x3 1x1 1x1
(5th convolutional layer)
Relu-5
Max pooling5
256
1
13 x 13 x 256
6x6x256
3x3
3x3
1x1
2x2
1x1
0x0
Fully connected layer-6(fc6)
Relu-6
4096 x1
4096 x1
Fully connected layer-7 (fc7)
Relu-7
4096 x1
4096 x1
Fully connected layer-8 (fc8)
Softmax layer
Output layer
1000x1
1000 class
8. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
72
4.3. The classification stage
The classifier is needed after feature extraction to find the corresponding label for every test
image. Different types of classifiers can be used for this task, for example, Support Vector
Machine, Softmax Regression, and Neural Network [26]. In this work, a multiclass Support
Vector Machine classifier has been used. A brief description of multiclass SVM is as follows:
Assuming we have the set of training data (x1,y1),(x2,y2),…(xn,yn) and we want to classify the set
into two classes where xi ∈ Rd
is the feature vector and yi ∈ {-1, +1} is the label class. The two
classes are linearly separable with a hyperplane w.x+b=0. With no other previous knowledge
about the data, SVM can find the optimal hyperplane as the one with the maximum margin
(which results in the minimum expected generalization error) [27].
The multi-class SVM can be implemented for a set of data with M classes, we can train M binary
classifiers that can distinguish each class against all other classes, then select the class that
classifies the test sample with the greatest margin (one-vs-all) [27].
Algorithm 2 lists all the required steps for the feature extraction and the classification stages.
Algorithm 2: Extracting feature using a Pre-trained CNN (Alex-Net) and classify the feature using the
SVM algorithm.
//Input: The input images
//Output: The recognition accuracy
1: Load input images and its labels
2: Split each category into the similar number of images
3: Load pre-trained CNN (Alex Net model)
4: Pre-process images For Alex Net model
5: Split the sets of the images into training and testing data.
6: Extract Features from the deeper layers of Alex-Net model.
7: Get training labels from the training set
8: Use the training features to train a multiclass SVM classifier
9: Extract features from test set
10: Use the trained classifier to predict the label for test set
11: Get the known labels for test set
12: Tabulate the results by a confusion matrix.
13: Convert confusion matrix into percentage form
14: Display the mean accuracy
5. EXPERIMENTAL RESULTS AND ANALYSIS
The proposed system is tested using four popular iris datasets: IITD iris databases [8], CASIA-
Iris-V1 [9], CASIA-Iris-thousand [10], and CASIA-Iris-Interval [11]. The iris images are
captured in these databases under different situations of pupil dilation, eyelids/eyelashes
occlusion, slight shadow of eyelids, specular reflection, etc.
The proposed system performance is evaluated according to the accuracy of the recognition rate.
The accuracy is the fraction of labels that the classifier predicts correctly. We tested the system on
60 subjects in each dataset. The specifications of these subjects for the four datasets are
summarized in Table 3.
9. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
73
Table 3. The specifications of the used datasets
dataset IITD CASIA-Iris-V1
CASIA-Iris-
Thousand
CASIA-Iris-
Interval
Number of
subjects
60 60 60 60
Samples per
subject
10 7 10 10
Number of images 600 420 600 600
Image size (pixels) (320 × 240) (320 × 280) (640 × 480) (320 × 280)
Image format BMP BMP JPEG JPEG
The experiments have been conducted on these datasets in two cases, the first case: using the iris
image after segmentation. In this case we resize the image to be 227-by-227 as required for the
for Alex-Net input.
The second case: using the iris image after normalization. In the normalization stage, the
normalization parameters were set manually, the radial resolution is set to 227 and the angular
resolution also is 227. So, the size of the image that comes from normalization stage is 227-by-
227, which is suitable for Alex-Net input.
The sets of the segmented and normalized images are split randomly into training and test data.
For each person 80% of images are used for the training and 20% of images are used for the
testing.
The features extracted from one of the deeper layers of Alex-Net named 'fc6' using the
"activations" method. The output of (fc6) layer is a 4096-dimensional vector. The 'MiniBatchSize'
is set to 22. The mini-batch is defined as a subset of the training data which used to assess the
gradient of the loss function in addition to updating the weights[25]. To speed-up the multiclass
SVM, the activations output is arranged as columns.
The system and its stages are implemented using MATLAB 2017 on a laptop with Core i7 CPU
running at 2.8GHz.
The learned filters on the first layer of the used convolutional neural network (Alex-Net) are
shown in Figure 6, It contains mostly edges and colors, which indicates that the filters at layer
'conv1' are edge detectors and color filters. The edge detectors are at different angles, which
allows the network to construct more complex features in the later layers. Moreover, the filters
learned on (fc6) layer is shown in Figure 7, This layer is towards the end of the network and
learns high-level combinations of the features learned by the earlier layers.
Figure 6. Filters learned on the first layer of CNN (Alex-Net) on IIT Iris Database
10. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
74
Figure 7. Filters learned on (fc6) layer of CNN (Alex-Net) on IIT Iris Database
As stated earlier, different layers encode different levels of visual content. To investigate the
performance due to each layer in the used CNN (Alex-Net), the recognition accuracy is estimated
after using the output from each layer as a feature vector to represent the iris. The recognition
accuracy is illustrated in Figure 8 and Figure 9 when the features are extracted from the iris
segmented image and from the iris normalized image respectively.
Figure 8. The recognition accuracy after segmentation stage versus different Alex-Net layer
As it can be seen from Figure 8. and Figure 9., in both cases of using the iris segmented image
and the iris normalized image, the features from the (fc6) layer have the highest recognition
accuracy and after that, the recognition accuracy drops. That because the higher layers of the
Alex-Net model, maybe cannot distinguish much between diverse iris patterns because they
capture only the abstract and high-level information, while the mid-level features in the fc6 layer
have additional distinguished power for same-class recognition.
The proposed system recognition accuracy and the required time to extract the features for each
iris image for the four databases are shown in Table 4. The recognition accuracy after the
segmentation stage is better than the recognition accuracy after the normalization stage that
because Alex-Net architecture can capture the discriminative visual features in the segmented iris
image better than the normalize iris images. but the time for extracting features after the
normalization stage is less than the time for extracting features after the segmentation stage.
11. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
75
Figure 9. The recognition accuracy after normalization stage versus different Alex-Net layer
Table 4. The recognition accuracy of the adopted iris image databases.
Database
After Segmentation After Normalization
Recognition
Accuracy
Time (S)
Recognition
Accuracy
Time (S)
IIT Delhi Iris Database 100% 0.06 98.33% 0.015
CASIA-Iris-V1 98% 0.08 85% 0.02
CASIA-Iris-Thousand 98% 0.06 96.6% 0.02
CASIA-Iris-Interval 89% 0.09 86.6% 0.02
The proposed system recognition accuracy for iris image after segmentation and the required time
to extract features for each iris image are compared to other systems as illustrated in Table 5, the
proposed iris recognition system has overall, outperformed than other feature extraction
algorithms which include, Intersecting Cortical Model (ICM) network [28], circular sector and
triangular DCT [29], discrete wavelet transformation (DWT) [30], Radon transform and
gradient-based isolation [31], Discrete Wavelet Trans Intersecting Cortical Model (ICM) network
form (DWT), Discrete Cosine Transform (DCT) [32], scattering transform and textural features
[6] and the feature extraction using pre-trained VGG-Net [5]. Also, the proposed system feature
extraction for each image takes less time than previous algorithms.
6. CONCLUSIONS
This paper evaluated the extracted learned features from a pre-trained Convolutional Neural
Network (Alex-Net) followed by multi-class SVM algorithm to perform iris recognition. The iris
is segmented using circular Hough transform and normalized using rubber sheet model. The
segmented and normalized image is fed as an input to the CNN (Alex-Net). The proposed system
is tested on public datasets (IITD iris databases, CASIA-Iris-V1, CASIA-Iris-thousand, and
CASIA-Iris- Interval), and a high accuracy rate is achieved. The results showed that the
recognition accuracy when extracting features from the segmented image is higher than when
extracting features from the normalized image.
12. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
76
In the future, we will evaluate the performance of the proposed algorithm using the different pre-
trained model in more iris datasets with other biometric recognition problems.
Table 5. Comparison Between the Performance of The Proposed Iris Recognition Scheme and The Other
Algorithms
Reference#
(Year)
Feature Extraction Database
Recognition
Accuracy
Time (S)
[28]
(2008)
Intersecting Cortical
Model (ICM) network
CASIA-Iris-V1 97.74% Not available
[29]
(2012)
circular sector and
triangular DCT
IIT Delhi Iris Database 97.12% Not available
[30]
(2013)
discrete wavelet
transformation (DWT)
IIT Delhi Iris Database 99.5 Not available
[31]
(2014)
Radon transform and
gradient-based isolation
IIT Delhi Iris Database 95.93% 0.10
CASIA-Iris-Interval 84.17% 0.44
[32]
(2015)
Discrete Wavelet
Transform (DWT) and
Discrete Cosine
Transform (DCT)
IIT Delhi Iris Database 97.81% 93.24
[6]
(2015)
Texture+ Scattering
Features
IIT Delhi Iris Database 99.2% Not available
[5]
(2016)
pre-trained VGG-Net
IIT Delhi Iris Database 99%
Not available
CASIA-Iris-Thousand 90%
The Proposed
scheme
(2018)
pre-trained Alex-Net
IIT Delhi Iris
Database
100% 0.06
CASIA-Iris-V1 98.3% 0.08
CASIA-Iris-
Thousand
98% 0.06
CASIA-Iris-Interval 89% 0.09
ACKNOWLEDGEMENTS
This work was supported by King Abdulaziz City for Science and Technology (KACST) under
Grant Number (PGP – 37 – 1858). Therefore, the authors wish to thank, KACST technical and
financial support.
REFERENCES
[1] M. Haghighat, S. Zonouz, and M. Abdel-Mottaleb, "CloudID: Trustworthy cloud-based and cross-
enterprise biometric identification," Expert Systems with Applications, vol. 42, pp. 7905-7916, 2015.
[2] D. Kesavaraja, D. Sasireka, and D. Jeyabharathi, "Cloud software as a service with iris
authentication," Journal of Global Research in Computer Science, vol. 1, pp. 16-22, 2010.
[3] N. Shah and P. Shrinath, "Iris Recognition System–A Review," International Journal of Computer and
Information Technology, vol. 3, 2014.
[4] A. B. Dehkordi and S. A. Abu-Bakar, "A review of iris recognition system," Jurnal Teknologi, vol.
77, 2015.
[5] S. Minaee, A. Abdolrashidiy, and Y. Wang, "An experimental study of deep convolutional features
for iris recognition," in Signal Processing in Medicine and Biology Symposium (SPMB), 2016 IEEE,
2016, pp. 1-6.
13. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
77
[6] S. Minaee, A. Abdolrashidi, and Y. Wang, "Iris recognition using scattering transform and textural
features," in Signal Processing and Signal Processing Education Workshop (SP/SPE), 2015 IEEE,
2015, pp. 37-42.
[7] S. Minaee, A. Abdolrashidi, and Y. Wang, "Face Recognition Using Scattering Convolutional
Network," arXiv preprint arXiv:1608.00059, 2016.
[8] IIT Delhi Database. Available: http://www4.comp.polyu.edu.hk/~csajaykr/IITD/Database_Iris.htm.
Accessed 14 April 2017.
[9] ( 2 April2017). CASIA Iris Image Database Version 1.0. Available:
http://www.idealtest.org/findDownloadDbByMode.do?mode=Iris. Accessed 12 April 2017.
[10] CASIA Iris Image Database Version 4.0 (CASIA-Iris-Thousand). Available:
http://biometrics.idealtest.org/dbDetailForUser.do?id=4. Accessed 17 April 2017.
[11] CASIA Iris Image Database Version 3.0 (CASIA-Iris-Interval). Available:
http://biometrics.idealtest.org/dbDetailForUser.do?id=3. Accessed 17 April2017.
[12] K. Nguyen, C. Fookes, A. Ross, and S. Sridharan, "Iris Recognition with Off-the-Shelf CNN
Features: A Deep Learning Perspective," IEEE Access, 2017.
[13] A. Romero, C. Gatta, and G. Camps-Valls, "Unsupervised deep feature extraction for remote sensing
image classification," IEEE Transactions on Geoscience and Remote Sensing, vol. 54, pp. 1349-1362,
2016.
[14] O. Oyedotun and A. Khashman, "Iris nevus diagnosis: convolutional neural network and deep belief
network," Turkish Journal of Electrical Engineering & Computer Sciences, vol. 25, pp. 1106-1115,
2017.
[15] A. S. Al-Waisy, R. Qahwaji, S. Ipson, S. Al-Fahdawi, and T. A. Nagem, "A multi-biometric iris
recognition system based on a deep learning approach," Pattern Analysis and Applications, pp. 1-20,
2017.
[16] J. Nagi, F. Ducatelle, G. A. Di Caro, D. Cireşan, U. Meier, A. Giusti, F. Nagi, J. Schmidhuber, and L.
M. Gambardella, "Max-pooling convolutional neural networks for vision-based hand gesture
recognition," in Signal and Image Processing Applications (ICSIPA), 2011 IEEE International
Conference on, 2011, pp. 342-347.
[17] D. Scherer, A. Müller, and S. Behnke, "Evaluation of pooling operations in convolutional
architectures for object recognition," Artificial Neural Networks–ICANN 2010, pp. 92-101, 2010.
[18] J. van Doorn, "Analysis of deep convolutional neural network architectures," 2014.
[19] C. L. Lam and M. Eizenman, "Convolutional neural networks for eye detection in remote gaze
estimation systems," 2008.
[20] S. Ahmad Radzi, K.-H. Mohamad, S. S. Liew, and R. Bakhteri, "Convolutional neural network for
face recognition with pose and illumination variation," International Journal of Engineering and
Technology (IJET), vol. 6, pp. 44-57, 2014.
[21] K. Itqan, A. Syafeeza, F. Gong, N. Mustafa, Y. Wong, and M. Ibrahim, "User identification system
based on finger-vein patterns using Convolutional Neural Network," ARPN Journal of Engineering
and Applied Sciences, vol. 11, pp. 3316-3319, 2016.
[22] S. Sangwan and R. Rani, "A Review on: Iris Recognition," (IJCSIT) International Journal of
Computer Science and Information Technologies, vol. 6, pp. 3871-3873, 2015
14. International Journal of Computer Science & Information Technology (IJCSIT) Vol 10, No 2, April 2018
78
[23] C. Jayachandra and H. V. Reddy, "Iris Recognition based on Pupil using Canny edge detection and K-
Means Algorithm," Int. J. Eng. Comput. Sci., vol. 2, pp. 221-225, 2013.
[24] L. A. Elrefaei, D. H. Hamid, A. A. Bayazed, S. S. Bushnak, and S. Y. Maasher, "Developing Iris
Recognition System for Smartphone Security," Multimedia Tools and Applications, pp. 1-25, 2017.
[25] A. Krizhevsky, I. Sutskever, and G. E. Hinton, "Imagenet classification with deep convolutional
neural networks," in Advances in neural information processing systems, 2012, pp. 1097-1105.
[26] S. Minaee and Y. Wang, "Palmprint Recognition Using Deep Scattering Convolutional Network,"
arXiv preprint arXiv:1603.09027, 2016.
[27] J. Weston and C. Watkins, "Multi-class support vector machines," Technical Report CSD-TR-98-04,
Department of Computer Science, Royal Holloway, University of London, May1998.
[28] G. Xu, Z. Zhang, and Y. Ma, "A novel method for iris feature extraction based on intersecting cortical
model network," Journal of Applied Mathematics and Computing, vol. 26, pp. 341-352, 2008.
[29] M. Abhiram, C. Sadhu, K. Manikantan, and S. Ramachandran, "Novel DCT based feature extraction
for enhanced iris recognition," in Communication, Information & Computing Technology (ICCICT),
2012 International Conference on, 2012, pp. 1-6.
[30] M. Elgamal and N. Al-Biqami, "An efficient feature extraction method for iris recognition based on
wavelet transformation," Int. J. Comput. Inf. Technol, vol. 2, pp. 521-527, 2013.
[31] B. Bharath, A. Vilas, K. Manikantan, and S. Ramachandran, "Iris recognition using radon transform
thresholding based feature extraction with Gradient-based Isolation as a pre-processing technique," in
Industrial and Information Systems (ICIIS), 2014 9th International Conference on, 2014, pp. 1-8.
[32] S. S. Dhage, S. S. Hegde, K. Manikantan, and S. Ramachandran, "DWT-based feature extraction and
radon transform based contrast enhancement for improved iris recognition," Procedia Computer
Science, vol. 45, pp. 256-265, 2015.
AUTHORS
Maram G. Alaslani Received her B.Sc. degree in Computer Science with Honors from King Abdulaziz
University in 2010. She works as Teaching Assistant from 2011 to date at Faculty of Computers and
Information Technology at King Abdulaziz University, Rabigh, Saudi Arabia. Now she is working in her
Master Degree at King Abdulaziz University, Jeddah, Saudi Arabia. She has a research interest in image
processing, pattern recognition, and neural network..
Lamiaa A. Elrefaei received her B.Sc. degree with honors in Electrical Engineering (Electronics and
Telecommunications) in 1997, her M.Sc. in 2003 and Ph.D. in 2008 in Electrical Engineering (Electronics)
from faculty of Engineering at Shoubra, Benha University, Egypt. She held a number of faculty positions at
Benha University, as Teaching Assistant from 1998 to 2003, as an Assistant Lecturer from 2003 to 2008,
and has been a lecturer from 2008 to date. She is currently an Associate Professor at the faculty of
Computing and Information Technology, King Abdulaziz University, Jeddah, Saudi Arabia. Her research
interests include computational intelligence, biometrics, multimedia security, wireless networks, and Nano
networks. She is a senior member of IEEE..