Facial recognition technology has been used in many fields such as security,
biometric identification, robotics, video surveillance, health, and commerce
due to its ease of implementation and minimal data processing time.
However, this technology is influenced by the presence of variations such as
pose, lighting, or occlusion. In this paper, we propose a new approach to
improve the accuracy rate of face recognition in the presence of variation or
occlusion, by combining feature extraction with a histogram of oriented
gradient (HOG), scale invariant feature transform (SIFT), Gabor, and the
Canny contour detector techniques, as well as a convolutional neural
network (CNN) architecture, tested with several combinations of the
activation function used (Softmax and Segmoïd) and the optimization
algorithm used during training (adam, Adamax, RMSprop, and stochastic
gradient descent (SGD)). For this, a preprocessing was performed on two
databases of our database of faces (ORL) and Sheffield faces used, then we
perform a feature extraction operation with the mentioned techniques and
then pass them to our used CNN architecture. The results of our simulations
show a high performance of the SIFT+CNN combination, in the case of the
presence of variations with an accuracy rate up to 100%.
Presentation given in the Seminar of B.Tech 6th Semester during session 2009-10 By Paramjeet Singh Jamwal, Poonam Kanyal, Rittitka Mittal and Surabhi Tyagi.
Presentation given in the Seminar of B.Tech 6th Semester during session 2009-10 By Paramjeet Singh Jamwal, Poonam Kanyal, Rittitka Mittal and Surabhi Tyagi.
Fundamental steps in Digital Image ProcessingShubham Jain
Fundamental Steps in Digital Image Processing: Image acquisition, enhancement, restoration, etc. For written notes and pdf visit: https://buzztech.in/fundamental-steps-in-digital-image-processing
Image processing, Noise, Noise Removal filtersKuppusamy P
Basics of images, Digital Images, Noise, Noise Removal filters
Reference:
Richard Szeliski, Computer Vision: Algorithms and Applications, Springer 2010
Disease prediction using machine learningJinishaKG
Github link :
https://github.com/jini-the-coder/Diseaseprediction
Blog link :
http://amigoscreation.blogspot.com/2020/07/disease-prediction-using-machine.html
Youtube link :
https://youtu.be/3YmAbta16yk
Fundamental steps in Digital Image ProcessingShubham Jain
Fundamental Steps in Digital Image Processing: Image acquisition, enhancement, restoration, etc. For written notes and pdf visit: https://buzztech.in/fundamental-steps-in-digital-image-processing
Image processing, Noise, Noise Removal filtersKuppusamy P
Basics of images, Digital Images, Noise, Noise Removal filters
Reference:
Richard Szeliski, Computer Vision: Algorithms and Applications, Springer 2010
Disease prediction using machine learningJinishaKG
Github link :
https://github.com/jini-the-coder/Diseaseprediction
Blog link :
http://amigoscreation.blogspot.com/2020/07/disease-prediction-using-machine.html
Youtube link :
https://youtu.be/3YmAbta16yk
Facial recognition based on enhanced neural networkIAESIJAI
Accurate automatic face recognition (FR) has only become a practical goal of biometrics research in recent years. Detection and recognition are the primary steps for identifying faces in this research, and The Viola-Jones algorithm implements to discover faces in images. This paper presents a neural network solution called modify bidirectional associative memory (MBAM). The basic idea is to recognize the image of a human's face, extract the face image, enter it into the MBAM, and identify it. The output ID for the face image from the network should be similar to the ID for the image entered previously in the training phase. The tests have conducted using the suggested model using 100 images. Results show that FR accuracy is 100% for all images used, and the accuracy after adding noise is the proportions that differ between the images used according to the noise ratio. Recognition results for the mobile camera images were more satisfactory than those for the Face94 dataset.
A face recognition system using convolutional feature extraction with linear...IJECEIAES
Face recognition is one of the important biometric authentication research areas for security purposes in many fields such as pattern recognition and image processing. However, the human face recognitions have the major problem in machine learning and deep learning techniques, since input images vary with poses of people, different lighting conditions, various expressions, ages as well as illumination conditions and it makes the face recognition process poor in accuracy. In the present research, the resolution of the image patches is reduced by the max pooling layer in convolutional neural network (CNN) and also used to make the model robust than other traditional feature extraction technique called local multiple pattern (LMP). The extracted features are fed into the linear collaborative discriminant regression classification (LCDRC) for final face recognition. Due to optimization using CNN in LCDRC, the distance ratio between the classes has maximized and the distance of the features inside the class reduces. The results stated that the CNN-LCDRC achieved 93.10% and 87.60% of mean recognition accuracy, where traditional LCDRC achieved 83.35% and 77.70% of mean recognition accuracy on ORL and YALE databases respectively for the training number 8 (i.e. 80% of training and 20% of testing data).
In this study, an end-to-end human iris recognition system is presented to automatically identify individuals for high level of security purposes. The deep learning technology based new 2D convolutional neural network (CNN) model is introduced for extracting the features and classifying the iris patterns. Firstly, the iris dataset is collected, preprocessed and augmented. The dataset are expanded and enhanced using data augmentation, histogram equalization (HE) and contrast-limited adaptive histogram equalization (CLAHE) techniques. Secondly, the features of the iris patterns were extracted and classified using CNN. The structure of CNN comprises of convolutional layers and ReLu layers for extracting the features, pooling layers for reducing the parameters, fully connected layer and Softmax layer for classifying the extracted features into N classes. For the training process and updating the weights, the backpropagation algorithm and adaptive moment estimation Adam optimizer are used. The experimental results carried out based on a graphics processing unit (GPU) and using Matlab. The overall training accuracy of the introduced system was 95.33% with a consumption time of 17.59 minutes for training set. While the testing accuracy 100% with a consumption time of 12 seconds. The introduced iris recognition system has been successfully applied.
Deep hypersphere embedding for real-time face recognitionTELKOMNIKA JOURNAL
With the advancement of human-computer interaction capabilities of robots, computer vision surveillance systems involving security yields a large impact in the research industry by helping in digitalization of certain security processes. Recognizing a face in the computer vision involves identification and classification of which faces belongs to the same person by means of comparing face embedding vectors. In an organization that has a large and diverse labelled dataset on a large number of epoch, oftentimes, creates a training difficulties involving incompatibility in different versions of face embedding that leads to poor face recognition accuracy. In this paper, we will design and implement robotic vision security surveillance system incorporating hybrid combination of MTCNN for face detection, and FaceNet as the unified embedding for face recognition and clustering.
Social distance and face mask detector system exploiting transfer learningIJECEIAES
As time advances, the use of deep learning-based object detection algorithms has also evolved leading to developments of new human-computer interactions, facilitating an exploration of various domains. Considering the automated process of detection, systems suitable for detecting violations are developed. One such applications is the social distancing and face mask detectors to control air-borne diseases. The objective of this research is to deploy transfer learning on object detection models for spotting violations in face masks and physical distance rules in real-time. The common drawbacks of existing models are low accuracy and inability to detect in real-time. The MobileNetV2 object detection model and YOLOv3 model with Euclidean distance measure have been used for detection of face mask and physical distancing. A proactive transfer learning approach is used to perform the functionality of face mask classification on the patterns obtained from the social distance detector model. On implementing the application on various surveillance footage, it was observed that the system could classify masked and unmasked faces and if social distancing was maintained or not with accuracies 99% and 94% respectively. The models exhibited high accuracy on testing and the system can be infused with the existing internet protocol (IP) cameras or surveillance systems for real-time surveillance of face masks and physical distancing rules effectively.
Efficient resampling features and convolution neural network model for image ...IJEECSIAES
The extended utilization of picture-enhancing or manipulating tools has led to ease of manipulating multimedia data which includes digital images. These manipulations will disturb the truthfulness and lawfulness of images, resulting in misapprehension, and might disturb social security. The image forensic approach has been employed for detecting whether or not an image has been manipulated with the usage of positive attacks which includes splicing, and copy-move. This paper provides a competent tampering detection technique using resampling features and convolution neural network (CNN). In this model range spatial filtering (RSF)-CNN, throughout preprocessing the image is divided into consistent patches. Then, within every patch, the resampling features are extracted by utilizing affine transformation and the Laplacian operator. Then, the extracted features are accumulated for creating descriptors by using CNN. A wide-ranging analysis is performed for assessing tampering detection and tampered region segmentation accuracies of proposed RSF-CNN based tampering detection procedures considering various falsifications and post-processing attacks which include joint photographic expert group (JPEG) compression, scaling, rotations, noise additions, and more than one manipulation. From the achieved results, it can be visible the RSF-CNN primarily based tampering detection with adequately higher accurateness than existing tampering detection methodologies.
Efficient resampling features and convolution neural network model for image ...nooriasukmaningtyas
The extended utilization of picture-enhancing or manipulating tools has led to ease of manipulating multimedia data which includes digital images. These manipulations will disturb the truthfulness and lawfulness of images, resulting in misapprehension, and might disturb social security. The image forensic approach has been employed for detecting whether or not an image has been manipulated with the usage of positive attacks which includes splicing, and copy-move. This paper provides a competent tampering detection technique using resampling features and convolution neural network (CNN). In this model range spatial filtering (RSF)-CNN, throughout preprocessing the image is divided into consistent patches. Then, within every patch, the resampling features are extracted by utilizing affine transformation and the Laplacian operator. Then, the extracted features are accumulated for creating descriptors by using CNN. A wide-ranging analysis is performed for assessing tampering detection and tampered region segmentation accuracies of proposed RSF-CNN based tampering detection procedures considering various falsifications and post-processing attacks which include joint photographic expert group (JPEG) compression, scaling, rotations, noise additions, and more than one manipulation. From the achieved results, it can be visible the RSF-CNN primarily based tampering detection with adequately higher accurateness than existing tampering detection methodologies.
Deep learning based masked face recognition in the era of the COVID-19 pandemicIJECEIAES
During the coronavirus disease 2019 (COVID-19) pandemic, monitoring for wearing masks obtains a crucial attention due to the effect of wearing masks to prevent the spread of coronavirus. This work introduces two deep learning models, the former based on pre-trained convolutional neural network (CNN) which called MobileNetv2, and the latter is a new CNN architecture. These two models have been used to detect masked face with three classes (correct, not correct, and no mask). The experiments conducted on benchmark dataset which is face mask detection dataset from Kaggle. Moreover, the comparison between two models is driven to evaluate the results of these two proposed models.
Convolutional neural network with binary moth flame optimization for emotion ...IAESIJAI
Electroencephalograph (EEG) signals have the ability of real-time reflecting brain activities. Utilizing the EEG signal for analyzing human emotional states is a common study. The EEG signals of the emotions aren’t distinctive and it is different from one person to another as every one of them has different emotional responses to same stimuli. Which is why, the signals of the EEG are subject dependent and proven to be effective for the subject dependent detection of the Emotions. For the purpose of achieving enhanced accuracy and high true positive rate, the suggested system proposed a binary moth flame optimization (BMFO) algorithm for the process of feature selection and convolutional neural networks (CNNs) for classifications. In this proposal, optimum features are chosen with the use of accuracy as objective function. Ultimately, optimally chosen features are classified after that with the use of a CNN for the purpose of discriminating different emotion states.
A novel ensemble model for detecting fake newsIAESIJAI
Due the growing proliferation of fake news over the past couple of years, our objective in this paper is to propose an ensemble model for the automatic classification of article news as being either real or fake. For this purpose, we opt for a blending technique that combines three models, namely bidirectional long short-term memory (Bi-LSTM), stochastic gradient descent classifier and ridge classifier. The implementation of the proposed model (i.e. BI-LSR) on real world datasets, has shown outstanding results. In fact, it achieved an accuracy score of 99.16%. Accordingly, this ensemble learning has proven to do perform better than individual conventional machine learning and deep learning models as well as many ensemble learning approaches cited in the literature.
K-centroid convergence clustering identification in one-label per type for di...IAESIJAI
Disease prediction is a high demand field which requires significant support from machine learning (ML) to enhance the result efficiency. The research works on application of K-means clustering supervised classification in disease prediction where each class only has one labeled data. The K-centroid convergence clustering identification (KC3 I) system is based on semi-K-means clustering but only requires single labeled data per class for the training process with the training dataset to update the centroid. The KC3 I model also includes a dictionary box to index all the input centroids before and after the updating process. Each centroid matches with a corresponding label inside this box. After the training process, each time the input features arrive, the trained centroid will put them to its cluster depending on the Euclidean distance, then convert them into the specific class name, which is coherent to that centroid index. Two validation stages were carried out and accomplished the expectation in terms of precision, recall, F1-score, and absolute accuracy. The last part demonstrates the possibility of feature reduction by selecting the most crucial feature with the extra tree classifier method. Total data are fed into the KC3 I system with the most important features and remain the same accuracy.
Plant leaf detection through machine learning based image classification appr...IAESIJAI
Since maize is a staple diet for people, especially vegetarians and vegans, maize leaf disease has a significant influence here on the food industry including maize crop productivity. Therefore, it should be understood that maize quality must be optimal; yet, to do so, maize must be safeguarded from several illnesses. As a result, there is a great demand for such an automated system that can identify the condition early on and take the appropriate action. Early disease identification is crucial, but it also poses a major obstacle. As a result, in this research project, we adopt the fundamental k-nearest neighbor (KNN) model and concentrate on building and developing the enhanced k-nearest neighbor (EKNN) model. EKNN aids in identifying several classes of disease. To gather discriminative, boundary, pattern, and structurally linked information, additional high-quality fine and coarse features are generated. This information is then used in the classification process. The classification algorithm offers high-quality gradient-based features. Additionally, the proposed model is assessed using the Plant-Village dataset, and a comparison with many standard classification models using various metrics is also done.
Backbone search for object detection for applications in intrusion warning sy...IAESIJAI
In this work, we propose a novel backbone search method for object detection for applications in intrusion warning systems. The goal is to find a compact model for use in embedded thermal imaging cameras widely used in intrusion warning systems. The proposed method is based on faster region-based convolutional neural network (Faster R-CNN) because it can detect small objects. Inspired by EfficientNet, the sought-after backbone architecture is obtained by finding the most suitable width scale for the base backbone (ResNet50). The evaluation metrics are mean average precision (mAP), number of parameters, and number of multiply–accumulate operations (MACs). The experimental results showed that the proposed method is effective in building a lightweight neural network for the task of object detection. The obtained model can keep the predefined mAP while minimizing the number of parameters and computational resources. All experiments are executed elaborately on the person detection in intrusion warning systems (PDIWS) dataset.
Deep learning method for lung cancer identification and classificationIAESIJAI
Lung cancer (LC) is calming many lives and is becoming a serious cause of concern. The detection of LC at an early stage assists the chances of recovery. Accuracy of detection of LC at an early stage can be improved with the help of a convolutional neural network (CNN) based deep learning approach. In this paper, we present two methodologies for Lung cancer detection (LCD) applied on Lung image database consortium (LIDC) and image database resource initiative (IDRI) data sets. Classification of these LC images is carried out using support vector machine (SVM), and deep CNN. The CNN is trained with i) multiple batches and ii) single batch for LC image classification as non cancer and cancer image. All these methods are being implemented in MATLAB. The accuracy of classification obtained by SVM is 65%, whereas deep CNN produced detection accuracy of 80% and 100% respectively for multiple and single batch training. The novelty of our experimentation is near 100% classification accuracy obtained by our deep CNN model when tested on 25 Lung computed tomography (CT) test images each of size 512×512 pixels in less than 20 iterations as compared to the research work carried out by other researchers using cropped LC nodule images.
Optically processed Kannada script realization with Siamese neural network modelIAESIJAI
Optical character recognition (OCR) is a technology that allows computers to recognize and extract text from images or scanned documents. It is commonly used to convert printed or handwritten text into machine-readable format. This Study presents an OCR system on Kannada Characters based on siamese neural network (SNN). Here the SNN, a Deep neural network which comprises of two identical convolutional neural network (CNN) compare the script and ranks based on the dissimilarity. When lesser dissimilarity score is identified, prediction is done as character match. In this work the authors use 5 classes of Kannada characters which were initially preprocessed using grey scaling and convert it to pgm format. This is directly input into the Deep convolutional network which is learnt from matching and non-matching image between the CNN with contrastive loss function in Siamese architecture. The Proposed OCR system uses very less time and gives more accurate results as compared to the regular CNN. The model can become a powerful tool for identification, particularly in situations where there is a high degree of variation in writing styles or limited training data is available.
Embedded artificial intelligence system using deep learning and raspberrypi f...IAESIJAI
Melanoma is a kind of skin cancer that originates in melanocytes responsible for producing melanin, it can be a severe and potentially deadly form of cancer because it can metastasize to other regions of the body if not detected and treated early. To facilitate this process, Recently, various computer-assisted low-cost, reliable, and accurate diagnostic systems have been proposed based on artificial intelligence (AI) algorithms, particularly deep learning techniques. This work proposed an innovative and intelligent system that combines the internet of things (IoT) with a Raspberry Pi connected to a camera and a deep learning model based on the deep convolutional neural network (CNN) algorithm for real-time detection and classification of melanoma cancer lesions. The key stages of our model before serializing to the Raspberry Pi: Firstly, the preprocessing part contains data cleaning, data transformation (normalization), and data augmentation to reduce overfitting when training. Then, the deep CNN algorithm is used to extract the features part. Finally, the classification part with applied Sigmoid Activation Function. The experimental results indicate the efficiency of our proposed classification system as we achieved an accuracy rate of 92%, a precision of 91%, a sensitivity of 91%, and an area under the curve- receiver operating characteristics (AUC-ROC) of 0.9133.
Deep learning based biometric authentication using electrocardiogram and irisIAESIJAI
Authentication systems play an important role in wide range of applications. The traditional token certificate and password-based authentication systems are now replaced by biometric authentication systems. Generally, these authentication systems are based on the data obtained from face, iris, electrocardiogram (ECG), fingerprint and palm print. But these types of models are unimodal authentication, which suffer from accuracy and reliability issues. In this regard, multimodal biometric authentication systems have gained huge attention to develop the robust authentication systems. Moreover, the current development in deep learning schemes have proliferated to develop more robust architecture to overcome the issues of tradition machine learning based authentication systems. In this work, we have adopted ECG and iris data and trained the obtained features with the help of hybrid convolutional neural network- long short-term memory (CNN-LSTM) model. In ECG, R peak detection is considered as an important aspect for feature extraction and morphological features are extracted. Similarly, gabor-wavelet, gray level co-occurrence matrix (GLCM), gray level difference matrix (GLDM) and principal component analysis (PCA) based feature extraction methods are applied on iris data. The final feature vector is obtained from MIT-BIH and IIT Delhi Iris dataset which is trained and tested by using CNN-LSTM. The experimental analysis shows that the proposed approach achieves average accuracy, precision, and F1-core as 0.985, 0.962 and 0.975, respectively.
Hybrid channel and spatial attention-UNet for skin lesion segmentationIAESIJAI
Melanoma is a type of skin cancer which has affected many lives globally. The American Cancer Society research has suggested that it a serious type of skin cancer and lead to mortality but it is almost 100% curable if it is detected and treated in its early stages. Currently automated computer vision-based schemes are widely adopted but these systems suffer from poor segmentation accuracy. To overcome these issue, deep learning (DL) has become the promising solution which performs extensive training for pattern learning and provide better classification accuracy. However, skin lesion segmentation is affected due to skin hair, unclear boundaries, pigmentation, and mole. To overcome this issue, we adopt UNet based deep learning scheme and incorporated attention mechanism which considers low level statistics and high-level statistics combined with feedback and skip connection module. This helps to obtain the robust features without neglecting the channel information. Further, we use channel attention, spatial attention modulation to achieve the final segmentation. The proposed DL based scheme is instigated on publically available dataset and experimental investigation shows that the proposed Hybrid Attention UNet approach achieves average performance as 0.9715, 0.9962, 0.9710.
Photoplethysmogram signal reconstruction through integrated compression sensi...IAESIJAI
The transmission of photoplethysmogram (PPG) signals in real-time is extremely challenging and facilitates the use of an internet of things (IoT) environment for healthcare- monitoring. This paper proposes an approach for PPG signal reconstruction through integrated compression sensing and basis function aware shallow learning (CSBSL). Integrated-CSBSL approach for combined compression of PPG signals via multiple channels thereby improving the reconstruction accuracy for the PPG signals essential in healthcare monitoring. An optimal basis function aware shallow learning procedure is employed on PPG signals with prior initialization; this is further fine-tuned by utilizing the knowledge of various other channels, which exploit the further sparsity of the PPG signals. The proposed method for learning combined with PPG signals retains the knowledge of spatial and temporal correlation. The proposed Integrated-CSBSL approach consists of two steps, in the first step the shallow learning based on basis function is carried out through training the PPG signals. The proposed method is evaluated using multichannel PPG signal reconstruction, which potentially benefits clinical applications through PPG monitoring and diagnosis.
Speaker identification under noisy conditions using hybrid convolutional neur...IAESIJAI
Speaker identification is biometrics that classifies or identifies a person from other speakers based on speech characteristics. Recently, deep learning models outperformed conventional machine learning models in speaker identification. Spectrograms of the speech have been used as input in deep learning-based speaker identification using clean speech. However, the performance of speaker identification systems gets degraded under noisy conditions. Cochleograms have shown better results than spectrograms in deep learning-based speaker recognition under noisy and mismatched conditions. Moreover, hybrid convolutional neural network (CNN) and recurrent neural network (RNN) variants have shown better performance than CNN or RNN variants in recent studies. However, there is no attempt conducted to use a hybrid CNN and enhanced RNN variants in speaker identification using cochleogram input to enhance the performance under noisy and mismatched conditions. In this study, a speaker identification using hybrid CNN and the gated recurrent unit (GRU) is proposed for noisy conditions using cochleogram input. VoxCeleb1 audio dataset with real-world noises, white Gaussian noises (WGN) and without additive noises were employed for experiments. The experiment results and the comparison with existing works show that the proposed model performs better than other models in this study and existing works.
Multi-channel microseismic signals classification with convolutional neural n...IAESIJAI
Identifying and classifying microseismic signals is essential to warn of mines’ dangers. Deep learning has replaced traditional methods, but labor-intensive manual identification and varying deep learning outcomes pose challenges. This paper proposes a transfer learning-based convolutional neural network (CNN) method called microseismic signals-convolutional neural network (MS-CNN) to automatically recognize and classify microseismic events and blasts. The model was instructed on a limited sample of data to obtain an optimal weight model for microseismic waveform recognition and classification. A comparative analysis was performed with an existing CNN model and classical image classification models such as AlexNet, GoogLeNet, and ResNet50. The outcomes demonstrate that the MS-CNN model achieved the best recognition and classification effect (99.6% accuracy) in the shortest time (0.31 s to identify 277 images in the test set). Thus, the MS-CNN model can efficiently recognize and classify microseismic events and blasts in practical engineering applications, improving the recognition timeliness of microseismic signals and further enhancing the accuracy of event classification.
Sophisticated face mask dataset: a novel dataset for effective coronavirus di...IAESIJAI
Efficient and accurate coronavirus disease (COVID-19) surveillance necessitates robust identification of individuals wearing face masks. This research introduces the sophisticated face mask dataset (SFMD), a comprehensive compilation of high-quality face mask images enriched with detailed annotations on mask types, fits, and usage patterns. Leveraging cutting-edge deep learning models—EfficientNet-B2, ResNet50, and MobileNet-V2—, we compare SFMD against two established benchmarks: the real-world masked face dataset (RMFD) and the masked face recognition dataset (MFRD). Across all models, SFMD consistently outperforms RMFD and MFRD in key metrics, including accuracy, precision, recall, and F1 score. Additionally, our study demonstrates the dataset's capability to cultivate robust models resilient to intricate scenarios like low-light conditions and facial occlusions due to accessories or facial hair.
Transfer learning for epilepsy detection using spectrogram imagesIAESIJAI
Epilepsy stands out as one of the common neurological diseases. The neural activity of the brain is observed using electroencephalography (EEG). Manual inspection of EEG brain signals is a slow and arduous process, which puts heavy load on neurologists and affects their performance. The aim of this study is to find the best result of classification using the transfer learning model that automatically identify the epileptic and the normal activity, to classify EEG signals by using images of spectrogram which represents the percentage of energy for each coefficient of the continuous wavelet. Dataset includes the EEG signals recorded at monitoring unit of epilepsy used in this study to presents an application of transfer learning by comparing three models Alexnet, visual geometry group (VGG19) and residual neural network ResNet using different combinations with seven different classifiers. This study tested the models and reached a different value of accuracy and other metrics used to judge their performances, and as a result the best combination has been achieved with ResNet combined with support vector machine (SVM) classifier that classified EEG signals with a high success rate using multiple performance metrics such as 97.22% accuracy and 2.78% the value of the error rate.
Deep neural network for lateral control of self-driving cars in urban environ...IAESIJAI
The exponential growth of the automotive industry clearly indicates that self-driving cars are the future of transportation. However, their biggest challenge lies in lateral control, particularly in urban bottlenecking environments, where disturbances and obstacles are abundant. In these situations, the ego vehicle has to follow its own trajectory while rapidly correcting deviation errors without colliding with other nearby vehicles. Various research efforts have focused on developing lateral control approaches, but these methods remain limited in terms of response speed and control accuracy. This paper presents a control strategy using a deep neural network (DNN) controller to effectively keep the car on the centerline of its trajectory and adapt to disturbances arising from deviations or trajectory curvature. The controller focuses on minimizing deviation errors. The Matlab/Simulink software is used for designing and training the DNN. Finally, simulation results confirm that the suggested controller has several advantages in terms of precision, with lateral deviation remaining below 0.65 meters, and rapidity, with a response time of 0.7 seconds, compared to traditional controllers in solving lateral control.
Attention mechanism-based model for cardiomegaly recognition in chest X-Ray i...IAESIJAI
Recently, cardiovascular diseases (CVDs) have become a rapidly growing problem in the world, especially in developing countries. The latter are facing a lifestyle change that introduces new risk factors for heart disease, that requires a particular and urgent interest. Besides, cardiomegaly is a sign of cardiovascular diseases that refers to various conditions; it is associated with the heart enlargement that can be either transient or permanent depending on certain conditions. Furthermore, cardiomegaly is visible on any imaging test including Chest X-Radiation (X-Ray) images; which are one of the most common tools used by Cardiologists to detect and diagnose many diseases. In this paper, we propose an innovative deep learning (DL) model based on an attention module and MobileNet architecture to recognize Cardiomegaly patients using the popular Chest X-Ray8 dataset. Actually, the attention module captures the spatial relationship between the relevant regions in Chest X-Ray images. The experimental results show that the proposed model achieved interesting results with an accuracy rate of 81% which makes it suitable for detecting cardiomegaly disease.
Efficient commodity price forecasting using long short-term memory modelIAESIJAI
Predicting commodity prices, particularly food prices, is a significant concern for various stakeholders, especially in regions that are highly sensitive to commodity price volatility. Historically, many machine learning models like autoregressive integrated moving average (ARIMA) and support vector machine (SVM) have been suggested to overcome the forecasting task. These models struggle to capture the multifaceted and dynamic factors influencing these prices. Recently, deep learning approaches have demonstrated considerable promise in handling complex forecasting tasks. This paper presents a novel long short-term memory (LSTM) network-based model for commodity price forecasting. The model uses five essential commodities namely bread, meat, milk, oil, and petrol. The proposed model focuses on advanced feature engineering which involves moving averages, price volatility, and past prices. The results reveal that our model outperforms traditional methods as it achieves 0.14, 3.04%, and 98.2% for root mean square error (RMSE), mean absolute percentage error (MAPE), and R-squared (R2 ), respectively. In addition to the simplicity of the model, which consists of an LSTM single-cell architecture that reduced the training time to a few minutes instead of hours. This paper contributes to the economic literature on price prediction using advanced deep learning techniques as well as provides practical implications for managing commodity price instability globally.
1-dimensional convolutional neural networks for predicting sudden cardiacIAESIJAI
Sudden cardiac arrest (SCA) is a serious heart problem that occurs without symptoms or warning. SCA causes high mortality. Therefore, it is important to estimate the incidence of SCA. Current methods for predicting ventricular fibrillation (VF) episodes require monitoring patients over time, resulting in no complications. New technologies, especially machine learning, are gaining popularity due to the benefits they provide. However, most existing systems rely on manual processes, which can lead to inefficiencies in disseminating patient information. On the other hand, existing deep learning methods rely on large data sets that are not publicly available. In this study, we propose a deep learning method based on one-dimensional convolutional neural networks to learn to use discrete fourier transform (DFT) features in raw electrocardiogram (ECG) signals. The results showed that our method was able to accurately predict the onset of SCA with an accuracy of 96% approximately 90 minutes before it occurred. Predictions can save many lives. That is, optimized deep learning models can outperform manual models in analyzing long-term signals.
A deep learning-based approach for early detection of disease in sugarcane pl...IAESIJAI
In many regions of the nation, agriculture serves as the primary industry. The farming environment now faces a number of challenges to farmers. One of the major concerns, and the focus of this research, is disease prediction. A methodology is suggested to automate a process for identifying disease in plant growth and warning farmers in advance so they can take appropriate action. Disease in crop plants has an impact on agricultural production. In this work, a novel DenseNet-support vector machine: explainable artificial intelligence (DNet-SVM: XAI) interpretation that combines a DenseNet with support vector machine (SVM) and local interpretable model-agnostic explanation (LIME) interpretation has been proposed. DNet-SVM: XAI was created by a series of modifications to DenseNet201, including the addition of a support vector machine (SVM) classifier. Prior to using SVM to identify if an image is healthy or un-healthy, images are first feature extracted using a convolution network called DenseNet. In addition to offering a likely explanation for the prediction, the reasoning is carried out utilizing the visual cue produced by the LIME. In light of this, the proposed approach, when paired with its determined interpretability and precision, may successfully assist farmers in the detection of infected plants and recommendation of pesticide for the identified disease.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Albert Hoitingh
In this session I delve into the encryption technology used in Microsoft 365 and Microsoft Purview. Including the concepts of Customer Key and Double Key Encryption.
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf91mobiles
91mobiles recently conducted a Smart TV Buyer Insights Survey in which we asked over 3,000 respondents about the TV they own, aspects they look at on a new TV, and their TV buying preferences.
Le nuove frontiere dell'AI nell'RPA con UiPath Autopilot™UiPathCommunity
In questo evento online gratuito, organizzato dalla Community Italiana di UiPath, potrai esplorare le nuove funzionalità di Autopilot, il tool che integra l'Intelligenza Artificiale nei processi di sviluppo e utilizzo delle Automazioni.
📕 Vedremo insieme alcuni esempi dell'utilizzo di Autopilot in diversi tool della Suite UiPath:
Autopilot per Studio Web
Autopilot per Studio
Autopilot per Apps
Clipboard AI
GenAI applicata alla Document Understanding
👨🏫👨💻 Speakers:
Stefano Negro, UiPath MVPx3, RPA Tech Lead @ BSP Consultant
Flavio Martinelli, UiPath MVP 2023, Technical Account Manager @UiPath
Andrei Tasca, RPA Solutions Team Lead @NTT Data
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Assuring Contact Center Experiences for Your Customers With ThousandEyes
A hybrid approach for face recognition using a convolutional neural network combined with feature extraction techniques
1. IAES International Journal of Artificial Intelligence (IJ-AI)
Vol. 12, No. 2, June 2023, pp. 627~640
ISSN: 2252-8938, DOI: 10.11591/ijai.v12.i2.pp627-640 627
Journal homepage: http://ijai.iaescore.com
A hybrid approach for face recognition using a convolutional
neural network combined with feature extraction techniques
Hicham Benradi, Ahmed Chater, Abdelali Lasfar
High School of Technology Salé, Mohammadia School of Engineering, Systems Analysis and Information Processing and Industrial
Management Laboratory, Mohammed V University, Rabat, Morocco
Article Info ABSTRACT
Article history:
Received Feb 6, 2022
Revised Sep 27, 2022
Accepted Oct 27, 2022
Facial recognition technology has been used in many fields such as security,
biometric identification, robotics, video surveillance, health, and commerce
due to its ease of implementation and minimal data processing time.
However, this technology is influenced by the presence of variations such as
pose, lighting, or occlusion. In this paper, we propose a new approach to
improve the accuracy rate of face recognition in the presence of variation or
occlusion, by combining feature extraction with a histogram of oriented
gradient (HOG), scale invariant feature transform (SIFT), Gabor, and the
Canny contour detector techniques, as well as a convolutional neural
network (CNN) architecture, tested with several combinations of the
activation function used (Softmax and Segmoïd) and the optimization
algorithm used during training (adam, Adamax, RMSprop, and stochastic
gradient descent (SGD)). For this, a preprocessing was performed on two
databases of our database of faces (ORL) and Sheffield faces used, then we
perform a feature extraction operation with the mentioned techniques and
then pass them to our used CNN architecture. The results of our simulations
show a high performance of the SIFT+CNN combination, in the case of the
presence of variations with an accuracy rate up to 100%.
Keywords:
Activation function
Convolutional neural network
Deep learning
Facial recognition
Optimization algorithms
Scale invariant feature
transform+convolutional neural
network
This is an open access article under the CC BY-SA license.
Corresponding Author:
Hicham Benradi
Laboratory of Systems Analysis and Information Processing and Industrial Management Laboratory
High School of Technology Salé, Mohammadia School of Engineering, University Mohammed V
Rabat, Morocco
Email: benradi.hicham@gmail.com
1. INTRODUCTION
Facial recognition is a technology that belongs to the field of computer vision. It is used in several
fields such as crime detection from video surveillance [1], security [2], drowsiness and fatigue detection [3],
biometric security [4], and also in robotics [5]. It aims at identifying and authenticating a person from an
image or a video sequence. This authentication is done from a process that starts with the detection and
extraction of the face from an image or a video sequence, then the extracted face is processed by feature
extraction techniques [6], [7]. This operation consists of performing a mathematical transformation calculated
on all the pixels of an image, allowing it to identify the visual properties of an image so that they can be used
for further processing. This operation can be performed using several techniques such as the histogram of
oriented gradient (HOG) descriptor [8], scale invariant feature transform (SIFT) which is a point of interest
detector, Gabor filter [9] which is a dedicated convolution filter for texture analysis, and CANNY edge
detector which is an image contour detector. The final phase of this process aims at authentication by
performing a comparison with other images stored in a database [10], using classifiers such as support vector
machine (SVM) [11], k-nearest neighbors (KNN) [12], or principal component analysis (PCA) [13].
2. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
628
Recently, several techniques have been implemented [14]–[17] , aiming at identifying an individual
from an image with maximum accuracy. However, the efficiency of a facial recognition system remains
ineffective in case of variations in the processed image such as pose, lighting, pre and sence of occlusion.
Recently, most of the work in face recognition uses the deep learning technique which is based on the
architecture of convolutional neural networks (CNN) inspired by biological neural networks, because of their
robustness in terms of analysis and feature extraction thanks to their deep architecture, allowing to perform a
very large number of computations on a single region of an image. They are used in several domains, such as
medicine [18]–[20], agriculture [21]–[23], and economics [24].
A CNN is an architecture that consists of several layers and parameters. There are two main parts in this
architecture, which are the feature extraction part and the classification part. The first part mainly uses convolution
layers to extract features and form a feature map and a pooling layer which reduces the dimension of the feature
map to reduce the computation rate. In the second part, a fully connected layer is used to assign each image to a
specific class that is suitable for it, based on the result of the first part, and another layer called dropout is used to
avoid overlearning our model in the training data set. It randomly drops several neurons during the model training
to reduce the size of the model. It also uses an activation function that selects the most relevant variables to pass to
the next neuron. There are several activation functions such as Relu Softmax, and Sigmoide, and each function is
used according to the context of the classification sought, either a binary or a multi-class classification. After the
design of the CNN, the architecture comes the last part which is the training of the model. This part contains an
important parameter, the optimization algorithm used, which reduces the error rate. There are several types of
optimizations algorithms, among which are d stochastic gradient descent (SGD).
In this paper, we propose a hybrid approach based on the best feature extraction algorithm with
different face variations, which is associated with the CNN architect modified according to the Softmax and
Sigmoide activations function that allocancelingcel the negative values and speed up the processing time and
finally evaluate by the following optimization techniques: Adam, Adamax, RMSprop, and SGD. To evaluate
our technique, we use two databases which are: Our database of faces (ORL) and Sheffield [25], with
different percentages of the base tested and trained, which present different variations (illumination, contrast,
occlusion, rotation). The results of our simulations allowed us to give good results in terms of accuracy rate
which reached 100% with different face variations.
2. RELATED WORK
CNN has attracted the attention of many researchers in the field of face recognition due to its
enormous capacity in terms of accuracy rate. Research works have been developed based on deep learning,
the authors in [26] proposed a face recognition system in the case of the presence of occlusion or noisy faces
that is based on deep learning using a deep neural network (DNN), for this the features are extracted in a
cascade from the images separately is then processed to select the most relevant then these are used by a
DNN for a classification. Experimental results showed that this method achieved an accuracy of 92.3%.
Another method has been proposed for face recognition under unfavorable conditions (difficult lighting, blur,
and low resolution) by [27] which uses CNNs to project the covariance matrices of Gabor waves into a
feature vector of the Euclidean space. This method effectively extracts fine features from an image and has
been shown to perform better than DNN. Another face recognition method was developed for use in a Big
data environment by [28] who optimized a face recognition algorithm that combines two feature extraction
techniques which are local binary pattern (LBP) algorithm and two-dimensional principle component
analysis (2DPCA) these features are subsequently merged to pass them to a CNN as input data. In the context
of big data, the accuracy of this technique could exceed 90%. The use of the linear discriminant analysis
(LDA) technique to generate a set of one-dimensional facial features from an original image dataset for
training the classifier of a one-dimensional deep convolutional neural network (1D-DCNN) was used in [29].
This method demonstrated a high accuracy performance that reached 100% accuracy.
3. METHODOLOGY
The proposed approach is based on four essential steps. First, a pre-processing of the data used.
Then a feature extraction operation is performed using techniques (HOG, SIFT, GABOR and CANNY).
Then, using a convolutional neural network architecture (CNN), we train our classification model. Finally, a
calculation of the accuracy rate will be performed.
3.1. Preprocessing of the data used
3.1.1. Database used
ORL dataset: ORL is a database of faces that contains 400 images in total distributed over 40
individuals, i.e. 10 images per individual. The conditions under which these images were taken differ either
3. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
629
by the details of the face (with or without glasses) or by the variance of the lighting or the facial expression
(smiling or not, eyes open or closed). All the images are in gray level with a size of 92x112 pixels. Figure 1
shows some images belonging to the database.
Sheffield dataset: Sheffield is a database of faces of 20 individuals and contains a total of 564 gray
level images of identical size of 220X220 pixels [25]. Each individual is represented by poses ranging from
profile to full face view taken under different conditions with variance either by gender, race, or appearance.
Figure 2 shows some images belonging to the database.
Figure 1. Example of images with variance from the ORL database
Figure 2. Layout from profile to front view from Sheffield database
3.1.2. Data preparations
A pre-processing is performed on all the images used. It aims to label each image with a
corresponding label for each set of images of the same individual, then transform them into the gray level,
then resizes them to an identical size which is 48×48 pixels to convert them into an array of pixel values and
reconvert them into float and finally to form two groups of data which are the first called Train for the
training part of the model and the second test which will be used to test the effectiveness of the model. The
Figure 3 shows the final form of the preprocessing performed on the images.
Figure 3. Preprocessing was performed on all the images of the ORL database
3.2. Feature extraction
3.2.1. Scale invariant feature transform (SIFT)
SIFT [30], is a very powerful and well-recognized point of interest detector in the field of facial
recognition, it is based on the calculation of the euclidean distance between two vectors to determine if they
correspond to the same points of interest in different images [31]. This technique goes through four main
steps Scale-space extrema detection, localization of key points, Orientation assignment, and extraction of key
point descriptors [30] detailed,
− In the first step, the different key points are identified in an image using the difference of Gaussians
(DOG) from multiple Gaussian images produced at different scales from the original image where the
DoGs are computed from the neighbors in the scale space.
− The second step consists of locating the different candidate keypoints based on extrema existing in the
DoGs, eliminating unstable keypoints with low contrast [30].
− The third step assigns a principal orientation to each key point.
− The final phase computes a highly distinctive descriptor for each keypoint.
4. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
630
The Figure 4 shows an example of key point detection by the SIFT technique. This detection was
performed using images from the ORL face database, with a size of 112×92 pixels. A set of parameters was
used to detect its key points, defined: the number of layers in each octave is 3, the contrast threshold is a value
of 0.04, the threshold for filtering the edges is a value of 10 and the value of the sigma of the Gaussian is 1.6.
Figure 4. Detection of key points by the SIFT
3.2.2. Histogram of oriented gradient (HOG)
HOG or histogram of the oriented gradient is a descriptor used in the field of image processing. It
was proposed by Dalal and Triggs in [8] and it aims to represent the appearance and local shape of an object
in the image by a distribution of the intensity of the gradient. To do this, the image must be resized to a size
of 64×128 pixels because it uses a detection window of the same size, then divide the image into small cells
and then calculate each cell the histogram of the gradient directions by calculating the magnitude using the
following formula,
||𝛻𝑓|| = √(
𝜕𝑓
𝜕𝑥
)
2
+ (
𝜕𝑓
𝜕𝑦
)
2
(1)
Where𝜕𝑥: the value of the gradient in the horizontal direction
𝜕𝑦: the value of the gradient in the vertical direction
Then the orientation of the gradient is calculated by
𝜃 = 𝑡𝑎𝑛−1
(
𝜕𝑓
𝜕𝑦
/
𝜕𝑓
𝜕𝑥
) (2)
And at the end, all these histograms will be combined to form a HOG descriptor.
Figure 5 shows an example of edge detection using the HOG technique. This detection was performed
using images from the ORL face database, with a size of 112×92 pixels. For this, the number of orientation bins
is 9, the cell size is 8×8 pixels, and several cells in each block for histogram normalization (2.2) :
Figure 5. Descriptor detection by the HOG
3.2.3. The Canny contour detector
Canny is an image processing algorithm whose goal is to detect the contour of an image [32]. It
allows to reduce the amount of data to be processed and this by extracting useful structural information. This
detection is achieved by following these steps,
1) The first step is to remove the noise using a Gaussian filter because the presence of noise in an image
influences the detection of the contour.
5. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
631
2) Then a search of the intensity gradient of the image is performed using a Sobel filter this is done in both
directions of the image (horizontal Gx and vertical Gy) and from these two images the gradient and the
direction of the edges for each pixel are identified using the formula,
𝐸𝑑𝑔𝑒_𝐺𝑟𝑎𝑑𝑖𝑒𝑛𝑡(𝐺) = √𝐺𝑥
2 + 𝐺𝑦
2 (3)
𝐴𝑛𝑔𝑙𝑒(𝜃) = 𝑡𝑎𝑛−1
(
𝐺𝑦
𝐺𝑥
) (4)
Where: Gx: is the first derivative in the horizontal direction.
Gy: is the first derivative in the vertical direction.
3) By analyzing the entire image, we eliminate the unwanted pixels that do not correspond to an edge. This
operation is performed by comparing a pixel with its neighborhood in the direction of the gradient.
Finally checked if it is a local maximum then it is an edge otherwise it is eliminated. Figure 6 is a
graphical presentation of the edge selection operation. The first case Figure 6(a) shows how the
CANNY detector proceeds to eliminate a pixel. The second case shows Figure 6(b) how CANNY
selects an edge.
(a) (b)
Figure 6. Determination of an edge (a) Case of elimination of minimal pixel B in the neighborhood (pixels A
and C) not corresponding to an edge and (b) Maximum pixel B in the neighborhood (pixels A and C)
corresponding to an edge
4) The last step is to select the best edges using the hysteresis threshold. To do this we need two threshold
values, minval and maxval, then all the edges with an intensity gradient greater than maxval are
considered as edges and in the case where the intensity gradient is less than minval is eliminated
because it is not an edge. In the case where the intensity gradient is located between the two values
minval and maxval, checking if it is connected to another pixel higher than maxval, in this case, it is
considered as an edge.
The Figure 7 shows an example of edge detection using the CANNY technique. This detection was
performed using images from the ENT face database, with a size of 112×92 pixels. For this, the threshold
value of the maximum intensity gradient used is 100 and the minimum value is 200,
Figure 7. Descriptor detection by CANNY
3.2.4. Gabor filter
A Gabor filter is a convolution filter used in computer vision and especially image processing for
texture analysis, edge detection, or feature extraction from an image [9]. They are special classes of bandpass
filter because it carries out a selection of the bands of frequencies to identify among them which one it must
6. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
632
allow and which one must be rejected. It represents a combination of Gaussian and sinusoidal terms where
the Gaussian component provides the weights and the sinusoidal component provides the direction, as the
following formula shows,
'2 2 '2
2
'
( , , , , , , ) exp exp 2
2
x y x
g x y i
+
= − +
(5)
Where: λ: wavelength of the sinusoidal component. θ: orientation of the filter. ψ: phase shift. σ: gaussian
filter. γ: spatial aspect ratio and: x' = x cos θ + y sin θ and y' = -x sin θ + y cos θ
The Gabor kernel mimics the visual cortex which means simulating using the Gabor kernel how do
we recognize the texture with our eyes that can be, which round it a bank of filters that can be designed to
detect the texture and extract the texture. The Figure 8 shows an example of detecting key points of images
from the ORL face database, of size 112×92 pixel by Gabor. For this purpose, six filters with six main
directions were used to extract the features from the images. Their angle is defined (0°, 30°, 60°, 90°, 120°,
and 150°) respectively. The size of each filter is 11 pixels, the standard deviation of the Gaussian function
Sigma= 1.5, and the wavelength of the sinusoidal factor Lambda= 3.
The Figure 9 summarizes the result of the feature extraction phase on the same images. For this we
used the HOG, SIFT, CANNY and GABOR techniques to extract features from images. It is noted that each
technique was used with a specific parameterization.
Figure 8. Descriptor detection by GABOR
Figure 9. Feature extraction from ORL database images
3.3. Model training using a convolutional neural network CNN
This phase is devoted to the creation of our CNN. Its objective is to train our model using the Train
part and to validate the model using the Test part. For this we have adopted the following architecture,
− Three convolution layers which contain respectively a filter value of 6, 16, and 64 with a kernel of 5x5
for the first two layers and 3x3 for the third and rectified linear unit (ReLU) as activation function.
7. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
633
− Each layer will be followed by a Maxpool layer with a value of (2.2), which will have the goal of
reducing the input size to half.
− A Flatten layer to flatten all values.
− Two fully connected dense layers the first one uses ReLU as an activation function followed by a
Dropout to avoid overfitting at a threshold set to 0.5 and the second one will use Softmax for the first
case and Sigmoid for the second case the as activation function for classification.
The Figure 10 is a summary diagram of the architecture of our neural network developed in our
method,
Figure 10. The architecture of the CNN neural network used
our elaborated CNN architecture has been tested with several combinations using two different activation
functions Softmax and Segmoïd, in the dense classification layer, and the optimization algorithms adam,
Adamax, RMSprop, and SGD, used in the training to evaluate which of these combinations gives better
results. Activation functions are functions capable of spatially modifying the representation of the data,
allowing it to switch from a linear to a non-linear form. For our approach, we used three types of activation
functions which are :
Rectified linear unit: Known for its simplicity which makes it the most used among the other
activation functions, the rectified linear unit or ReLU function [33] is a function that aims at determining the
maximum between x and 0.
x if x > 0 and
0 if x < 0
Fonction_ReLU(x) = max (x, 0) (6)
Softmax: This is a function often used by classification models for multi-class problems. It treats
each vector independently of the others and allows the transformation of a real vector into a probability
vector. The input axis on which Softmax is applied is defined by the axis argument.
fonction_Softmax(x) = exp(x)/tf.reduce_sum(exp(x)) (7)
fonction_Softmax(x) = exp(x)/sum(exp(xi)) (8)
Sigmoid: it is a function that is used for binary classification in the case where the model will have
to determine only two labels because the results are always between 0 and 1. It uses the following formula,
8. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
634
fonction_Sigmoid(x) = 1/(1 + exp(-x)) (9)
during the training of the model four optimization algorithms were used which are:
Adam: proposed by Kingma & Ba in 2014 in their paper [34], Adam is an optimization algorithm
based on stochastic gradient descent. Its role is to update the weights of a neural network iteratively,
according to the training data. For that it uses the following formula,
1
n n n
n
v
m
+
= −
+ (10)
Where: θn+1: weight of time n+1; θn: weight of time n; vn: sum of the square of the past gradients
α: step size parameter; ϵ: a constant; mn: the aggregate of the gradient of time n
Adamax: this is an extension of the Adam optimization algorithm. Adamax [34] updates the weights
of a neural network based on the infinite norm of past gradients. For this it uses the following formula,
𝜽𝒏+𝟏 = 𝜽𝒏 −
𝜼
𝒖𝒏
𝒎
∧
𝒏 (11)
Where: θn+1: weight of time n+1; θn: weight of time n; mn: gradient aggregate of time n
𝜂: 0.002; un: is used to denote the infinite norm constraint vt,
𝑢𝑛 = 𝛽2
∞
𝑣𝑛−1 + (1 − 𝛽2
∞)|𝑔𝑛|∞
(12)
𝑢𝑛 = 𝑚𝑎𝑥(𝛽2. 𝑣𝑛−1, |𝑔𝑛|) (13)
Where: β1 and β2 are default values (β1=0.9 and β2=0.999)
RMSprop: Proposed by Geoffrey Hinton, RMSpro is an optimization technique used in the training
phase. It was designed for mini-batch learning as a stochastic technique to solve the gradient explosion
problems using complex functions. It is based on the gradient and specifically the normalization of the
gradient. The step size is modified to avoid the explosion in the case of large gradients the step size is
decreased and to avoid the disappearance in the case of small gradients the step size is increased to create an
equilibrium. In other words, it aims to change the learning rate according to the size of the gradients.
RMSprop update rule,
𝐸[𝑔2]𝑛 = 𝛽𝐸[𝑔2]𝑛−1 + (1 − 𝛽) (
𝛿𝑐
𝛿𝑤
)
2
(14)
𝜃𝑛+1 = 𝜃𝑛 −
𝜂
√𝐸[𝑔2]𝑛
𝛿𝑐
𝛿𝑤
(15)
Where: E[g]: moving average of the square gradients
δC/δw: gradient of the cost function concerning the weight
𝜂: learning rate; β: moving average parameter (0.9)
Stochastic gradient descent (SGD): is an optimization algorithm. It eliminates the redundant
computations performed by batch gradient descent in the case of large data sets which recalculates the
gradients before each parameter update and performs one update at a time using the following formula,
𝜃 = 𝜃 − 𝜂. 𝛻𝜃𝑗(𝜃; 𝑥(𝑖)
; 𝑦(𝑖)
) (16)
Where: θ : model parameters ; 𝜂: learning rate ; ∇θJ(θ) : objective function
𝑥(𝑖)
: training set; label
3.4. Calculation of the accuracy rate
The last step of our method is devoted to the evaluation of the model by calculating the accuracy
rate. For this, we used the categorical_crossentropy function as a loss function. It is generally used in multi-
class classification and its purpose is to decide in which class among all existing classes a result from a CNN
can be assigned using the following formula,
9. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
635
1
.log
output
size
i
i
Loss y
i
y
=
= −
(17)
Where: output size: number of scalar values in the model output
i
y
: scalar value in the output of the model; corresponding target value
The following diagram as shown in Figure 11 summarizes the different steps of our approach,
Figure 11. Summary diagram of the steps of our method
4. RESULTS AND DISCUSSION
The results of our simulations were performed on two databases ORL and Sheffield. The purpose of
these simulations is to evaluate the accuracy rate for each database. All models were trained using 150
epochs and a batch size of 40 for the ORL database and 58 for the Sheffield database.
4.1. Evaluation of the proposed method using the ORL database
The results of our simulations performed on the Test part of the ORL database using the Softmax
activation function gave satisfactory results in terms of accuracy rate in the case of using Adam, Adamax,
and RMSprop optimizers with a value higher than 80% for the feature extraction techniques HOG, SIFT,
Gabor and Canny. On the other hand, using the SGD optimization algorithm, the accuracy gave less
compared to the others. As we can see from the Table 1 we can say that the best accuracy is obtained using
the Adam optimizer with an accuracy rate varying between 80% as the minimum value in the case of Canny
and 93.75% as the maximum value in the case of Gabor.In the case of using the Sigmoid activation function,
the results obtained by the HOG technique are the lowest compared to the other feature extraction techniques
used, especially in the case of using SGD as the optimization algorithm. The reading made on the Table 1,
which represents graphically the totality of our simulation results, shows that when using the combination of
SIFT with the CNN the results obtained are satisfactory and stable whatever the combination (activation
function/optimization algorithm) used, with an accuracy rate that reached 92.5%.
The results of our simulations are the Figure 12, in terms of accuracy and loss rate of the
SIFT+CNN method, Figure 12(a) presents the Softmax/RMSprop combination use case and Figure 12(b)
presents the Sigmoid/Adamax combination use case. The accuracy on the training set reaches a value of
100% and a value of 91.25% on the test set for validation in the Softmax/RMSprop case and the loss function
reaches a minimum value of 0.0679 on the training set and 0.187 on the validation set. In the case of using
the Sigmoid/Adamx combination, a value of 98.12% was achieved on the training set and 92.5% for the
validation set. For the loss function, it reached a minimum value of 0.0013 on the training set and 0.0047 on
the validation set.
4.2. Evaluation of the proposed method using the Sheffield database
The results of our simulation performed using the Softmax activation function in the function of
optimization algorithms (Adam, Adamax, RMSprop, and SGD), allowed us to record good results in terms of
accuracy rate. In the case of using the three optimization algorithms adam, Adamax and RMSprop, the
accuracy rate exceeded the value of 97.39%. It is also noted that the combination SIFT+CNN has recorded
the highest accuracy rates (adam = 100%, Adamx = 99.13%, RMSprop = 100% and SGD=98.26%).
10. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
636
Table 1. The performance of accuracy rate in (%) according to the ORL face database
Fonction d'activation Softmax Fonction d'activation Sigmoid
Optimisateur Adam Adamax RMSprop SGD Adam Adamax RMSprop SGD
HOG + CNN 91.25 92.5 91.25 0 5 91.25 97.5 0
LBP + CNN 68.75 75 77.45 0 45 63.75 3.75 0
SIFT + CNN 88.75 88.75 91.25 86.25 88.75 92.5 91.25 92.5
GABOR + CNN 93.75 88.75 90 88.75 90 87.5 90 92.5
CANNY + CNN 80 86.25 87.5 83.75 82.5 83.75 87.5 87.5
(a)
(b)
Figure 12. Accuracy and loss rates as a function of the number of epochs according to the SIFT +CNN
technique, (a) Softmax+RMSprop, and (b) Sigmoid+Adamax
For the case of using the Sigmoid activation function, the results obtained show that the accuracy
rate decreases when using the SGD optimization algorithm. On the other hand, using other algorithms gives
satisfactory results, especially when using the Adamax algorithm, because all the rates recorded are high,
varying between 96.52% and 99.13%. The results as shown in the Table 2, shows a remarkable performance
of the feature extraction techniques Sift, Gabor, and Canny. On the other hand, the rates recorded by the Hog
techniques are less satisfactory.
11. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
637
Table 2. The performance of accuracy rate in (%) according to the Sheffield face database
Fonction d'activation Softmax Fonction d'activation Sigmoid
Optimisateur adam Adamax RMSprop SGD adam Adamax RMSprop SGD
HOG + CNN 97.39 99.13 97.39 92.17 9.56 98.26 99.13 10.43
LBP + CNN 89.56 87.82 97.39 10.43 91.3 96.52 6.08 4.34
SIFT + CNN 100 99.13 100 98.26 98.26 99.13 99.13 97.39
GABOR + CNN 98.26 99.13 99.13 98.26 98.26 99.13 99.13 97.39
CANNY + CNN 98.26 97.4 99.13 96.52 97.4 96.52 98.26 98.26
The results of our simulations are shown in the Figure 13 in appendix, in terms of accuracy and loss
rates of the SIFT+CNN method, Figure 13(a) shows the use case of Softmax/RMSprop combination and
Figure 13(b) shows the use case of Sigmoid/Adamax combination. The accuracy on the training set reaches a
value of 99.56% and a value of 100% on the test set for validation in the Softmax/RMSprop case and the loss
function reaches a minimum value of 0.0090 on the training set and 0.0039 on the validation set. In the case
of using the Sigmoid/Adamx combination, a value of 99.34% was achieved on the training set and 99.13%
for the validation set. For the loss function, it reached a minimum value of 0.0258 on the training set and
0.0511 on the validation set.
(a)
(b)
Figure 13. Accuracy and loss rates as a function of the number of epochs according to the SIFT +CNN
technique, (a) Softmax+RMSprop, and (b) Sigmoid+Adamax
12. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
638
To better evaluate the performance of our approach, a comparative study with other existing
approaches was performed. The Table 3 summarizes the accuracy rate results of our approach and other
approaches applied on the same two datasets ORL and Sheffield. It results that our approach is competitive
and achieves a better result in terms of recognition rate with the presence of different variations.
Table 3. Comparison of recognition rate of the proposed method with another existing approach
ORL dataset Sheffield dataset
Method Accuracy (%) Method Accuracy (%)
NSST [15] 99.32 SURF+SVM [35] 97.87
LBP+CNN [36] 100 Parameterless SLPP [37] 95.6
DNNs [38] 99.07 2DJLNDA + CNN [39] 89.87
PCA + SVM [40] 98.75 V-LGS + LSAD [41] 95
SIFT [42] 91.2 UFSELM+L2.1 [43] 76.89
(MF_GF_HE)_PCA_ MultiSVMs [44] 91.6 AS-LRC [45] 85.42
Proposed method 100 Proposed method 100
5. CONCLUSION
Facial recognition is a field of artificial intelligence that aims to identify individuals from an image.
It is a complicated task, for applications such as the identification of individuals by video surveillance when
variations or occlusions are present. In this paper, we propose a hybrid approach based on the best feature
extraction algorithm with different face variations, which is associated with the CNN architect modified
according to the Softmax and Sigmoide activation function that allows to cancel the negative values and to
speed up the processing time, and finally evaluated by the following optimization techniques: adam,
Adamax, RMSprop and SGD.The results obtained showed a remarkable performance when using
SIFT+CNN with an accuracy rate of up to 100%. We also note that when using the Softmax activation
function and the Adam, Adamax, and RMSprop optimization algorithms, the results are satisfactory in terms
of accuracy rate. This approach can be used in the field of security by video surveillance to identify
individuals in case of the presence of variations. It is noted that testing on other databases representing new
cases of variations is important to better evaluate our proposed model.
REFERENCES
[1] K. K. Kumar, Y. Kasiviswanadham, D. V. S. N. V. Indira, P. Priyanka palesetti, and Ch. V. Bhargavi, “Criminal face
identification system using deep learning algorithm multi-task cascade neural network (MTCNN),” Materials Today:
Proceedings, 2021, doi: 10.1016/j.matpr.2021.06.373.
[2] P. Wang, P. Wang, and E. Fan, “Violence detection and face recognition based on deep learning,” Pattern Recognition Letters,
vol. 142, pp. 20–24, 2021, doi: 10.1016/j.patrec.2020.11.018.
[3] N. N. A. Mangshor, I. A. A. Majid, S. Ibrahim, and N. Sabri, “A real-time drowsiness and fatigue recognition using support
vector machine,” IAES International Journal of Artificial Intelligence, vol. 9, no. 4, pp. 584–590, 2020, doi:
10.11591/ijai.v9.i4.pp584-590.
[4] E. Maiorana, “Deep learning for EEG-based biometric recognition,” Neurocomputing, vol. 410, pp. 374–386, 2020, doi:
10.1016/j.neucom.2020.06.009.
[5] E. Yaghoubi, D. Borza, J. Neves, A. Kumar, and H. Proença, “An attention-based deep learning model for multiple pedestrian
attributes recognition,” Image and Vision Computing, vol. 102, 2020, doi: 10.1016/j.imavis.2020.103981.
[6] V. Vijayan and P. Kp, “A comparative analysis of RootSIFT and SIFT methods for drowsy features extraction,” Procedia
Computer Science, vol. 171, pp. 436–445, 2020, doi: 10.1016/j.procs.2020.04.046.
[7] A. Chater and A. Lasfar, “New approach to the identification of the easy expression recognition system by robust techniques
(SIFT, PCA-SIFT, ASIFT and SURF),” Telkomnika (Telecommunication Computing Electronics and Control), vol. 18, no. 2, pp.
695–704, 2020, doi: 10.12928/TELKOMNIKA.V18I2.13726.
[8] N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” Proceedings - 2005 IEEE Computer Society
Conference on Computer Vision and Pattern Recognition, CVPR 2005, vol. I, pp. 886–893, 2005, doi: 10.1109/CVPR.2005.177.
[9] S. Fischer, G. Cristóbal, and R. Redondo, “Sparse overcomplete gabor wavelet representation based on local competitions,” IEEE
Transactions on Image Processing, vol. 15, no. 2, pp. 265–272, 2006, doi: 10.1109/TIP.2005.860614.
[10] A. Chater and A. Lasfar, “Comparison of robust methods for extracting descriptors and facial matching,” 2019 International
Conference on Wireless Technologies, Embedded and Intelligent Systems, WITS 2019, 2019, doi: 10.1109/WITS.2019.8723858.
[11] L. Shi, X. Wang, and Y. Shen, “Research on 3D face recognition method based on LBP and SVM,” Optik, vol. 220, 2020, doi:
10.1016/j.ijleo.2020.165157.
[12] M. Sowmiya Manoj and S. Arulselvi, “Palm print identification and classification using KNN algorithm,” Materials Today:
Proceedings, 2021, doi: 10.1016/j.matpr.2021.01.804.
[13] Y. Ren, X. Xu, G. Feng, and X. Zhang, “Non-Interactive and secure outsourcing of PCA-Based face recognition,” Computers and
Security, vol. 110, 2021, doi: 10.1016/j.cose.2021.102416.
[14] A. H. Thary Al-Ghrairi, A. Abdulwahhab Mohammed, and E. Z. Sameen, “Face detection and recognition with 180 degree
rotation based on principal component analysis algorithm,” IJ-AI, vol. 11, no. 2, p. 593, Jun. 2022, doi:
10.11591/ijai.v11.i2.pp593-602.
13. Int J Artif Intell ISSN: 2252-8938
A hybrid approach for face recognition using a … (Hicham Benradi)
639
[15] Y. S and N. Purnachand, “Convolutional neural network-based face recognition using non-subsampled shearlet transform and
histogram of local feature descriptors,” IJ-AI, vol. 10, no. 4, p. 1079, Dec. 2021, doi: 10.11591/ijai.v10.i4.pp1079-1090.
[16] A. Chater, H. Benradi, and A. Lasfar, “Method of optimization of the fundamental matrix by technique speeded up robust features
application of different stress images,” IJECE, vol. 12, no. 2, p. 1429, Apr. 2022, doi: 10.11591/ijece.v12i2.pp1429-1436.
[17] B. Hicham, C. Ahmed, and L. Abdelali, “Face recognition method combining SVM machine learning and scale invariant feature
transform,” E3S Web Conf., vol. 351, p. 01033, 2022, doi: 10.1051/e3sconf/202235101033.
[18] W. Chen et al., “Deep diagnostic agent forest (DDAF): A deep learning pathogen recognition system for pneumonia based on
CT,” Computers in Biology and Medicine, vol. 141, 2022, doi: 10.1016/j.compbiomed.2021.105143.
[19] A. Salehzadeh, A. P. Calitz, and J. Greyling, “Human activity recognition using deep electroencephalography learning,”
Biomedical Signal Processing and Control, vol. 62, 2020, doi: 10.1016/j.bspc.2020.102094.
[20] M. Polsinelli, L. Cinque, and G. Placidi, “A light CNN for detecting COVID-19 from CT scans of the chest,” Pattern Recognition
Letters, vol. 140, pp. 95–100, 2020, doi: 10.1016/j.patrec.2020.10.001.
[21] F. Jiang, Y. Lu, Y. Chen, D. Cai, and G. Li, “Image recognition of four rice leaf diseases based on deep learning and support
vector machine,” Computers and Electronics in Agriculture, vol. 179, 2020, doi: 10.1016/j.compag.2020.105824.
[22] M. Agarwal, S. K. Gupta, and K. K. Biswas, “Development of efficient CNN model for Tomato crop disease identification,”
Sustainable Computing: Informatics and Systems, vol. 28, 2020, doi: 10.1016/j.suscom.2020.100407.
[23] C. Chen, W. Zhu, and T. Norton, “Behaviour recognition of pigs and cattle: Journey from computer vision to deep learning,”
Computers and Electronics in Agriculture, vol. 187, 2021, doi: 10.1016/j.compag.2021.106255.
[24] J. Zhao, “Efficiency of corporate debt financing based on machine learning and convolutional neural network,” Microprocessors
and Microsystems, vol. 83, 2021, doi: 10.1016/j.micpro.2021.103998.
[25] Eds. H. Wechsler, P. J. Phillips, V. Bruce, F. F. Soulié, and T. S. Huang, “Face recognition: from theory to applications,”
Springer Berlin Heidelberg, 1998.
[26] N. Soni, E. K. Sharma, and A. Kapoor, “Hybrid meta-heuristic algorithm based deep neural network for face recognition,”
Journal of Computational Science, vol. 51, 2021, doi: 10.1016/j.jocs.2021.101352.
[27] C. Li, Y. Huang, W. Huang, and F. Qin, “Learning features from covariance matrix of gabor wavelet for face recognition under
adverse conditions,” Pattern Recognition, vol. 119, 2021, doi: 10.1016/j.patcog.2021.108085.
[28] Y. Zhu and Y. Jiang, “Optimization of face recognition algorithm based on deep learning multi feature fusion driven by big data,”
Image and Vision Computing, vol. 104, 2020, doi: 10.1016/j.imavis.2020.104023.
[29] J. Mohammed Sahan, E. I. Abbas, and Z. M. Abood, “A facial recognition using a combination of a novel one-dimension deep
CNN and LDA,” Materials Today: Proceedings, 2021, doi: 10.1016/j.matpr.2021.07.325.
[30] D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no.
2, pp. 91–110, 2004, doi: 10.1023/B:VISI.0000029664.99615.94.
[31] A. E.-T. A. Chater, and A. Lasfar, “Face recognition using feature extraction and similarity measures,” Technology Reports of
Kansai University, vol. 62, no. 3, p. 10, 2020.
[32] J. Canny, “A computational approach to edge detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol.
PAMI-8, no. 6, pp. 679–698, 1986, doi: 10.1109/TPAMI.1986.4767851.
[33] L. Deng et al., “Recent advances in deep learning for speech research at Microsoft,” ICASSP, IEEE International Conference on
Acoustics, Speech and Signal Processing - Proceedings, pp. 8604–8608, 2013, doi: 10.1109/ICASSP.2013.6639345.
[34] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” 3rd International Conference on Learning
Representations, ICLR 2015 - Conference Track Proceedings, 2015.
[35] B. Anand, “Face recognition using SURF features and SVM Classifier,” p. 8.
[36] J. Tang, Q. Su, B. Su, S. Fong, W. Cao, and X. Gong, “Parallel ensemble learning of convolutional neural networks and local
binary patterns for face recognition,” Computer Methods and Programs in Biomedicine, vol. 197, p. 105622, Dec. 2020, doi:
10.1016/j.cmpb.2020.105622.
[37] F. Dornaika and A. Assoum, “Enhanced and parameterless Locality Preserving Projections for face recognition,”
Neurocomputing, vol. 99, pp. 448–457, Jan. 2013, doi: 10.1016/j.neucom.2012.07.016.
[38] M. N. ElBedwehy, G. M. Behery, and R. Elbarougy, “Face recognition based on relative gradient magnitude strength,” Arab J Sci
Eng, vol. 45, no. 12, Art. no. 12, Dec. 2020, doi: 10.1007/s13369-020-04538-y.
[39] J. Yu, H. Liu, and X. Zheng, “Two-dimensional joint local and nonlocal discriminant analysis-based 2D image feature extraction
for deep learning,” Neural Comput & Applic, vol. 32, no. 10, Art. no. 10, May 2020, doi: 10.1007/s00521-019-04085-0.
[40] xinzheng Chen, L. Song, and C. Qiu, “Face recognition by feature extraction and classification,” in 2018 12th IEEE International
Conference on Anti-counterfeiting, Security, and Identification (ASID), Xiamen, China, Nov. 2018, pp. 43–46. doi:
10.1109/ICASID.2018.8693198.
[41] R. D. Rakshit, S. C. Nath, and D. R. Kisku, “Face identification using some novel local descriptors under the influence of facial
complexities,” Expert Systems with Applications, vol. 92, pp. 82–94, Feb. 2018, doi: 10.1016/j.eswa.2017.09.038.
[42] K. V. Arya, S. S. Rajput, and S. Upadhyay, “Noise-robust low-resolution face recognition using SIFT features,” in Computational
Intelligence: Theories, Applications and Future Directions - Volume II, vol. 799, N. K. Verma and A. K. Ghosh, Eds. Singapore:
Springer Singapore, 2019, pp. 645–655. doi: 10.1007/978-981-13-1135-2_49.
[43] J. Chen, Y. Zeng, Y. Li, and G.-B. Huang, “Unsupervised feature selection based extreme learning machine for clustering,”
Neurocomputing, vol. 386, pp. 198–207, Apr. 2020, doi: 10.1016/j.neucom.2019.12.065.
[44] H. M. Maw, S. M. Thu, and M. T. Mon, “Face recognition based on illumination invariant techniques model,” in 2019
International Conference on Advanced Information Technologies (ICAIT), Yangon, Myanmar, Nov. 2019, pp. 120–125. doi:
10.1109/AITC.2019.8921027.
[45] Z. Zhang et al., “Adaptive structure-constrained robust latent low-rank coding for image recovery,” in 2019 IEEE International
Conference on Data Mining (ICDM), Beijing, China, Nov. 2019, pp. 846–855. doi: 10.1109/ICDM.2019.00095.
14. ISSN: 2252-8938
Int J Artif Intell, Vol. 12, No. 2, June 2023: 627-640
640
BIOGRAPHIES OF AUTHORS
Hicham Benradi born on October 07, 1985. He holds a bachelor's degree in
Mobile Application Engineering from the Ecole Supérieure de Technologie in Salé, and a
master's degree in Data Engineering and Software Development from the Faculty of Science at
Mohamed V University. He is a Ph.D. student at the Mohammadia School of Engineering in
Rabat. His research focuses on facial recognition methods and image processing. He can be
contacted at the following email address: benradi.hicham@gmail.com.
Ahmed Chater born on December 30, 1986, in Taounate. Degree in Engineering
Sciences and Techniques, specialty: Image Processing, Laboratory of Systems Analysis,
Information Processing and Industrial Management, Mohamed V University Rabat
(Mohammadia School). My research focuses on the Segmentation and restoration of different
types of color and grayscale images. Classification and recognition of facial expressions and
machine learning. He can be contacted at email: ahmedchater11@gmail.com.
Abdelali Lasfar was born on January 10, 1971 in Salé. He is a Professor of Higher
Education at Mohammed V Agdal University, Salé Higher School of Technology, Morocco.
His research focuses on compression methods, indexing by image content and image indexing,
and knowledge extraction from images. He can be contacted at email: ali.lasfar@gmail.com.