This document outlines the PhD thesis of Taleb ALASHKAR on 3D dynamic facial sequence analysis for face recognition and emotion detection. The thesis proposes frameworks for 4D face recognition and 4D spontaneous emotion detection using subspace representations of 3D facial sequences and modeling trajectories on the Grassmann manifold. Experimental results on public databases show the frameworks achieve better recognition and detection performance than state-of-the-art methods, especially in expression-independent face recognition and early detection of emotions like happiness and pain.
The document discusses challenges and approaches for facial emotion recognition. It aims to develop a model-based approach for real-time driver emotion recognition on an embedded platform using parallel processing. Model-based approaches can overcome issues like illumination and pose variations. The document reviews several state-of-the-art methods and discusses challenges like occlusion, lighting distortions, and complex backgrounds. It describes exploring both 2D and 3D techniques for facial feature extraction and expression recognition.
Facial expression recognition based on image featureTasnim Tara
This document presents a method for facial expression recognition based on image features. It discusses existing works that use techniques like PCA and Gabor wavelets for feature extraction and Euclidean distance for classification. The proposed method uses Gaussian filtering, radial symmetry transform, and edge projection for feature extraction, and calculates a feature vector based on geometric facial parameters to classify expressions using Euclidean distance. It aims to recognize six basic expressions accurately from the JAFFE database and could be developed for real-time video recognition in the future.
An Enhanced Independent Component-Based Human Facial Expression Recognition ...أحلام انصارى
This document presents a facial expression recognition system that uses enhanced independent component analysis and fisher linear discriminant analysis (EICA-FLDA) for feature extraction from video frames, and hidden Markov models (HMM) for expression recognition. The system is tested on the Cohn-Kanade facial expression database and achieves a mean recognition rate of 93.23% for six universal expressions (anger, joy, sad, disgust, fear, surprise). Facial expression recognition has applications in human-computer interaction domains like online gaming.
This document presents information on face detection techniques. It discusses image segmentation as a preprocessing step for face detection. Some common segmentation methods are thresholding, edge-based segmentation, and region-based segmentation. Face detection can be classified as implicit/pattern-based or explicit/knowledge-based. Implicit methods use techniques like templates, PCA, LDA, and neural networks, while explicit methods exploit cues like color, motion, and facial features. One method discussed is human skin color-based face detection, which filters for skin-colored regions and finds facial parts within those regions. Advantages include speed and independence from training data, while disadvantages include sensitivity to lighting and accessories.
1. The document proposes a hybrid approach to facial expression recognition that combines appearance features extracted using Local Directional Number descriptors with geometric features based on distances between facial landmark points.
2. The features are classified independently using SVMs and the scores are fused at the decision level using product rule fusion to identify facial expressions in images.
3. Experiments on the CK+ and JAFFE databases show the hybrid approach achieves better recognition rates than using appearance or geometric features individually.
This document discusses using Local Binary Patterns (LBP) for facial expression recognition from images. It summarizes that LBP is used to extract features from facial regions that are divided into cells. Histograms of these LBP features are concatenated into a feature vector for each image. Support Vector Machines (SVM) are used to classify the images into six basic expressions based on these vectors, achieving better results than template matching or Linear Discriminant Analysis. The document also proposes a method for expression-invariant face recognition that classifies an input image, neutralizes it, then searches a neutral image dataset to find potential matches.
Facial Expression Recognition Using Local Binary Pattern and Support Vector M...AM Publications
Facial expression analysis is a remarkable and demanding problem, and impacts significant applications in various fields like human-computer interaction and data-driven animation. Developing an efficient facial representation from the original face images is a crucial step for achieving facial expression recognition. Facial representation based on statistical local features, Local Binary Patterns (LBP) is practically assessed. Several machine learning techniques were thoroughly observed on various databases. LBP features- which are effectual and competent for facial expression recognition are generally used by researchers Cohn Kanade is the database for present work and the programming language used is MATLAB. Firstly, face area is divided in small regions, by which histograms, Local Binary Patterns (LBP) are extracted and then concatenated into single feature vector. This feature vector outlines a well-organized representation of face and is helpful in determining the resemblance among images.
The document discusses challenges and approaches for facial emotion recognition. It aims to develop a model-based approach for real-time driver emotion recognition on an embedded platform using parallel processing. Model-based approaches can overcome issues like illumination and pose variations. The document reviews several state-of-the-art methods and discusses challenges like occlusion, lighting distortions, and complex backgrounds. It describes exploring both 2D and 3D techniques for facial feature extraction and expression recognition.
Facial expression recognition based on image featureTasnim Tara
This document presents a method for facial expression recognition based on image features. It discusses existing works that use techniques like PCA and Gabor wavelets for feature extraction and Euclidean distance for classification. The proposed method uses Gaussian filtering, radial symmetry transform, and edge projection for feature extraction, and calculates a feature vector based on geometric facial parameters to classify expressions using Euclidean distance. It aims to recognize six basic expressions accurately from the JAFFE database and could be developed for real-time video recognition in the future.
An Enhanced Independent Component-Based Human Facial Expression Recognition ...أحلام انصارى
This document presents a facial expression recognition system that uses enhanced independent component analysis and fisher linear discriminant analysis (EICA-FLDA) for feature extraction from video frames, and hidden Markov models (HMM) for expression recognition. The system is tested on the Cohn-Kanade facial expression database and achieves a mean recognition rate of 93.23% for six universal expressions (anger, joy, sad, disgust, fear, surprise). Facial expression recognition has applications in human-computer interaction domains like online gaming.
This document presents information on face detection techniques. It discusses image segmentation as a preprocessing step for face detection. Some common segmentation methods are thresholding, edge-based segmentation, and region-based segmentation. Face detection can be classified as implicit/pattern-based or explicit/knowledge-based. Implicit methods use techniques like templates, PCA, LDA, and neural networks, while explicit methods exploit cues like color, motion, and facial features. One method discussed is human skin color-based face detection, which filters for skin-colored regions and finds facial parts within those regions. Advantages include speed and independence from training data, while disadvantages include sensitivity to lighting and accessories.
1. The document proposes a hybrid approach to facial expression recognition that combines appearance features extracted using Local Directional Number descriptors with geometric features based on distances between facial landmark points.
2. The features are classified independently using SVMs and the scores are fused at the decision level using product rule fusion to identify facial expressions in images.
3. Experiments on the CK+ and JAFFE databases show the hybrid approach achieves better recognition rates than using appearance or geometric features individually.
This document discusses using Local Binary Patterns (LBP) for facial expression recognition from images. It summarizes that LBP is used to extract features from facial regions that are divided into cells. Histograms of these LBP features are concatenated into a feature vector for each image. Support Vector Machines (SVM) are used to classify the images into six basic expressions based on these vectors, achieving better results than template matching or Linear Discriminant Analysis. The document also proposes a method for expression-invariant face recognition that classifies an input image, neutralizes it, then searches a neutral image dataset to find potential matches.
Facial Expression Recognition Using Local Binary Pattern and Support Vector M...AM Publications
Facial expression analysis is a remarkable and demanding problem, and impacts significant applications in various fields like human-computer interaction and data-driven animation. Developing an efficient facial representation from the original face images is a crucial step for achieving facial expression recognition. Facial representation based on statistical local features, Local Binary Patterns (LBP) is practically assessed. Several machine learning techniques were thoroughly observed on various databases. LBP features- which are effectual and competent for facial expression recognition are generally used by researchers Cohn Kanade is the database for present work and the programming language used is MATLAB. Firstly, face area is divided in small regions, by which histograms, Local Binary Patterns (LBP) are extracted and then concatenated into single feature vector. This feature vector outlines a well-organized representation of face and is helpful in determining the resemblance among images.
This document describes various algorithms used to build a facial emotion recognition system, including Haar cascade, HOG, Eigenfaces, and Fisherfaces. It explains how each algorithm works, such as how Haar cascade detects facial features and HOG extracts histograms of gradients. The system is trained on the CK+ dataset and uses Eigenface and Fisherface classifiers to classify emotions, achieving higher accuracy (86.54%) with Fisherfaces. It provides code snippets of key steps like cropping, resizing images, splitting data, and predicting emotions.
This document presents a literature review and proposed work plan for face recognition using a back propagation neural network. It summarizes the Viola-Jones face detection algorithm which uses Haar features and an integral image for real-time detection. The algorithm has high detection rates with low false positives. Future work will apply back propagation neural networks to extract features and recognize faces from a database of facial images in order to build a facial recognition system.
Recognition of Partially Occluded Face Using Gradientface and Local Binary Pa...Win Yu
This document proposes a method for recognizing partially occluded faces using Gradientface and Local Binary Patterns (LBP). It first detects occluded regions using a MultiLayer Perceptron classifier, then applies Gradientface preprocessing to normalize for illumination before extracting LBP features from non-occluded regions for recognition. Experiments on the AR and ORL face databases show the method can effectively recognize faces with sunglasses and scarf occlusions.
BEB801 Project. Facial expression recognition android application. Student: Alexander Fernicola. Supervisor: Professor Vinod Chandran. Describes the first steps of building an android application that can detect the users facial expression in an image or in video.
This document provides an introduction and overview of face recognition and detection. It discusses how face recognition involves identifying faces in images and can operate in verification or identification modes. Key steps in face recognition processing are discussed, including detection, alignment, feature extraction, and matching. Analysis of faces in subspaces is also covered, as are technical challenges such as variability in facial appearance and complexity of face manifolds. Neural networks, AdaBoost methods, and dealing with head rotations in detection are also outlined.
Predicting Emotions through Facial Expressions twinkle singh
This document describes a facial expression recognition system with two parts: face recognition and facial expression recognition. It discusses using principal component analysis (PCA) and linear discriminative analysis (LDA) for face recognition, and PCA to extract eigenfaces for facial expression recognition. The system first performs face detection, then extracts facial expression data and classifies the expression. MATLAB is used as the tool for its faster programming capabilities.
A comparative review of various approaches for feature extraction in Face rec...Vishnupriya T H
This document provides an overview of various approaches for feature extraction in face recognition. It discusses common feature extraction algorithms such as PCA, DCT, LDA, and ICA. PCA is aimed at data compression while ensuring no information loss. DCT transforms images from spatial to frequency domains. LDA maximizes between-class variations and minimizes within-class variations. ICA determines statistically independent variables and minimizes higher-order dependencies. The document reviews several papers comparing the performance of these algorithms individually and in combination for face recognition applications.
This document provides a literature review of various techniques for automatic facial expression recognition. It discusses approaches such as principal component analysis (PCA), linear discriminant analysis (LDA), independent component analysis (ICA), 2D PCA, global eigen approaches using color images, subpattern extended 2D PCA, multilinear image analysis, color subspace LDA, 2D Gabor filter banks, and local Gabor binary patterns. It provides a table comparing the performance and disadvantages of these different methods. Recently, tensor perceptual color frameworks have been introduced that apply tensor concepts and perceptual color spaces to improve recognition performance under varying illumination conditions.
Face Recognition Human Computer Interactionines beltaief
Face recognition is a popular area of computer vision research that involves identifying or verifying a person from a digital image. There are two main tasks: face identification which matches a face to known individuals, and face verification which confirms whether an image matches a claimed identity. Common approaches to face recognition include detecting faces, extracting facial features, and classifying expressions. Popular techniques are face detection algorithms, principal component analysis, linear discriminant analysis, kernel methods, and template matching. Applications of face recognition include automatic face tagging, gaming, price comparisons, image search, and security systems.
This document discusses facial expression recognition and the challenges that remain. It provides an overview of the current state-of-the-art techniques for facial expression recognition, which still struggle with accuracy when tested on naturalistic data rather than posed images. The document outlines a proposed pipeline for facial expression recognition that combines deep learning techniques for feature fusion and representation learning to help address these challenges and improve recognition accuracy on real-world data. Samples of datasets used for training and evaluating facial expression recognition systems are also presented.
https://telecombcn-dl.github.io/2017-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
Face recognition technology uses machine learning algorithms to identify or verify a person's identity from digital images or video frames. The process involves detecting faces, applying preprocessing techniques like filtering and scaling, training classifiers using labeled face images, and then classifying new faces. Common machine learning algorithms used include K-nearest neighbors, naive Bayes, decision trees, and locally weighted learning. The proposed system detects faces, builds a tabular dataset from pixel values, trains classifiers, and evaluates performance on a test set. Software applies techniques like detection, alignment, normalization, and matching to encode faces for comparison. Face recognition has advantages like convenience and low cost, and applications in security, banking, and more.
Face recognition across non uniform motion blur, illumination, and posePvrtechnologies Nellore
The document proposes a method for face recognition in the presence of non-uniform motion blur from hand-held cameras. It models a blurred face as a convex combination of geometrically transformed gallery images. It develops an algorithm using the assumption of sparse camera motion and an l1-norm constraint. The framework is extended to handle illumination variations by exploiting the bi-convex set of images from blurring and illumination changes. The method is also extended to account for pose variations and uses a multi-scale implementation for efficient computation and memory usage.
The document summarizes a design seminar project on human face identification. The objectives of the project were to develop a computational model for face recognition that can work under varying poses and apply it to problems like criminal identification, security, and image processing. The research methodology used eigenface methods based on information theory. The project involved developing a face identification system with features like adding images to a database, clipping images, updating details, and searching for matches. It provides screenshots of the system interface and discusses the software and hardware requirements and limitations of the approach. The conclusion states that the system can efficiently find faces without exhaustive searching and face recognition will have many applications in smart environments.
The document summarizes research on automated face detection and recognition. It discusses common applications of face detection such as webcam tracking and photo tagging. Face recognition can be used for biometrics, mugshot databases, and detecting fake IDs. The document then compares human and computer abilities in face detection/recognition and describes challenges computers face representing multidimensional face data. It provides a brief history of the field and covers common approaches to face detection and recognition including eigenfaces, Fisherfaces, neural networks, Gabor wavelets, and active shape models. The document also discusses challenges of 3D, video, and comparing face recognition systems.
This document summarizes research on deep learning approaches for face recognition. It describes the DeepFace model from Facebook, which used a deep convolutional network trained on 4.4 million faces to achieve state-of-the-art accuracy on the Labeled Faces in the Wild (LFW) dataset. It also summarizes the DeepID2 and DeepID3 models from Chinese University of Hong Kong, which employed joint identification-verification training of convolutional networks and achieved performance comparable or superior to DeepFace on LFW. Evaluation metrics for face verification and identification tasks are also outlined.
Face recognization using artificial nerual networkDharmesh Tank
This document presents an overview of face recognition using artificial neural networks. It discusses the basic concepts of face recognition, issues with existing systems, and proposes a new system using discrete cosine transform (DCT) for feature extraction and an artificial neural network with backpropagation for classification. DCT is used to extract illumination invariant features and reduce dimensionality. The neural network is trained on these features to recognize faces. Thresholding rules are also introduced to improve recognition performance. Real-time applications of face recognition like Microsoft's Project Natal are mentioned.
CDS is the criminal face identification by capsule neural network.
Solving the common problems in image recognition such as illumination problem, scale variability, and to fight against a most common problem like pose problem, we are introducing Face Reconstruction System.
Real time multi face detection using deep learningReallykul Kuul
This document proposes a framework for real-time multiple face recognition using deep learning on an embedded GPU system. The framework includes face detection using a CNN, face tracking to reduce processing time, and a state-of-the-art deep CNN for face recognition. Experimental results showed the system can recognize up to 8 faces simultaneously in real-time, with processing times up to 0.23 seconds and a minimum recognition rate of 83.67%.
4837410 automatic-facial-emotion-recognitionNgaire Taylor
This document summarizes an automatic facial emotion recognition system. It begins with an introduction to facial expression recognition and importance of understanding emotions. It then discusses related work on universal emotions and facial feature analysis. The system uses a facial tracker to extract features from tracked facial landmarks. Two classifiers, Naive Bayes and TAN, are used to classify emotions and results are visualized. The system includes a face detector for initialization and uses evaluation on recognition accuracy for different classifiers and dependency.
Facial emotion recognition uses active shape modeling to identify 5 classes of emotions from facial images. It reconstructs facial models by labeling landmark features, performing shape modeling through principal component analysis and model fitting, and classifying emotions. The methodology was tested on a set of labeled images and achieved over 80% accuracy in emotion recognition according to the results.
This document describes various algorithms used to build a facial emotion recognition system, including Haar cascade, HOG, Eigenfaces, and Fisherfaces. It explains how each algorithm works, such as how Haar cascade detects facial features and HOG extracts histograms of gradients. The system is trained on the CK+ dataset and uses Eigenface and Fisherface classifiers to classify emotions, achieving higher accuracy (86.54%) with Fisherfaces. It provides code snippets of key steps like cropping, resizing images, splitting data, and predicting emotions.
This document presents a literature review and proposed work plan for face recognition using a back propagation neural network. It summarizes the Viola-Jones face detection algorithm which uses Haar features and an integral image for real-time detection. The algorithm has high detection rates with low false positives. Future work will apply back propagation neural networks to extract features and recognize faces from a database of facial images in order to build a facial recognition system.
Recognition of Partially Occluded Face Using Gradientface and Local Binary Pa...Win Yu
This document proposes a method for recognizing partially occluded faces using Gradientface and Local Binary Patterns (LBP). It first detects occluded regions using a MultiLayer Perceptron classifier, then applies Gradientface preprocessing to normalize for illumination before extracting LBP features from non-occluded regions for recognition. Experiments on the AR and ORL face databases show the method can effectively recognize faces with sunglasses and scarf occlusions.
BEB801 Project. Facial expression recognition android application. Student: Alexander Fernicola. Supervisor: Professor Vinod Chandran. Describes the first steps of building an android application that can detect the users facial expression in an image or in video.
This document provides an introduction and overview of face recognition and detection. It discusses how face recognition involves identifying faces in images and can operate in verification or identification modes. Key steps in face recognition processing are discussed, including detection, alignment, feature extraction, and matching. Analysis of faces in subspaces is also covered, as are technical challenges such as variability in facial appearance and complexity of face manifolds. Neural networks, AdaBoost methods, and dealing with head rotations in detection are also outlined.
Predicting Emotions through Facial Expressions twinkle singh
This document describes a facial expression recognition system with two parts: face recognition and facial expression recognition. It discusses using principal component analysis (PCA) and linear discriminative analysis (LDA) for face recognition, and PCA to extract eigenfaces for facial expression recognition. The system first performs face detection, then extracts facial expression data and classifies the expression. MATLAB is used as the tool for its faster programming capabilities.
A comparative review of various approaches for feature extraction in Face rec...Vishnupriya T H
This document provides an overview of various approaches for feature extraction in face recognition. It discusses common feature extraction algorithms such as PCA, DCT, LDA, and ICA. PCA is aimed at data compression while ensuring no information loss. DCT transforms images from spatial to frequency domains. LDA maximizes between-class variations and minimizes within-class variations. ICA determines statistically independent variables and minimizes higher-order dependencies. The document reviews several papers comparing the performance of these algorithms individually and in combination for face recognition applications.
This document provides a literature review of various techniques for automatic facial expression recognition. It discusses approaches such as principal component analysis (PCA), linear discriminant analysis (LDA), independent component analysis (ICA), 2D PCA, global eigen approaches using color images, subpattern extended 2D PCA, multilinear image analysis, color subspace LDA, 2D Gabor filter banks, and local Gabor binary patterns. It provides a table comparing the performance and disadvantages of these different methods. Recently, tensor perceptual color frameworks have been introduced that apply tensor concepts and perceptual color spaces to improve recognition performance under varying illumination conditions.
Face Recognition Human Computer Interactionines beltaief
Face recognition is a popular area of computer vision research that involves identifying or verifying a person from a digital image. There are two main tasks: face identification which matches a face to known individuals, and face verification which confirms whether an image matches a claimed identity. Common approaches to face recognition include detecting faces, extracting facial features, and classifying expressions. Popular techniques are face detection algorithms, principal component analysis, linear discriminant analysis, kernel methods, and template matching. Applications of face recognition include automatic face tagging, gaming, price comparisons, image search, and security systems.
This document discusses facial expression recognition and the challenges that remain. It provides an overview of the current state-of-the-art techniques for facial expression recognition, which still struggle with accuracy when tested on naturalistic data rather than posed images. The document outlines a proposed pipeline for facial expression recognition that combines deep learning techniques for feature fusion and representation learning to help address these challenges and improve recognition accuracy on real-world data. Samples of datasets used for training and evaluating facial expression recognition systems are also presented.
https://telecombcn-dl.github.io/2017-dlcv/
Deep learning technologies are at the core of the current revolution in artificial intelligence for multimedia data analysis. The convergence of large-scale annotated datasets and affordable GPU hardware has allowed the training of neural networks for data analysis tasks which were previously addressed with hand-crafted features. Architectures such as convolutional neural networks, recurrent neural networks and Q-nets for reinforcement learning have shaped a brand new scenario in signal processing. This course will cover the basic principles and applications of deep learning to computer vision problems, such as image classification, object detection or image captioning.
Face recognition technology uses machine learning algorithms to identify or verify a person's identity from digital images or video frames. The process involves detecting faces, applying preprocessing techniques like filtering and scaling, training classifiers using labeled face images, and then classifying new faces. Common machine learning algorithms used include K-nearest neighbors, naive Bayes, decision trees, and locally weighted learning. The proposed system detects faces, builds a tabular dataset from pixel values, trains classifiers, and evaluates performance on a test set. Software applies techniques like detection, alignment, normalization, and matching to encode faces for comparison. Face recognition has advantages like convenience and low cost, and applications in security, banking, and more.
Face recognition across non uniform motion blur, illumination, and posePvrtechnologies Nellore
The document proposes a method for face recognition in the presence of non-uniform motion blur from hand-held cameras. It models a blurred face as a convex combination of geometrically transformed gallery images. It develops an algorithm using the assumption of sparse camera motion and an l1-norm constraint. The framework is extended to handle illumination variations by exploiting the bi-convex set of images from blurring and illumination changes. The method is also extended to account for pose variations and uses a multi-scale implementation for efficient computation and memory usage.
The document summarizes a design seminar project on human face identification. The objectives of the project were to develop a computational model for face recognition that can work under varying poses and apply it to problems like criminal identification, security, and image processing. The research methodology used eigenface methods based on information theory. The project involved developing a face identification system with features like adding images to a database, clipping images, updating details, and searching for matches. It provides screenshots of the system interface and discusses the software and hardware requirements and limitations of the approach. The conclusion states that the system can efficiently find faces without exhaustive searching and face recognition will have many applications in smart environments.
The document summarizes research on automated face detection and recognition. It discusses common applications of face detection such as webcam tracking and photo tagging. Face recognition can be used for biometrics, mugshot databases, and detecting fake IDs. The document then compares human and computer abilities in face detection/recognition and describes challenges computers face representing multidimensional face data. It provides a brief history of the field and covers common approaches to face detection and recognition including eigenfaces, Fisherfaces, neural networks, Gabor wavelets, and active shape models. The document also discusses challenges of 3D, video, and comparing face recognition systems.
This document summarizes research on deep learning approaches for face recognition. It describes the DeepFace model from Facebook, which used a deep convolutional network trained on 4.4 million faces to achieve state-of-the-art accuracy on the Labeled Faces in the Wild (LFW) dataset. It also summarizes the DeepID2 and DeepID3 models from Chinese University of Hong Kong, which employed joint identification-verification training of convolutional networks and achieved performance comparable or superior to DeepFace on LFW. Evaluation metrics for face verification and identification tasks are also outlined.
Face recognization using artificial nerual networkDharmesh Tank
This document presents an overview of face recognition using artificial neural networks. It discusses the basic concepts of face recognition, issues with existing systems, and proposes a new system using discrete cosine transform (DCT) for feature extraction and an artificial neural network with backpropagation for classification. DCT is used to extract illumination invariant features and reduce dimensionality. The neural network is trained on these features to recognize faces. Thresholding rules are also introduced to improve recognition performance. Real-time applications of face recognition like Microsoft's Project Natal are mentioned.
CDS is the criminal face identification by capsule neural network.
Solving the common problems in image recognition such as illumination problem, scale variability, and to fight against a most common problem like pose problem, we are introducing Face Reconstruction System.
Real time multi face detection using deep learningReallykul Kuul
This document proposes a framework for real-time multiple face recognition using deep learning on an embedded GPU system. The framework includes face detection using a CNN, face tracking to reduce processing time, and a state-of-the-art deep CNN for face recognition. Experimental results showed the system can recognize up to 8 faces simultaneously in real-time, with processing times up to 0.23 seconds and a minimum recognition rate of 83.67%.
4837410 automatic-facial-emotion-recognitionNgaire Taylor
This document summarizes an automatic facial emotion recognition system. It begins with an introduction to facial expression recognition and importance of understanding emotions. It then discusses related work on universal emotions and facial feature analysis. The system uses a facial tracker to extract features from tracked facial landmarks. Two classifiers, Naive Bayes and TAN, are used to classify emotions and results are visualized. The system includes a face detector for initialization and uses evaluation on recognition accuracy for different classifiers and dependency.
Facial emotion recognition uses active shape modeling to identify 5 classes of emotions from facial images. It reconstructs facial models by labeling landmark features, performing shape modeling through principal component analysis and model fitting, and classifying emotions. The methodology was tested on a set of labeled images and achieved over 80% accuracy in emotion recognition according to the results.
The document discusses subspace indexing on Grassmannian manifolds for large scale visual identification. It proposes using local subspace models built on neighborhoods defined by queries, but notes issues with computational complexity and lack of optimality. It then introduces Grassmannian and Stiefel manifolds to characterize subspace similarity and define distances. A model hierarchical tree is proposed to index subspaces through iterative merging based on distances on the Grassmannian manifold.
The document discusses research into facial expressions in both humans and dogs. It describes Charles Darwin's theory that facial expressions are innate and evolved to communicate emotions. Recent work supports this theory. The Facial Action Coding System was developed to categorize facial movements based on the contracting muscles. Six universal human emotions - happiness, sadness, surprise, fear, disgust, and anger - each correspond to distinct facial expressions involving specific muscles. Additional research found that dogs also use facial expressions like ear positioning to communicate emotions when seeing their owner, a stranger, or disliked objects.
The document summarizes a student project that aims to identify 5 classes of emotions - neutral, joy, sadness, surprise, and anger - from facial images using active shape modeling. It does this by labeling landmark features, generating reference and mean models, selecting features for each emotion, and classifying emotions. The results are presented in a confusion matrix and future work is outlined to improve the model using texture, better classifiers, semi-supervised learning, and real-time applications.
Face Recognition for Personal Photos using Online Social Network Context and ...Wesley De Neve
Thanks to easy-to-use multimedia devices and cheap storage and bandwidth, present-day social media applications host staggering numbers of personal photos. As the number of personal photos shared on social media applications continues to accelerate, the problem of organizing and retrieving relevant photos becomes more apparent for consumers. Automatic face recognition assists in bringing order to collections of personal photos. However, personal photos pose a plethora of challenges for automatic face recognition. Face images may widely differ in terms of lighting, expressions, and pose. As a result, the accuracy of appearance-based techniques for automatic face recognition in collections of personal photos cannot be considered satisfactory.
This talk aims at providing insight into timely developments in the area of socially-aware face recognition. We first discuss how online social network context can be used to substantially improve the effectiveness of appearance-based techniques for automatic face recognition, as recently demonstrated by researchers of Harvard University. Next, we pay attention to collaborative face recognition in decentralized online social networks, as studied at KAIST. For both of the aforementioned topics, we present experimental results obtained for real-world collections of personal photos, contributed by volunteers who are members of online social networks such as Facebook and Cyworld. Finally, we conclude our talk with an outline of future applications of socially-aware face recognition, including augmented identity and socially-aware robots.
Irene Andersen is a female bodybuilder from Sweden who began bodybuilding in the late 1970s. She was born in Denmark in 1966 and moved to Sweden at age 2. She has trained in various sports like jazz ballet, judo, and bodybuilding since age 15. Between 1990-1996 she took breaks from training to have children but returned to the gym when they were older. She began competing in bodybuilding in 2003 at age 37 and obtained her pro card in 2005.
Academic Entrepreneurship at UCY,
by Mr. Christis Christoforou, MBA principal for accelyservices.
The results and the methodoloty of an extensive survey that were conducted at the university of Cyprus will be presented.
This document discusses concepts and solutions for a face recognition tool. It covers collecting face images, managing the image files, using the tool for staff attendance recording, and potential solutions for the tool including a web interface with server, macOS app, or command line only. It also provides examples of naming conventions for image files and discusses checks to validate captured images.
Deformable Facial Models and 3D Face Reconstruction Methods: A surveyLakshmi Sarvani Videla
Deformable Facial Model Construction for non-rigid motion tracking, 3D Face Reconstruction Methods, Geometry-Based Methods , Stereo methods, Shape from Motion models, Face Models, Cylindrical Model, Ellipsoidal Model, Planar Model
, Facial deformable models, Holistic models, Part based models, Eigenfaces, Active Shape Models, Combined Appearance Models, comparison of 3D facial features,list of 3d face databases containing 3D static expressions
Real-time Face Recognition & Detection Systems 1Suvadip Shome
The document discusses face detection and recognition. It begins with definitions of face detection as identifying faces in images and locating them, as well as distinguishing faces from non-faces. Face recognition is then defined as identifying known faces versus unknown faces. The document outlines the difference between detection and recognition. It then discusses various methods for face detection, including knowledge-based, feature-based, template matching and appearance-based methods. Finally, it provides an overview of the principal component analysis (PCA) algorithm for face recognition, outlining the main steps of converting images to vectors, normalizing, calculating eigenvectors, selecting eigenfaces, and representing faces as combinations of eigenvectors.
KaoNet: Face Recognition and Generation App using Deep LearningVan Huy
KaoNet is a face recognition and generation app using deep learning. It uses convolutional neural networks (CNNs) for face recognition and generative adversarial networks (GANs) for face generation. The app was trained on a dataset of celebrity faces collected from online sources. Initial results for face recognition were poor due to overfitting and limited data. Expanding the dataset improved validation accuracy to 98%. The GAN was also able to generate realistic looking faces after training.
Face recognition and deep learning โดย ดร. สรรพฤทธิ์ มฤคทัต NECTECBAINIDA
Face recognition and deep learning โดย ดร. สรรพฤทธิ์ มฤคทัต NECTEC
คณะสถิติประยุกต์ สถาบันบัณฑิตพัฒนบริหารศาสตร์ ร่วมกับ Data Science Thailand ร่วมกันจัดงาน The First NIDA Business Analytics and Data Sciences Contest/Conference
Junyu Tech is a Chinese company that develops face recognition software and related SDKs. The document describes Junyu Tech's face recognition principles, performance testing results, applications, and additional modules. It also provides details on Junyu Tech's patents, professional team, and participation in tech shows.
This document discusses face recognition technology. It begins by outlining the topics that will be covered, which include why FRT is used, how it is implemented, applications, opposition to it, and its future. It then describes the accuracy and passive identification capabilities of FRT. It explains the verification and identification processes and the components involved, including the enrollment module, database, and identification module. It outlines the five step process of acquiring an image, locating the face, analyzing the facial image, comparing it to stored images, and determining a match or no match. Finally, it discusses some applications and opposition to FRT and envisions its future advancement and integration.
This document discusses manifolds and kernels on manifolds. It defines manifolds and different types of manifolds such as topological manifolds, differentiable manifolds, and Riemannian manifolds. It then discusses Hilbert spaces, kernels, and reproducing kernel Hilbert spaces. It explains that defining kernels on manifolds allows applying kernel methods to nonlinear manifolds. It discusses challenges in defining positive definite kernels on manifolds using geodesic distance and provides conditions for when the Gaussian RBF kernel is positive definite on a manifold. It also covers applications to pedestrian detection and visual object categorization using kernels on the manifold of symmetric positive definite matrices.
Humans often use faces to recognize individuals, and advancements in computing capability over the past few decades now enable similar recognitions automatically. Early facial recognition algorithms used simple geometric models, but the recognition process has now matured into a science of sophisticated mathematical representations and matching processes. Major advancements and initiatives in the past 10 to 15 years have propelled facial recognition technology into the spotlight. Facial recognition can be used for both verification and identification.
Diffusion Deformable Model for 4D Temporal Medical Image GenerationBoahKim2
This document describes a diffusion deformable model for generating 4D temporal medical images. The model uses a diffusion probabilistic model combined with a registration model to generate intermediate deformed images along a continuous trajectory between a source and target image. The model was tested on cardiac MRI data and shown to outperform existing deformation models in generating dynamic deformations and intermediate frames, as measured by quantitative metrics and qualitative evaluation. The approach provides a promising new tool for analyzing changes in anatomical structures over time.
1) Computer vision techniques such as stereo vision and object detection can be used for applications in robotics.
2) Stereo vision involves calculating depth from two images using techniques like block matching and disparity maps. It can be implemented efficiently on GPUs.
3) Object detection methods include feature detection using SIFT or SURF descriptors followed by matching and geometry validation using techniques like RANSAC. These allow objects to be detected and their pose estimated.
This document describes a proposed multimodal biometric authentication system using face, fingerprint, palm print, and palm vein modalities. It provides background on each biometric, describes related work fusing different modalities, and outlines the proposed system. The system will extract features from each biometric, fuse the features at the level of extraction, and calculate distance metrics to perform authentication. Preliminary results show recognition rates over 98% for 50-500 people when fusing modalities.
Montage4D: Interactive Seamless Fusion of Multiview Video TexturesRuofei Du
Project Site: http://montage4d.com
The commoditization of virtual and augmented reality devices and the availability of inexpensive consumer depth cameras have catalyzed a resurgence of interest in spatiotemporal performance capture. Recent systems like Fusion4D and Holoportation address several crucial problems in the real-time fusion of multiview depth maps into volumetric and deformable representations. Nonetheless, stitching multiview video textures onto dynamic meshes remains challenging due to imprecise geometries, occlusion seams, and critical time constraints. In this paper, we present a practical solution towards real-time seamless texture montage for dynamic multiview reconstruction. We build on the ideas of dilated depth discontinuities and majority voting from Holoportation to reduce ghosting effects when blending textures. In contrast to their approach, we determine the appropriate blend of textures per vertex using view-dependent rendering techniques, so as to avert fuzziness caused by the ubiquitous normal-weighted blending. By leveraging geodesics-guided diffusion and temporal texture fields, our algorithm mitigates spatial occlusion seams while preserving temporal consistency. Experiments demonstrate significant enhancement in rendering quality, especially in detailed regions such as faces. We envision a wide range of applications for Montage4D, including immersive telepresence for business, training, and live entertainment.
This document provides a synopsis for a project on emotion detection from facial expressions. It outlines the objectives to develop an automatic emotion detection system using machine learning algorithms to analyze facial expressions in video frames and compare them to a database to classify emotions. The technical details discuss using a facial tracker and extracting features to represent expressions. Classification algorithms like KNN, SVM, and voting will be used for recognition and mapping expressions to emotions. Future work may include 3D processing, speech recognition, and detecting micro-expressions.
In this paper the process of 3D modelling from video is presented. Analysed previous research related to
this process, and specifically described algorithms for detecting and matching key points. We described
their advantages and disadvantages, and made a critical analysis of algorithms. In this paper, the three
detectors (SUSAN, Plessey and Förstner) are tested and compare. We used video taken with hand held
camera of a cube and compare these detectors on it (taking into account their parameters of accuracy and
repeatability). In conclusion, we practically made 3D model of the cube from video used these detectors in
the first step of the process and three algorithms (RANSAC, MSAC and MLESAC) for matching data.
This document summarizes an international journal article that proposes a two-phase algorithm for face recognition in the frequency domain using discrete cosine transform (DCT) and discrete Fourier transform (DFT). The algorithm works in two phases: the first phase uses Euclidean distance to determine the K nearest neighbor training samples of a test sample. The second phase represents the test sample as a linear combination of the K nearest neighbors and classifies the sample based on which class representation has the smallest deviation from the test sample. Experimental results on FERET and ORL face databases show the two-phase algorithm based on DCT and DFT outperforms other methods like two-phase sparse representation and PCA/LDA in terms of classification accuracy.
This document summarizes 10 research papers on various techniques for facial expression recognition. The papers cover topics like using local gray code patterns and kernel canonical correlation analysis to extract facial features and recognize expressions. Other techniques discussed include using facial animation parameters and hidden Markov models, active appearance models to track facial features over video sequences, and using geometric deformation features and support vector machines to recognize expressions in image sequences. The document provides an overview of the different approaches researchers have taken and their relative performances on standard datasets.
Movement Tracking in Real-time Hand Gesture RecognitionPranav Kulkarni
To translate the gesture performed by the user in a
video sequence into meaningful symbols/commands, feature
extraction is the first and most crucial step in such systems
which measures the detected hand positions and its movement
track. We propose an efficient approach based on inter-frame
difference (IDF) to handle the hand movement tracking, which
is shown to be more robust in the accuracy aspect compared to
skin-color based approaches. Computational efficiency is
another attractive property that our approach greatly
improves the processing frame rate to fulfil the demand of a
real-time hand gesture recognition system.
This document discusses the history and fundamentals of visual odometry (VO) and simultaneous localization and mapping (SLAM). It provides an overview of key developments in VO from the 1980s to present day, including the first real-time VO implementation on a robot in 1980 and use of VO on the Mars rovers in 2004. The document also summarizes the differences between VO, SFM, and V-SLAM, and describes common approaches to feature extraction, motion estimation, and optimization in VO pipelines.
Lip Reading by Using 3-D Discrete Wavelet Transform with Dmey WaveletCSCJournals
Lip movement is an useful way to communicate with machines and it is extremely helpful in noisy environments. However, the recognition of lip motion is a difficult task since the region of interest (ROI) is nonlinear and noisy. In the proposed lip reading method we have used two stage feature extraction mechanism which is précised, discriminative and computation efficient. The first stage is to convert video frame data into 3 dimension space and the second stage trims down the raw information space by using 3 Dimension Discrete Wavelet Transform (DWT). These features are smaller in size to give rise a novel lip reading system. In addition to the novel feature extraction technique, we have also compared the performance of Back Propagation Neural Network (BPNN) and Support Vector Machine(SVM) classifier. CUAVE database and Tulips database are used for experimentation. Experimental results show that 3-D DWT feature mining is better than 2-D DWT. 3-D DWT with Dmey wavelet results are better than 3-D DWT Db4. Results of experimentation show that 3-D DWT-Dmey along with BNNN classifier outperforms SVM.
This document summarizes two papers presented at the SIGGRAPH 2014 conference. The first paper proposes a parametric wave field coding technique to precompute and compress sound propagation simulations in complex 3D environments. It represents impulse responses using four perceptual parameters that can be interpolated over space. The second paper describes an interactive algorithm for simulating higher-order diffraction and diffuse reflections in large dynamic scenes using ray tracing. It reuses ray paths over time and employs edge culling and visibility graphs to improve performance.
This doctoral dissertation examines facial skin motion properties from video for modeling and applications. It presents two methods for computing strain patterns from video: a finite difference method and a finite element method. The finite element method incorporates material properties of facial tissues by modeling their Young's modulus values. Experiments show strain patterns are discriminative and stable features for facial expression recognition, age estimation, and person identification. The dissertation also develops a method for expression invariant face matching by modeling Young's modulus from multiple expressions.
The document describes a project to develop a gender voice recognition system using machine learning. It aims to achieve higher accuracy than existing MLP models. The proposed system uses logistic regression and fast Fourier transform for noise cancellation. It achieves 96.74% accuracy on test data, higher than existing systems. The document outlines the aim, abstract, introduction, literature review on existing approaches, proposed system description using algorithms like logistic regression and FFT, requirements, UML diagrams, advantages of automatic gender recognition, limitations, output, references, and conclusions.
The document discusses continuous human action recognition in ambient assisted living scenarios. It proposes an approach that uses action zones, which are the most discriminative segments of an action, rather than whole action sequences. Action zones are learned from training data and used to recognize actions in continuous video streams using a sliding window approach. The recognition thresholds and parameters are optimized using an evolutionary algorithm to maximize recognition performance. The approach is validated on public datasets and aims to enable long-term continuous human behavior analysis in ambient assisted living environments.
Ioannis Pitas, Professor, Aristotle University of Thessaloniki, Department of Informatics (IEEE Fellow), Semantic 3DTV Content Analysis and Description
https://telecombcn-dl.github.io/2017-dlsl/
Winter School on Deep Learning for Speech and Language. UPC BarcelonaTech ETSETB TelecomBCN.
The aim of this course is to train students in methods of deep learning for speech and language. Recurrent Neural Networks (RNN) will be presented and analyzed in detail to understand the potential of these state of the art tools for time series processing. Engineering tips and scalability issues will be addressed to solve tasks such as machine translation, speech recognition, speech synthesis or question answering. Hands-on sessions will provide development skills so that attendees can become competent in contemporary data analytics tools.
Similar to 3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection (20)
Rainfall intensity duration frequency curve statistical analysis and modeling...bijceesjournal
Using data from 41 years in Patna’ India’ the study’s goal is to analyze the trends of how often it rains on a weekly, seasonal, and annual basis (1981−2020). First, utilizing the intensity-duration-frequency (IDF) curve and the relationship by statistically analyzing rainfall’ the historical rainfall data set for Patna’ India’ during a 41 year period (1981−2020), was evaluated for its quality. Changes in the hydrologic cycle as a result of increased greenhouse gas emissions are expected to induce variations in the intensity, length, and frequency of precipitation events. One strategy to lessen vulnerability is to quantify probable changes and adapt to them. Techniques such as log-normal, normal, and Gumbel are used (EV-I). Distributions were created with durations of 1, 2, 3, 6, and 24 h and return times of 2, 5, 10, 25, and 100 years. There were also mathematical correlations discovered between rainfall and recurrence interval.
Findings: Based on findings, the Gumbel approach produced the highest intensity values, whereas the other approaches produced values that were close to each other. The data indicates that 461.9 mm of rain fell during the monsoon season’s 301st week. However, it was found that the 29th week had the greatest average rainfall, 92.6 mm. With 952.6 mm on average, the monsoon season saw the highest rainfall. Calculations revealed that the yearly rainfall averaged 1171.1 mm. Using Weibull’s method, the study was subsequently expanded to examine rainfall distribution at different recurrence intervals of 2, 5, 10, and 25 years. Rainfall and recurrence interval mathematical correlations were also developed. Further regression analysis revealed that short wave irrigation, wind direction, wind speed, pressure, relative humidity, and temperature all had a substantial influence on rainfall.
Originality and value: The results of the rainfall IDF curves can provide useful information to policymakers in making appropriate decisions in managing and minimizing floods in the study area.
Embedded machine learning-based road conditions and driving behavior monitoringIJECEIAES
Car accident rates have increased in recent years, resulting in losses in human lives, properties, and other financial costs. An embedded machine learning-based system is developed to address this critical issue. The system can monitor road conditions, detect driving patterns, and identify aggressive driving behaviors. The system is based on neural networks trained on a comprehensive dataset of driving events, driving styles, and road conditions. The system effectively detects potential risks and helps mitigate the frequency and impact of accidents. The primary goal is to ensure the safety of drivers and vehicles. Collecting data involved gathering information on three key road events: normal street and normal drive, speed bumps, circular yellow speed bumps, and three aggressive driving actions: sudden start, sudden stop, and sudden entry. The gathered data is processed and analyzed using a machine learning system designed for limited power and memory devices. The developed system resulted in 91.9% accuracy, 93.6% precision, and 92% recall. The achieved inference time on an Arduino Nano 33 BLE Sense with a 32-bit CPU running at 64 MHz is 34 ms and requires 2.6 kB peak RAM and 139.9 kB program flash memory, making it suitable for resource-constrained embedded systems.
Advanced control scheme of doubly fed induction generator for wind turbine us...IJECEIAES
This paper describes a speed control device for generating electrical energy on an electricity network based on the doubly fed induction generator (DFIG) used for wind power conversion systems. At first, a double-fed induction generator model was constructed. A control law is formulated to govern the flow of energy between the stator of a DFIG and the energy network using three types of controllers: proportional integral (PI), sliding mode controller (SMC) and second order sliding mode controller (SOSMC). Their different results in terms of power reference tracking, reaction to unexpected speed fluctuations, sensitivity to perturbations, and resilience against machine parameter alterations are compared. MATLAB/Simulink was used to conduct the simulations for the preceding study. Multiple simulations have shown very satisfying results, and the investigations demonstrate the efficacy and power-enhancing capabilities of the suggested control system.
Applications of artificial Intelligence in Mechanical Engineering.pdfAtif Razi
Historically, mechanical engineering has relied heavily on human expertise and empirical methods to solve complex problems. With the introduction of computer-aided design (CAD) and finite element analysis (FEA), the field took its first steps towards digitization. These tools allowed engineers to simulate and analyze mechanical systems with greater accuracy and efficiency. However, the sheer volume of data generated by modern engineering systems and the increasing complexity of these systems have necessitated more advanced analytical tools, paving the way for AI.
AI offers the capability to process vast amounts of data, identify patterns, and make predictions with a level of speed and accuracy unattainable by traditional methods. This has profound implications for mechanical engineering, enabling more efficient design processes, predictive maintenance strategies, and optimized manufacturing operations. AI-driven tools can learn from historical data, adapt to new information, and continuously improve their performance, making them invaluable in tackling the multifaceted challenges of modern mechanical engineering.
Optimizing Gradle Builds - Gradle DPE Tour Berlin 2024Sinan KOZAK
Sinan from the Delivery Hero mobile infrastructure engineering team shares a deep dive into performance acceleration with Gradle build cache optimizations. Sinan shares their journey into solving complex build-cache problems that affect Gradle builds. By understanding the challenges and solutions found in our journey, we aim to demonstrate the possibilities for faster builds. The case study reveals how overlapping outputs and cache misconfigurations led to significant increases in build times, especially as the project scaled up with numerous modules using Paparazzi tests. The journey from diagnosing to defeating cache issues offers invaluable lessons on maintaining cache integrity without sacrificing functionality.
artificial intelligence and data science contents.pptxGauravCar
What is artificial intelligence? Artificial intelligence is the ability of a computer or computer-controlled robot to perform tasks that are commonly associated with the intellectual processes characteristic of humans, such as the ability to reason.
› ...
Artificial intelligence (AI) | Definitio
An improved modulation technique suitable for a three level flying capacitor ...IJECEIAES
This research paper introduces an innovative modulation technique for controlling a 3-level flying capacitor multilevel inverter (FCMLI), aiming to streamline the modulation process in contrast to conventional methods. The proposed
simplified modulation technique paves the way for more straightforward and
efficient control of multilevel inverters, enabling their widespread adoption and
integration into modern power electronic systems. Through the amalgamation of
sinusoidal pulse width modulation (SPWM) with a high-frequency square wave
pulse, this controlling technique attains energy equilibrium across the coupling
capacitor. The modulation scheme incorporates a simplified switching pattern
and a decreased count of voltage references, thereby simplifying the control
algorithm.
Introduction- e - waste – definition - sources of e-waste– hazardous substances in e-waste - effects of e-waste on environment and human health- need for e-waste management– e-waste handling rules - waste minimization techniques for managing e-waste – recycling of e-waste - disposal treatment methods of e- waste – mechanism of extraction of precious metal from leaching solution-global Scenario of E-waste – E-waste in India- case studies.
IEEE Aerospace and Electronic Systems Society as a Graduate Student Member
3D Dynamic Facial Sequences Analsysis for face recognition and emotion detection
1. 3D DYNAMIC FACIAL SEQUENCES ANALYSIS
FOR FACE RECOGNITION AND EMOTION
DETECTION
PhD Candidate: Taleb ALASHKAR
Supervisor: Prof. Mohamed DAOUDI
Co-Supervisor: Dr. Boulbaba BEN AMOR
1
Taleb ALASHKAR PhD Defense 2-Nov-2015
2. WHY FACE ANALYSIS?
?
DB
ID: 15081986
Identity Recognition
Facial Expressions
29252212
Age Estimation
Physical State Monitoring
FatiguePain
AngrySurprisedHappy
WHY 3D FACE?
Illumination
Pose
3D
2D
WHY 3D DYNAMIC?
3D static
3D Dynamic
vs
vs
Year
….
22
3. MOTIVATION AND CHALLENGES
Motivation to 4D (3D+t) Face Analysis
Robustness to illumination changes and pose variations
Availability of cost-effective (Kinect-like) and high
resolution (Di4D) 3D dynamic sensors
Richness in shape and deformation
Challenges of 4D Face Analysis
Noisy data (from acquisition and sensor accuracy)
Missing data (single-view scanners)
Volume of data (sequence of 3D meshes)
Low-resolution frames (Kinect-like sensors)
Compact spatio-temporal representation robust to
noise and missing data wich allows 4D face analysis
3
5. OUTLINE
4D Face Recognition
State of the art
4D face recognition framework
Experiments and results
4D Spontaneous Emotion Detection
State of the art
Trajectories on Grassmann manifold
Spontaneous emotional state detection from depth video
Spontaneous pain detection from 4D high resolution video
Conclusion and Future Work
5
6. OUTLINE
4D Face Recognition
State of the art
4D face recognition framework
Experiments and results
4D Spontaneous Emotion Detection
State of the art
Trajectories on Grassmann manifold
Spontaneous emotional state detection from depth video
Spontaneous pain detection from 4D high resolution video
Conclusion and Future Work
6
7. FACE RECOGNITION FROM 4D DATA
State of the Art
Frame SetSuper ResolutionSpatio-Temporal
Low resolution (Kinect)
Illumination/FE
Temporal information
Complex enrollment
(Lie et al., 2013)
Low resolution (Kinect)
Constant expression
Temporal information
3D frames alignment
(Berretti et al., 2014)(Sun et al., 2010)
One Kinect
frame
3D HR
scanner 77
Outperforms 2D video/3D static
Space-time representation
Time consuming
Tracking/model adaptation/
conformal mapping/ST HMM
8. 4D FACE RECOGNITION
4D Face Recognition Approach
Data
processing
Training
Test
time
Modeling
Identity
Subspace
Modeling (k-SVD)
Curvature-maps
Extraction
time
Mean
Curvature
Computation
time
?
= Span{ , ,...., }
Dictionary
Classification
Grassmann
Dictionary
Learning
Grassmann
SRC Sparse
Coding
time
time
8
9. 4D FACE RECOGNITION
Where K1 and K2 is the two principle curvatures at each vertex.
Spatial Feature Extraction
Capture the local facial shape
Invariant to the scale, rotation and mesh resolution
Ability to capture the non-rigid facial deformation 9
10. 4D FACE RECOGNITION
Spatio-Temporal Subspace Representation
n×m
k-SVD
3D dynamic original Data
Matrix manifold
Curvature Map
Reshape
Subspace
Representation
k < m
Compact low dimensional representation
Robust against noise and missing data
Availability of geometric statistical tools
Why subspace representation?
1 2 ………... k …. m
10
11. 4D FACE RECOGNITION
Matrix Manifolds1:
Stiefel manifold
All possible k-dimensional subspaces n-dimensional space.
Defined distance on Stiefel given by:
Grassmann manifold
It is a quotient space of the Stiefel manifold with an equivalence
constraint:
X=Y if Span(X)=Span(Y), or
Exist orthonormal k×k matrix SO(k) w.r.t X=Y*SO(k)
[1]. P.A Absil et al,. “Optimization algorithms on matrix manifolds”, 2008.
11
12. 4D FACE RECOGNITION
Grassmann Manifold Geometry
Non-linear manifold
A tangent space can be defined at any point on
the manifold.
Algorithmic tools to compute the Log and Exp
maps functions.
Distances on Grassmann1
Canonical Correlation/Principle Angles1
Geodesic Distance:
[1]. Hamm et al., ICML, 2008.
12
13. 4D FACE RECOGNITION
Statistical Analysis on Grassmann manifold:
p1
p2
Log
Exp
v1
v2Tµ
µ
v3
p3
v
Intrinsic methods
Grassmann Nearest Neighbor (GNN) Classification
Training:
1. Compute karcher mean1 for every subject
in the training data (Gallery).
Testing:
2. Compare the probe with the mean of each
class using one defined distance on
Grassmann.
3. The closest mean to the probe gives the
targeted subject.
[1]. H. Karcher. PAAM, 1977.
1313
14. 4D FACE RECOGNITION
Statistical Analysis on Grassmann manifold:
14
Extrinsic methods1
Grassmann manifold embedding into linear space
Less computational time than Intrinsic
Projection mapping
[1]. Harandi et al., CVPR, 2013
14
15. 4D FACE RECOGNITION
Sparse Coding and Dictionary Learning
Suitable for data with sparse structure
Learning over-complete rich dictionary
Robustness against noise and missing data
Efficient Sparse Representation Classifier (SRC)1
[1]. Wright et al. ,PAMI, 2009
15
16. 4D FACE RECOGNITION
Experimental Results
Database
Bu4DFE Database1
101 subjects / 6 expressions (sequence) per subject
About 100 frames per sequence
Experimental Protocol (Sun et al, 2010)
60 subjects / sub-sequence of size w=6 / shifting step 3
Expression Dependent (ED): ½ of the expression training , ½ testing
Expression Independent (EI): 1 expression training, 5 testing
[1]. Yin et al. , FG, 2008
16
17. I. Grassmann Nearest Neighbor (GNN) classifier (w=6)
ED performance are better
than EI results
GNN is based on the mean
for each class (statistical
summary).
4D FACE RECOGNITION
Experimental Results
II. Grassmann Sparse Representation (GSR) classifier (w=6, EI)
Consider the face dynamics improves the recognition performance
in Expression Independent Dictionary representation
3.1%
17
18. 4D FACE RECOGNITION
18
GSR > GGDA (variant of the GDA)
GSR < Sun et al. (10%) but
- it is computationally much less expensive
- Landmarks free
Experimental Results
III. Grassmann sparse representation (GSR) classifier
GGDA is a variant of Grassmann Discriminant Analysis (proposed in [1]. Harandi et al. , CVPR, 2011.
ST-HMM is the 4D FR approach propose in [2]. Sun et al., IEEE T-Cybernetics, 2010.
Robustness to the temporal
evolution (neutral-apex or
apex-neutral)
1
2
Expression DependentExpression Independent
18
1
19. 4D FACE RECOGNITION
Expression Independent
Training by 1 vs. training by 5 expressions
- Richness of the dictionary learned
- The sparse representation (code) of a new observation can be
covered efficiently from available atoms (except for surprise)
Experimental Results
IV. Grassmann Sparse Representation (GSR) classifier
9.2%
19
20. OUTLINE
4D Face Recognition
State of the art
4D face recognition framework
Experiments and results
4D Spontaneous Emotion Detection
State of the art
Trajectories on Grassmann manifold
Spontaneous emotional state detection from depth video
Spontaneous pain detection from 4D high resolution video
Conclusion and Future Work
20
21. 4D SPONTANEOUS EMOTION DETECTION
Objectives
Proposing early detection framework for
spontaneous emotion from 3D dynamic
sequences in a continuous emotions space.
Challenges:
Spontaneous emotion of interest detection
Early emotion detection as soon as possible
3D (depth/high resolution) video
Arousal-valence chart
21
22. 4D SPONTANEOUS EMOTION DETECTION
22
3D Facial Deformation 3D Feature Tracking
State of the Art
Global deformation
Subtle changes
Nose tip
Acted FE
(Ben Amor et al., 2014)
Non-Rigid
Deformation
Parameterization facial
Deformation
Local Spatial
Feature Tracking
Landmarks
Tracking
Global deformation
Automatic
Time consuming
Acted FE
(Sandbach et al., 2011)
Robust to Noise
Real time
landmarks tracking
Acted FE
(Berretti et al., 2012)
Robust to noise
Fast performance
landmarks tracking
Acted FE
(Xue et al., 2015)
22
23. 4D SPONTANEOUS EMOTION DETECTION
Trajectories analysis on matrix manifold approach
Dividing the 3D video into subsequences
Subspace representation for each subsequence
Time parameterized trajectories on Matrix manifold
Temporal evolution through trajectory is computed
SO-SVM early event classifier applied
23
24. 4D SPONTANEOUS EMOTION DETECTION
Spontaneous emotion detection from depth video
- Upper part of the body vs. the face only
Face vs. Face+ Upper Part
- Depth vs. 2D video data
24
25. 4D SPONTANEOUS EMOTION DETECTION
Spontaneous emotion detection from depth video
Depth video representation as
trajectory on Grassmann.
Geodesic distances between
successive subspaces of the
trajectory is computed.
Geometric Motion History (GMH)
gives the temporal evolution of the
depth video.
SO-SVM1 early event detection is
applied on the GMHs signals.
[1]. Hoai et de la Torre. IJCV ,2014
……....
25
26. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
Database:
The experiments are conducted on
Cam3D Kinect database1 which
contains depth videos for spontaneous
emotions
Protocol:
Two emotions will be detected
(Happiness vs. others and
Thinking/Unsure vs. others).
Targeted videos will be divided into two
halves, for training and testing.
Each emotion of interest will be
concatenated with two different others
randomly to have 100 samples for training
and testing.
[1]. Mahmoud et al., ACII,2011
26
27. 4D SPONTANEOUS EMOTION DETECTION
Evaluation Criteria
True Positive (TP) Rate: is the fraction of
time series that the detector fires during the
event of interest.
False Positive (FP) Rate: is the fraction
of time series that the detector fires before
the event of interest starts
I. ROC Curve: is the function of TPR
against FPR by varying the detection
threshold. Area Under ROC Curve
(AUC)
II. AMOC curve: is to evaluate the
timeliness of detection.
27
28. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
I. Grassmann vs. Stiefel manifold
Happiness detection
experiment
Thinking/unsure detection
experiment
28
29. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
II. Upper part of the body vs. face alone
29
30. 4D SPONTANEOUS EMOTION DETECTION
Physical pain detection from high resolution 4D-faces
Spontaneous pain detection out of facial expressions.
3D dynamic high resolution data is available.
Early detection of pain using SO-SVM framework
30
31. 4D SPONTANEOUS EMOTION DETECTION
Physical pain detection from high resolution 4D-faces
Depth-based Grassmann trajectories
Trajectory representation of the 3D video.
Velocity vectors computed between
successive subspaces.
Local Deformation Histogram (LDH) is
computed.
LDHs is concatenated.
The beginning and the end of the pain is
defined.
SO-SVM early detection.
1st Component of
the velocity
31
32. 4D SPONTANEOUS EMOTION DETECTION
Physical pain detection from high resolution 4D-faces
3D landmark-based Grassmann trajectories (Baseline)
3D physical pain video is divided into
subsequences.
Facial landmarks coordinates (x,y,z) is
used as facial descriptor (83)
Every subsequence is represented as
subspace.
Geodesic distance is quantified
between successive subspaces to build
the GMH.
The beginning and the end of the pain
is defined.
SO-SVM early detection.
32
33. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
Database:
BP4D-Spontaneous database1
41 subjects/ 8 Tasks
AUs annotation
Protocol:
28 physical pain videos are used.
14 for training and 14 for testing.
2-cross validation is applied.
The beginning and the end of pain is
determined according to action units
activation formula.
1,2
[1]. Zhang et al., IVCJ, 2014
[2] Prkachin et al., Pain, 2008.
AU4: Brow Lowering
AU6: Cheek raising
AU7: Tightening of eyelids
AU9: Wrinkling of nose
AU10: Raising of upper lip
33
34. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
I. Effect of the smoothing and pose normalization
AUC=0.75
AUC=0.70
AUC=0.63
AUC=0.78
AUC=0.74
AUC=0.70
Geodesicdistance/Normofthevelocity
- Increasing the derivation step
improves the results
- Normalizing the head pose improves
the results
34
35. 4D SPONTANEOUS EMOTION DETECTION
Experimental Analysis
II. Landmarks vs. Depth method
AUC=0.80
AUC=0.78
35
37. OUTLINE
4D Face Recognition
State of the art
4D face recognition framework
Experiments and results
4D Spontaneous Emotion Detection
State of the art
Trajectories on Grassmann manifold
Spontaneous emotional state detection from depth video
Spontaneous pain detection from 4D high resolution video
Conclusion and Future Work
37
38. CONCLUSION AND FUTURE WORK
Conclusions
Common geometric framework with two different
representations
4D Face Recognition
Efficient subspace representation for 4D data
Exploiting the shape and its dynamic improves the results
Enrich the dictionary improves the results
4D Emotion Detection
Modeling 3D video as time-parameterized curves (Trajectories) on
Grassmann manifold.
Upper part of the body outperforms the face alone.
Local approach (LDH) outperforms global approach (Distances)
Coupling geometric features (velocities) with advanced ML techniques
(Early-event detector)
38
39. CONCLUSION AND FUTURE WORK
Limitations
Lack of frame-to-frame vertex-level dense correspondence
Not considering the texture channel (available)
Limited number of subjects in the DB/lack of spontaneous DB
Perspectives and Future Work
Dense non-rigid registration/tracking
Investigating high order derivatives along the trajectories
39
40. PUBLICATION LIST
Submitted Journal
1. T. Alashkar, B. Ben Amor, M. Daoudi and S. Berretti , “Analyzing Trajectories on
Grassmann Manifolds for Spontaneous Emotion Detection ”, Submitted to IEEE
Transaction on Affective Computing, Sep-2015.
2. T. Alashkar, B. Ben Amor, M. Daoudi and S. Berretti “Modeling Shape Dynamics on
Grassmann Manifolds for 4D Face Recognition”, In preparation.
International Conferences and Workshops
1. T. Alashkar, B. Ben Amor, S. Berretti and M. Daoudi, “Analyzing Trajectories on
Grassmann Manifold for Early Emotion Detection from Depth Videos” in FG
2015.
2. T. Alashkar, B. Ben Amor, M. Daoudi and S. Berretti, “A 3D Dynamic Database for
Unconstrained Face Recognition ” in 3D Body Scanning Technology International
Conference 2014.
3. T. Alashkar, B. Ben Amor, M. Daoudi and S. Berretti, “A Grassmannian Framework
for Face Recognition of 3D Dynamic Sequences with Challenging Conditions ”
in Springer NORDIA Workshop in ECCV 2014.
National Conference
1. T. Alashkar, B. Ben Amor, S. Berretti and M. Daoudi, “Analyse des trajectoires sur
une Grassmannienne pour la détection d’émotions dans des vidéos de
profondeur” in ORASIS 2015.
40
41. Thank You and You are Welcome
41
Shawrma Taboula Kebba Yabrak
Syrian SweetsNamouraMtabalHommos
Salle: F024