This document summarizes a capstone report that describes the creation of a facial recognition system using a convolutional neural network to identify emotions and gender from images of faces. The system achieves 93% accuracy for emotion recognition and nearly 90% accuracy for gender detection. It analyzes grayscale images of faces that are 48 by 48 pixels in size to identify these attributes. The system is deployed through a web application built with Flask to allow real-time analysis and interpretation of video feeds or images. The goal is to demonstrate how this technology can be applied in business contexts like customer relationship management systems to enhance customer experiences.
https://ijitce.com/index.php
Our journal maintains rigorous peer review standards. Each submitted article undergoes a thorough evaluation by experts in the respective field. This stringent review process helps ensure that only high-quality and scientifically sound research is accepted for publication. Researchers can trust that the articles they find in IJITCE have been critically assessed for validity, significance, and originality.
Facial emotion recognition using enhanced multi-verse optimizer methodIJECEIAES
In recent years, facial emotion recognition has gained significant improvement and attention. This technology utilizes advanced algorithms to analyze facial expressions, enabling computers to detect and interpret human emotions accurately. Its applications span over a wide range of fields, from improving customer service through sentiment analysis, to enhancing mental health support by monitoring emotional states. However, there are several challenges in facial emotion recognition, including variability in individual expressions, cultural differences in emotion display, and privacy concerns related to data collection and usage. Lighting conditions, occlusions, and the need for diverse datasets also impacts accuracy. To solve these issues, an enhanced multi-verse optimizer (EMVO) technique is proposed to improve the efficiency of recognizing emotions. Moreover, EMVO is used to improve the convergence speed, exploration-exploitation balance, solution quality, and the applicability in different types of optimization problems. Two datasets were used to collect the data, namely YouTube and surrey audio-visual expressed emotion (SAVEE) datasets. Then, the classification is done using the convolutional neural networks (CNN) to improve the performance of emotion recognition. When compared to the existing methods shuffled frog leaping algorithm-incremental wrapper-based subset selection (SFLA-IWSS), hierarchical deep neural network (H-DNN) and unique preference learning (UPL), the proposed method achieved better accuracies, measured at 98.65% and 98.76% on the YouTube and SAVEE datasets, respectively.
IRJET- A Survey on Image Retrieval using Machine LearningIRJET Journal
This document summarizes a research paper on using machine learning for image retrieval. It discusses using optical character recognition (OCR) to extract text from images and then translate that text into machine-readable formats. The document reviews related work on combining visual and text features for image retrieval. It also provides an overview of the proposed system architecture, which combines color, texture, and edge features for image retrieval and evaluates performance using metrics like sensitivity, specificity, retrieval score, and accuracy.
The objective of emotion recognition is
identifying emotions of a human. The emotion can be
captured either from face or from verbal communication. In
this work we focus on identifying human emotion from
facial expressions. Facial emotion recognition is one of the
useful task and can be used as a base for many real-time
applications.
This document presents a facial expression recognition system that identifies and classifies seven basic expressions: happy, surprise, fear, disgust, sad, anger, and a neutral state. The system consists of four main parts: image acquisition, pre-processing, feature extraction, and classification. It was developed using both OpenCV and a web-based JavaScript approach. The system was tested on both real-time and pre-recorded video streams and can identify emotions in images and video input from a webcam in real-time. Evaluation showed the JavaScript implementation using a generalized dataset provided more accurate real-time predictions compared to the OpenCV approach.
A Study on Face Expression Observation Systemsijtsrd
Human expressions can convey a great deal of information. We cannot learn every language in the world, but we can decipher the majority of human expressions. The state of a users conduct in various settings and scenarios can be inferred from their facial expressions. Through various human computer interface and programming approaches, facial expression can be digitized. Face detection, feature extraction, and kind of expression determination are all parts of the facial expression perception process. Verbal and non verbal forms of communication are both possible. Through their emotions, people can communicate nonverbally Weve given a broad summary of the various facial expression perception processes in the literature in this article. Jyoti | Neeraj Chawaria | Ekta "A Study on Face Expression Observation Systems" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-5 , August 2022, URL: https://www.ijtsrd.com/papers/ijtsrd50450.pdf Paper URL: https://www.ijtsrd.com/computer-science/computer-graphics/50450/a-study-on-face-expression-observation-systems/jyoti
This document provides a summary of a minor project report on image recognition submitted in partial fulfillment of the requirements for a Bachelor of Technology degree in Computer Science and Engineering. The report was submitted by Bhaskar Tripathi and Joel Jose in October 2018 under the supervision of Dr. P. Mohamed Fathimal, Assistant Professor in the Department of Computer Science and Engineering at SRM Institute of Science and Technology. The report includes acknowledgements, a table of contents, and chapters on the introduction, project details, tools and technologies used, proposed system architecture, modules and functionality.
This document discusses pattern recognition and its techniques and use cases. It defines pattern recognition as the process of distinguishing and segmenting data according to set criteria or common elements. The main techniques discussed are statistical, syntactic, and template matching models. Key uses of pattern recognition include data analytics, natural language processing, optical character recognition, image recognition, voice recognition, and sentiment analysis. The document concludes that pattern recognition is key to further advances in computational technology and big data analytics.
https://ijitce.com/index.php
Our journal maintains rigorous peer review standards. Each submitted article undergoes a thorough evaluation by experts in the respective field. This stringent review process helps ensure that only high-quality and scientifically sound research is accepted for publication. Researchers can trust that the articles they find in IJITCE have been critically assessed for validity, significance, and originality.
Facial emotion recognition using enhanced multi-verse optimizer methodIJECEIAES
In recent years, facial emotion recognition has gained significant improvement and attention. This technology utilizes advanced algorithms to analyze facial expressions, enabling computers to detect and interpret human emotions accurately. Its applications span over a wide range of fields, from improving customer service through sentiment analysis, to enhancing mental health support by monitoring emotional states. However, there are several challenges in facial emotion recognition, including variability in individual expressions, cultural differences in emotion display, and privacy concerns related to data collection and usage. Lighting conditions, occlusions, and the need for diverse datasets also impacts accuracy. To solve these issues, an enhanced multi-verse optimizer (EMVO) technique is proposed to improve the efficiency of recognizing emotions. Moreover, EMVO is used to improve the convergence speed, exploration-exploitation balance, solution quality, and the applicability in different types of optimization problems. Two datasets were used to collect the data, namely YouTube and surrey audio-visual expressed emotion (SAVEE) datasets. Then, the classification is done using the convolutional neural networks (CNN) to improve the performance of emotion recognition. When compared to the existing methods shuffled frog leaping algorithm-incremental wrapper-based subset selection (SFLA-IWSS), hierarchical deep neural network (H-DNN) and unique preference learning (UPL), the proposed method achieved better accuracies, measured at 98.65% and 98.76% on the YouTube and SAVEE datasets, respectively.
IRJET- A Survey on Image Retrieval using Machine LearningIRJET Journal
This document summarizes a research paper on using machine learning for image retrieval. It discusses using optical character recognition (OCR) to extract text from images and then translate that text into machine-readable formats. The document reviews related work on combining visual and text features for image retrieval. It also provides an overview of the proposed system architecture, which combines color, texture, and edge features for image retrieval and evaluates performance using metrics like sensitivity, specificity, retrieval score, and accuracy.
The objective of emotion recognition is
identifying emotions of a human. The emotion can be
captured either from face or from verbal communication. In
this work we focus on identifying human emotion from
facial expressions. Facial emotion recognition is one of the
useful task and can be used as a base for many real-time
applications.
This document presents a facial expression recognition system that identifies and classifies seven basic expressions: happy, surprise, fear, disgust, sad, anger, and a neutral state. The system consists of four main parts: image acquisition, pre-processing, feature extraction, and classification. It was developed using both OpenCV and a web-based JavaScript approach. The system was tested on both real-time and pre-recorded video streams and can identify emotions in images and video input from a webcam in real-time. Evaluation showed the JavaScript implementation using a generalized dataset provided more accurate real-time predictions compared to the OpenCV approach.
A Study on Face Expression Observation Systemsijtsrd
Human expressions can convey a great deal of information. We cannot learn every language in the world, but we can decipher the majority of human expressions. The state of a users conduct in various settings and scenarios can be inferred from their facial expressions. Through various human computer interface and programming approaches, facial expression can be digitized. Face detection, feature extraction, and kind of expression determination are all parts of the facial expression perception process. Verbal and non verbal forms of communication are both possible. Through their emotions, people can communicate nonverbally Weve given a broad summary of the various facial expression perception processes in the literature in this article. Jyoti | Neeraj Chawaria | Ekta "A Study on Face Expression Observation Systems" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-6 | Issue-5 , August 2022, URL: https://www.ijtsrd.com/papers/ijtsrd50450.pdf Paper URL: https://www.ijtsrd.com/computer-science/computer-graphics/50450/a-study-on-face-expression-observation-systems/jyoti
This document provides a summary of a minor project report on image recognition submitted in partial fulfillment of the requirements for a Bachelor of Technology degree in Computer Science and Engineering. The report was submitted by Bhaskar Tripathi and Joel Jose in October 2018 under the supervision of Dr. P. Mohamed Fathimal, Assistant Professor in the Department of Computer Science and Engineering at SRM Institute of Science and Technology. The report includes acknowledgements, a table of contents, and chapters on the introduction, project details, tools and technologies used, proposed system architecture, modules and functionality.
This document discusses pattern recognition and its techniques and use cases. It defines pattern recognition as the process of distinguishing and segmenting data according to set criteria or common elements. The main techniques discussed are statistical, syntactic, and template matching models. Key uses of pattern recognition include data analytics, natural language processing, optical character recognition, image recognition, voice recognition, and sentiment analysis. The document concludes that pattern recognition is key to further advances in computational technology and big data analytics.
What are Cognitive Applications? What is exciting about them? They represent a whole new way of human computer interaction and acting on data insights. Introducing IBM Watson and how to develop Cognitive applications. AI, Machine Learning compared and contrasted.
MTCNN BASED AUTOMATIC ATTENDANCE SYSTEM USING FACE RECOGNITIONIRJET Journal
This document summarizes a research paper that proposes an automated attendance system using facial recognition technology. It begins by outlining the limitations of current manual and RFID card-based attendance systems. It then describes a new system that uses MTCNN for face detection and CNN for facial recognition. The system captures images and identifies recognized students as present by matching faces to a database of stored images. The document provides details on the various stages of the proposed method, including face detection using MTCNN, face alignment, feature extraction with FaceNet, and classification with SVM. It presents the overall algorithm and concludes by discussing modelling and analysis.
Age and Gender Prediction and Human countIRJET Journal
This document presents a system for age and gender prediction and human counting using deep convolutional neural networks. The system is trained on the Adience benchmark dataset containing over 26,000 images labeled with age and gender. A caffe model is used to perform the predictions, which are then deployed on a website. The model achieves 88% accuracy for age and gender prediction. The system also counts the number of faces in an image using a simple CNN architecture. The purpose is to demonstrate how machine learning can be used to detect age, gender, and count humans in images with reduced manual effort.
This document proposes integrating human fingerprint expert knowledge with machine learning algorithms to improve quantitative fingerprint analysis. The researchers will design experiments to extract relevant information from experts and use it to enhance statistical fingerprint matching models. Their goal is to demonstrate that expert knowledge can provide important context that current algorithms lack, such as knowledge of important fingerprint regions and the ability to focus on specific features. They will apply expert-guided data to various statistical analyses and techniques adapted from natural language processing to integrate expert-derived temporal sequences into the matching process.
The document describes a Driverless ML API that was created to automate machine learning workflows including feature engineering, model validation, tuning, selection, and deployment. The API uses machine learning interpretability techniques to provide visualizations and explanations of models. It aims to help scale data science efforts and enable both expert and junior data scientists to more quickly develop accurate, production-ready models. Key capabilities of the API include automated exploratory data analysis, feature selection and engineering, model selection and hyperparameter tuning using GPUs for faster training, and model interpretability visualizations.
Public Figure Recognition Using SVM and Computer Vision TechniquesIRJET Journal
This document describes a study that developed a public figure recognition system using support vector machines (SVM) and computer vision techniques. The system achieved 90.24% accuracy on test data. It used Haar cascade classifiers to extract facial features from images and wavelet transformations to isolate key features. Unlike systems using convolutional neural networks, the SVM model requires less training data while still providing accurate classifications. The study suggests this system could also be applied to student identification in academic settings by verifying student enrollment. It addresses limitations of other systems, such as data and computational requirements, while demonstrating versatility across applications.
Text pre-processing of multilingual for sentiment analysis based on social ne...IJECEIAES
Sentiment analysis (SA) is an enduring area for research especially in the field of text analysis. Text pre-processing is an important aspect to perform SA accurately. This paper presents a text processing model for SA, using natural language processing techniques for twitter data. The basic phases for machine learning are text collection, text cleaning, pre-processing, feature extractions in a text and then categorize the data according to the SA techniques. Keeping the focus on twitter data, the data is extracted in domain specific manner. In data cleaning phase, noisy data, missing data, punctuation, tags and emoticons have been considered. For pre-processing, tokenization is performed which is followed by stop word removal (SWR). The proposed article provides an insight of the techniques, that are used for text pre-processing, the impact of their presence on the dataset. The accuracy of classification techniques has been improved after applying text preprocessing and dimensionality has been reduced. The proposed corpus can be utilized in the area of market analysis, customer behaviour, polling analysis, and brand monitoring. The text pre-processing process can serve as the baseline to apply predictive analysis, machine learning and deep learning algorithms which can be extended according to problem definition.
Quantifying human experience for increased intelligence within work teams an...Katri Saarikivi
We are looking for industry partners for a research project exploring the possibilities of the sciences and technologies of emotions and interaction in human-centered business.
Did you know that a recent study by McKinsey & Company highlighted that 84% of organizations are concerned about bias in their AI algorithms? However, there's a solution to this problem. Upholding best practices can significantly mitigate biases in AI for enterprises, particularly given the challenges posed by compliance and the rapid dissemination of information through digital media.
In this E42 Blog post, we delve into an array of best practices to mitigate bias and hallucinations in AI models. A few of these best practices include:
Model optimization: This practice focuses on enhancing model performance and reducing bias through various optimization techniques
Understanding model architecture: This involves a deep dive into the structure of AI models to identify and rectify biases
Human interactions: This emphasizes on the critical role of human feedback in the training loop in ensuring unbiased AI outcomes
On-premises large language models: This practice involves utilizing on-premises LLMs to maintain control over data and model training
A Smart Receptionist Implementing Facial Recognition and Voice InteractionCSCJournals
The purpose of this research is to implement a smart receptionist system with facial recognition and voice interaction using deep learning. The facial recognition component is implemented using real time image processing techniques, and it can be used to learn new faces as well as detect and recognize existing faces. The first time a customer uses this system, it will take the person’s facial data to create a unique user facial model, and this model will be triggered if the person comes the second time. The recognition is done in real time and after which voice interaction will be applied. Voice interaction is used to provide a life-like human communication and improve user experience. Our proposed smart receptionist system could be integrated into the self check-in kiosks deployed in hospitals or smart buildings to streamline the user recognition process and provide customized user interactions. This system could also be used in smart home environment where smart cameras have been deployed and voice assistants are in place.
This document describes a proposed Web-Based Applicants' Matching System (WBAMS) that would match suitable job applicants to appropriate jobs. Key points:
- The current manual recruitment process at the Federal College of Chemical and Leather Technology has problems like high workload, lack of information.
- The proposed WBAMS would allow applicants to apply online and store applicant and job data in a central database. It would use criteria to match applicants to jobs and shortlist the most suitable candidates.
- The system would be designed using technologies like MySQL database, Apache web server. A three-tier architecture model is proposed for communication between the database, application logic and user interface layers.
Age and Gender Classification using Convolutional Neural NetworkIRJET Journal
This document describes a study that aims to accurately identify the gender and age range of facial images using convolutional neural networks. It begins with an introduction to age and gender classification and some of the challenges. It then discusses the system analysis and design, including the use of convolutional neural networks and machine learning techniques. The implementation section notes that Python, TensorFlow, Keras and OpenCV will be used to build a convolutional neural network model to detect faces in images and predict age and gender through training on available datasets. The overall goal is to develop an accurate system for age and gender detection from facial images.
IRJET- Social Network Mental Disorders Detection Via Online Social Media MiningIRJET Journal
This document summarizes a research paper about detecting social network mental disorders via online social media mining. The paper proposes a system to analyze user posts and interactions on social media like Facebook to detect signs of mental disorders like information overload or cyber-relationship addiction. The system uses techniques like sentiment analysis of posts using CNNs and classification of user stress levels using TSVM. If a user is detected as having a mental disorder, the system can recommend nearby hospitals on a map and send the user precautionary information to avoid disorders and promote well-being. The goal is to help detect mental disorders earlier through social media analysis to enable timely clinical intervention.
Facial Recognition based Attendance System: A SurveyIRJET Journal
The document describes a facial recognition-based attendance system. It discusses how traditional attendance methods are inefficient and error-prone. The proposed system utilizes facial recognition technology, machine learning algorithms, and a database to accurately identify and track attendance in real-time. It analyzes different facial recognition and machine learning methods like OpenCV, cosine similarity, and distance metrics to efficiently and securely match facial templates. The system aims to provide a customized, user-friendly and robust alternative to traditional attendance tracking methods in educational institutions and organizations.
This document proposes integrating human fingerprint expert knowledge with machine learning algorithms to improve quantitative fingerprint analysis. The researchers will design experiments to extract relevant information from experts and use it to guide statistical models. Their goal is to demonstrate how expert knowledge, represented properly, can enhance fingerprint matching performance compared to current automated approaches alone. In particular, they will apply experts' understanding of important fingerprint regions and ability to focus on specific features to statistical modeling.
IRJET- Deep Learning Based Card-Less Atm Using Fingerprint And Face Recogniti...IRJET Journal
This document summarizes a proposed system for a cardless ATM that uses fingerprint and face recognition for authentication instead of an ATM card. The system would use CNN models for face recognition and minutiae feature extraction for fingerprint recognition. It provides background on deep learning techniques and discusses related work applying deep learning for tasks like face recognition, fingerprint spoofing detection, and improving machine learning algorithms. The proposed cardless ATM system aims to address security issues with the current card-based system and allow multiple family members to access the same account from different locations using their biometrics instead of an ATM card.
This document describes a proposed next-generation attendance management system using face recognition technology. The system aims to automate the attendance tracking process for students in an educational institution. It works by enrolling students into the system by capturing their facial images and extracting facial features to create unique encodings for each student. When taking attendance, the system uses these encodings to identify students' faces in captured images or video and marks their attendance automatically. This reduces the burden on teachers of manually taking and entering attendance. The system also calculates and stores attendance records in a database for viewing and analysis. The proposed system aims to address limitations of traditional attendance tracking methods like errors, time consumption and lack of data analysis capabilities.
This document proposes and evaluates methods for fusing 3D ear and face biometrics at the score level and feature level for personal authentication. Local 3D features are extracted from ear and face data and fused using root mean square distance matching at the feature level. At the score level, matching scores from ear and face modalities are fused using weighted sum rule techniques. Experiments on a database of 990 ear and face images from 60 individuals show that the multimodal biometrics systems using feature level or score level fusion techniques have lower equal error rates compared to unimodal ear or face systems alone, demonstrating improved performance from fusing the biometric modalities.
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...ijtsrd
This paper presents an integrated system for emotion detection using facial detection and recognition. we have taken into account the fact that emotions are most widely represented with eyes and mouth expressions. In this research effort, we implement a general convolutional neural network CNN building framework for designing real time CNNs. We validate our models by creating a real time vision system that accomplishes the tasks of face detection, emotion classification, and generating the content according to the emotion or mood of the person simultaneously in one blended step using our proposed CNN architecture. Our proposed model consisted of modules such as image processing, Feature extraction, feature classification, and recommendation process. The images used in the experiment are pre processed with various image processing methods like canny edge detection, histogram equalization, fit ellipse, and FER dataset is mediated for conducting the experiments. With a trained profile that can be updated flexibly, a user can detect his her behavior on a real time basis. It utilizes the state of the art of face detection and recognition algorithms. Sanket Godbole | Jaivardhan Shelke "AI Therapist – Emotion Detection using Facial Detection and Recognition & Showing Content According to Emotions" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-6 , October 2020, URL: https://www.ijtsrd.com/papers/ijtsrd33267.pdf Paper Url: https://www.ijtsrd.com/computer-science/artificial-intelligence/33267/ai-therapist-–-emotion-detection-using-facial-detection-and-recognition-and-showing-content-according-to-emotions/sanket-godbole
This document discusses a product analyst advisor software that uses natural language processing techniques like sentiment analysis to analyze customer reviews and sentiments about products. It extracts reviews from various websites about a product being researched and processes the data to provide useful insights. The insights help users easily select the best available option. The system architecture involves scraping live data from websites, using deep learning algorithms to analyze reviews for sentiments, and displaying product insights. It uses BERT for sentiment analysis and frameworks like Django and ReactJS. Web scraping is used to extract review data for analysis and providing recommendations to users.
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
More Related Content
Similar to Deep_Learning_Innovations_In_Facial_Analysis
What are Cognitive Applications? What is exciting about them? They represent a whole new way of human computer interaction and acting on data insights. Introducing IBM Watson and how to develop Cognitive applications. AI, Machine Learning compared and contrasted.
MTCNN BASED AUTOMATIC ATTENDANCE SYSTEM USING FACE RECOGNITIONIRJET Journal
This document summarizes a research paper that proposes an automated attendance system using facial recognition technology. It begins by outlining the limitations of current manual and RFID card-based attendance systems. It then describes a new system that uses MTCNN for face detection and CNN for facial recognition. The system captures images and identifies recognized students as present by matching faces to a database of stored images. The document provides details on the various stages of the proposed method, including face detection using MTCNN, face alignment, feature extraction with FaceNet, and classification with SVM. It presents the overall algorithm and concludes by discussing modelling and analysis.
Age and Gender Prediction and Human countIRJET Journal
This document presents a system for age and gender prediction and human counting using deep convolutional neural networks. The system is trained on the Adience benchmark dataset containing over 26,000 images labeled with age and gender. A caffe model is used to perform the predictions, which are then deployed on a website. The model achieves 88% accuracy for age and gender prediction. The system also counts the number of faces in an image using a simple CNN architecture. The purpose is to demonstrate how machine learning can be used to detect age, gender, and count humans in images with reduced manual effort.
This document proposes integrating human fingerprint expert knowledge with machine learning algorithms to improve quantitative fingerprint analysis. The researchers will design experiments to extract relevant information from experts and use it to enhance statistical fingerprint matching models. Their goal is to demonstrate that expert knowledge can provide important context that current algorithms lack, such as knowledge of important fingerprint regions and the ability to focus on specific features. They will apply expert-guided data to various statistical analyses and techniques adapted from natural language processing to integrate expert-derived temporal sequences into the matching process.
The document describes a Driverless ML API that was created to automate machine learning workflows including feature engineering, model validation, tuning, selection, and deployment. The API uses machine learning interpretability techniques to provide visualizations and explanations of models. It aims to help scale data science efforts and enable both expert and junior data scientists to more quickly develop accurate, production-ready models. Key capabilities of the API include automated exploratory data analysis, feature selection and engineering, model selection and hyperparameter tuning using GPUs for faster training, and model interpretability visualizations.
Public Figure Recognition Using SVM and Computer Vision TechniquesIRJET Journal
This document describes a study that developed a public figure recognition system using support vector machines (SVM) and computer vision techniques. The system achieved 90.24% accuracy on test data. It used Haar cascade classifiers to extract facial features from images and wavelet transformations to isolate key features. Unlike systems using convolutional neural networks, the SVM model requires less training data while still providing accurate classifications. The study suggests this system could also be applied to student identification in academic settings by verifying student enrollment. It addresses limitations of other systems, such as data and computational requirements, while demonstrating versatility across applications.
Text pre-processing of multilingual for sentiment analysis based on social ne...IJECEIAES
Sentiment analysis (SA) is an enduring area for research especially in the field of text analysis. Text pre-processing is an important aspect to perform SA accurately. This paper presents a text processing model for SA, using natural language processing techniques for twitter data. The basic phases for machine learning are text collection, text cleaning, pre-processing, feature extractions in a text and then categorize the data according to the SA techniques. Keeping the focus on twitter data, the data is extracted in domain specific manner. In data cleaning phase, noisy data, missing data, punctuation, tags and emoticons have been considered. For pre-processing, tokenization is performed which is followed by stop word removal (SWR). The proposed article provides an insight of the techniques, that are used for text pre-processing, the impact of their presence on the dataset. The accuracy of classification techniques has been improved after applying text preprocessing and dimensionality has been reduced. The proposed corpus can be utilized in the area of market analysis, customer behaviour, polling analysis, and brand monitoring. The text pre-processing process can serve as the baseline to apply predictive analysis, machine learning and deep learning algorithms which can be extended according to problem definition.
Quantifying human experience for increased intelligence within work teams an...Katri Saarikivi
We are looking for industry partners for a research project exploring the possibilities of the sciences and technologies of emotions and interaction in human-centered business.
Did you know that a recent study by McKinsey & Company highlighted that 84% of organizations are concerned about bias in their AI algorithms? However, there's a solution to this problem. Upholding best practices can significantly mitigate biases in AI for enterprises, particularly given the challenges posed by compliance and the rapid dissemination of information through digital media.
In this E42 Blog post, we delve into an array of best practices to mitigate bias and hallucinations in AI models. A few of these best practices include:
Model optimization: This practice focuses on enhancing model performance and reducing bias through various optimization techniques
Understanding model architecture: This involves a deep dive into the structure of AI models to identify and rectify biases
Human interactions: This emphasizes on the critical role of human feedback in the training loop in ensuring unbiased AI outcomes
On-premises large language models: This practice involves utilizing on-premises LLMs to maintain control over data and model training
A Smart Receptionist Implementing Facial Recognition and Voice InteractionCSCJournals
The purpose of this research is to implement a smart receptionist system with facial recognition and voice interaction using deep learning. The facial recognition component is implemented using real time image processing techniques, and it can be used to learn new faces as well as detect and recognize existing faces. The first time a customer uses this system, it will take the person’s facial data to create a unique user facial model, and this model will be triggered if the person comes the second time. The recognition is done in real time and after which voice interaction will be applied. Voice interaction is used to provide a life-like human communication and improve user experience. Our proposed smart receptionist system could be integrated into the self check-in kiosks deployed in hospitals or smart buildings to streamline the user recognition process and provide customized user interactions. This system could also be used in smart home environment where smart cameras have been deployed and voice assistants are in place.
This document describes a proposed Web-Based Applicants' Matching System (WBAMS) that would match suitable job applicants to appropriate jobs. Key points:
- The current manual recruitment process at the Federal College of Chemical and Leather Technology has problems like high workload, lack of information.
- The proposed WBAMS would allow applicants to apply online and store applicant and job data in a central database. It would use criteria to match applicants to jobs and shortlist the most suitable candidates.
- The system would be designed using technologies like MySQL database, Apache web server. A three-tier architecture model is proposed for communication between the database, application logic and user interface layers.
Age and Gender Classification using Convolutional Neural NetworkIRJET Journal
This document describes a study that aims to accurately identify the gender and age range of facial images using convolutional neural networks. It begins with an introduction to age and gender classification and some of the challenges. It then discusses the system analysis and design, including the use of convolutional neural networks and machine learning techniques. The implementation section notes that Python, TensorFlow, Keras and OpenCV will be used to build a convolutional neural network model to detect faces in images and predict age and gender through training on available datasets. The overall goal is to develop an accurate system for age and gender detection from facial images.
IRJET- Social Network Mental Disorders Detection Via Online Social Media MiningIRJET Journal
This document summarizes a research paper about detecting social network mental disorders via online social media mining. The paper proposes a system to analyze user posts and interactions on social media like Facebook to detect signs of mental disorders like information overload or cyber-relationship addiction. The system uses techniques like sentiment analysis of posts using CNNs and classification of user stress levels using TSVM. If a user is detected as having a mental disorder, the system can recommend nearby hospitals on a map and send the user precautionary information to avoid disorders and promote well-being. The goal is to help detect mental disorders earlier through social media analysis to enable timely clinical intervention.
Facial Recognition based Attendance System: A SurveyIRJET Journal
The document describes a facial recognition-based attendance system. It discusses how traditional attendance methods are inefficient and error-prone. The proposed system utilizes facial recognition technology, machine learning algorithms, and a database to accurately identify and track attendance in real-time. It analyzes different facial recognition and machine learning methods like OpenCV, cosine similarity, and distance metrics to efficiently and securely match facial templates. The system aims to provide a customized, user-friendly and robust alternative to traditional attendance tracking methods in educational institutions and organizations.
This document proposes integrating human fingerprint expert knowledge with machine learning algorithms to improve quantitative fingerprint analysis. The researchers will design experiments to extract relevant information from experts and use it to guide statistical models. Their goal is to demonstrate how expert knowledge, represented properly, can enhance fingerprint matching performance compared to current automated approaches alone. In particular, they will apply experts' understanding of important fingerprint regions and ability to focus on specific features to statistical modeling.
IRJET- Deep Learning Based Card-Less Atm Using Fingerprint And Face Recogniti...IRJET Journal
This document summarizes a proposed system for a cardless ATM that uses fingerprint and face recognition for authentication instead of an ATM card. The system would use CNN models for face recognition and minutiae feature extraction for fingerprint recognition. It provides background on deep learning techniques and discusses related work applying deep learning for tasks like face recognition, fingerprint spoofing detection, and improving machine learning algorithms. The proposed cardless ATM system aims to address security issues with the current card-based system and allow multiple family members to access the same account from different locations using their biometrics instead of an ATM card.
This document describes a proposed next-generation attendance management system using face recognition technology. The system aims to automate the attendance tracking process for students in an educational institution. It works by enrolling students into the system by capturing their facial images and extracting facial features to create unique encodings for each student. When taking attendance, the system uses these encodings to identify students' faces in captured images or video and marks their attendance automatically. This reduces the burden on teachers of manually taking and entering attendance. The system also calculates and stores attendance records in a database for viewing and analysis. The proposed system aims to address limitations of traditional attendance tracking methods like errors, time consumption and lack of data analysis capabilities.
This document proposes and evaluates methods for fusing 3D ear and face biometrics at the score level and feature level for personal authentication. Local 3D features are extracted from ear and face data and fused using root mean square distance matching at the feature level. At the score level, matching scores from ear and face modalities are fused using weighted sum rule techniques. Experiments on a database of 990 ear and face images from 60 individuals show that the multimodal biometrics systems using feature level or score level fusion techniques have lower equal error rates compared to unimodal ear or face systems alone, demonstrating improved performance from fusing the biometric modalities.
AI Therapist – Emotion Detection using Facial Detection and Recognition and S...ijtsrd
This paper presents an integrated system for emotion detection using facial detection and recognition. we have taken into account the fact that emotions are most widely represented with eyes and mouth expressions. In this research effort, we implement a general convolutional neural network CNN building framework for designing real time CNNs. We validate our models by creating a real time vision system that accomplishes the tasks of face detection, emotion classification, and generating the content according to the emotion or mood of the person simultaneously in one blended step using our proposed CNN architecture. Our proposed model consisted of modules such as image processing, Feature extraction, feature classification, and recommendation process. The images used in the experiment are pre processed with various image processing methods like canny edge detection, histogram equalization, fit ellipse, and FER dataset is mediated for conducting the experiments. With a trained profile that can be updated flexibly, a user can detect his her behavior on a real time basis. It utilizes the state of the art of face detection and recognition algorithms. Sanket Godbole | Jaivardhan Shelke "AI Therapist – Emotion Detection using Facial Detection and Recognition & Showing Content According to Emotions" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-4 | Issue-6 , October 2020, URL: https://www.ijtsrd.com/papers/ijtsrd33267.pdf Paper Url: https://www.ijtsrd.com/computer-science/artificial-intelligence/33267/ai-therapist-–-emotion-detection-using-facial-detection-and-recognition-and-showing-content-according-to-emotions/sanket-godbole
This document discusses a product analyst advisor software that uses natural language processing techniques like sentiment analysis to analyze customer reviews and sentiments about products. It extracts reviews from various websites about a product being researched and processes the data to provide useful insights. The insights help users easily select the best available option. The system architecture involves scraping live data from websites, using deep learning algorithms to analyze reviews for sentiments, and displaying product insights. It uses BERT for sentiment analysis and frameworks like Django and ReactJS. Web scraping is used to extract review data for analysis and providing recommendations to users.
Similar to Deep_Learning_Innovations_In_Facial_Analysis (20)
Infrastructure Challenges in Scaling RAG with Custom AI modelsZilliz
Building Retrieval-Augmented Generation (RAG) systems with open-source and custom AI models is a complex task. This talk explores the challenges in productionizing RAG systems, including retrieval performance, response synthesis, and evaluation. We’ll discuss how to leverage open-source models like text embeddings, language models, and custom fine-tuned models to enhance RAG performance. Additionally, we’ll cover how BentoML can help orchestrate and scale these AI components efficiently, ensuring seamless deployment and management of RAG systems in the cloud.
How to Get CNIC Information System with Paksim Ga.pptxdanishmna97
Pakdata Cf is a groundbreaking system designed to streamline and facilitate access to CNIC information. This innovative platform leverages advanced technology to provide users with efficient and secure access to their CNIC details.
Your One-Stop Shop for Python Success: Top 10 US Python Development Providersakankshawande
Simplify your search for a reliable Python development partner! This list presents the top 10 trusted US providers offering comprehensive Python development services, ensuring your project's success from conception to completion.
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Unlocking Productivity: Leveraging the Potential of Copilot in Microsoft 365, a presentation by Christoforos Vlachos, Senior Solutions Manager – Modern Workplace, Uni Systems
AI-Powered Food Delivery Transforming App Development in Saudi Arabia.pdfTechgropse Pvt.Ltd.
In this blog post, we'll delve into the intersection of AI and app development in Saudi Arabia, focusing on the food delivery sector. We'll explore how AI is revolutionizing the way Saudi consumers order food, how restaurants manage their operations, and how delivery partners navigate the bustling streets of cities like Riyadh, Jeddah, and Dammam. Through real-world case studies, we'll showcase how leading Saudi food delivery apps are leveraging AI to redefine convenience, personalization, and efficiency.
In the rapidly evolving landscape of technologies, XML continues to play a vital role in structuring, storing, and transporting data across diverse systems. The recent advancements in artificial intelligence (AI) present new methodologies for enhancing XML development workflows, introducing efficiency, automation, and intelligent capabilities. This presentation will outline the scope and perspective of utilizing AI in XML development. The potential benefits and the possible pitfalls will be highlighted, providing a balanced view of the subject.
We will explore the capabilities of AI in understanding XML markup languages and autonomously creating structured XML content. Additionally, we will examine the capacity of AI to enrich plain text with appropriate XML markup. Practical examples and methodological guidelines will be provided to elucidate how AI can be effectively prompted to interpret and generate accurate XML markup.
Further emphasis will be placed on the role of AI in developing XSLT, or schemas such as XSD and Schematron. We will address the techniques and strategies adopted to create prompts for generating code, explaining code, or refactoring the code, and the results achieved.
The discussion will extend to how AI can be used to transform XML content. In particular, the focus will be on the use of AI XPath extension functions in XSLT, Schematron, Schematron Quick Fixes, or for XML content refactoring.
The presentation aims to deliver a comprehensive overview of AI usage in XML development, providing attendees with the necessary knowledge to make informed decisions. Whether you’re at the early stages of adopting AI or considering integrating it in advanced XML development, this presentation will cover all levels of expertise.
By highlighting the potential advantages and challenges of integrating AI with XML development tools and languages, the presentation seeks to inspire thoughtful conversation around the future of XML development. We’ll not only delve into the technical aspects of AI-powered XML development but also discuss practical implications and possible future directions.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
HCL Notes and Domino License Cost Reduction in the World of DLAUpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-notes-and-domino-license-cost-reduction-in-the-world-of-dlau/
The introduction of DLAU and the CCB & CCX licensing model caused quite a stir in the HCL community. As a Notes and Domino customer, you may have faced challenges with unexpected user counts and license costs. You probably have questions on how this new licensing approach works and how to benefit from it. Most importantly, you likely have budget constraints and want to save money where possible. Don’t worry, we can help with all of this!
We’ll show you how to fix common misconfigurations that cause higher-than-expected user counts, and how to identify accounts which you can deactivate to save money. There are also frequent patterns that can cause unnecessary cost, like using a person document instead of a mail-in for shared mailboxes. We’ll provide examples and solutions for those as well. And naturally we’ll explain the new licensing model.
Join HCL Ambassador Marc Thomas in this webinar with a special guest appearance from Franz Walder. It will give you the tools and know-how to stay on top of what is going on with Domino licensing. You will be able lower your cost through an optimized configuration and keep it low going forward.
These topics will be covered
- Reducing license cost by finding and fixing misconfigurations and superfluous accounts
- How do CCB and CCX licenses really work?
- Understanding the DLAU tool and how to best utilize it
- Tips for common problem areas, like team mailboxes, functional/test users, etc
- Practical examples and best practices to implement right away
Generating privacy-protected synthetic data using Secludy and MilvusZilliz
During this demo, the founders of Secludy will demonstrate how their system utilizes Milvus to store and manipulate embeddings for generating privacy-protected synthetic data. Their approach not only maintains the confidentiality of the original data but also enhances the utility and scalability of LLMs under privacy constraints. Attendees, including machine learning engineers, data scientists, and data managers, will witness first-hand how Secludy's integration with Milvus empowers organizations to harness the power of LLMs securely and efficiently.
TrustArc Webinar - 2024 Global Privacy SurveyTrustArc
How does your privacy program stack up against your peers? What challenges are privacy teams tackling and prioritizing in 2024?
In the fifth annual Global Privacy Benchmarks Survey, we asked over 1,800 global privacy professionals and business executives to share their perspectives on the current state of privacy inside and outside of their organizations. This year’s report focused on emerging areas of importance for privacy and compliance professionals, including considerations and implications of Artificial Intelligence (AI) technologies, building brand trust, and different approaches for achieving higher privacy competence scores.
See how organizational priorities and strategic approaches to data security and privacy are evolving around the globe.
This webinar will review:
- The top 10 privacy insights from the fifth annual Global Privacy Benchmarks Survey
- The top challenges for privacy leaders, practitioners, and organizations in 2024
- Key themes to consider in developing and maintaining your privacy program
Monitoring and Managing Anomaly Detection on OpenShift.pdfTosin Akinosho
Monitoring and Managing Anomaly Detection on OpenShift
Overview
Dive into the world of anomaly detection on edge devices with our comprehensive hands-on tutorial. This SlideShare presentation will guide you through the entire process, from data collection and model training to edge deployment and real-time monitoring. Perfect for those looking to implement robust anomaly detection systems on resource-constrained IoT/edge devices.
Key Topics Covered
1. Introduction to Anomaly Detection
- Understand the fundamentals of anomaly detection and its importance in identifying unusual behavior or failures in systems.
2. Understanding Edge (IoT)
- Learn about edge computing and IoT, and how they enable real-time data processing and decision-making at the source.
3. What is ArgoCD?
- Discover ArgoCD, a declarative, GitOps continuous delivery tool for Kubernetes, and its role in deploying applications on edge devices.
4. Deployment Using ArgoCD for Edge Devices
- Step-by-step guide on deploying anomaly detection models on edge devices using ArgoCD.
5. Introduction to Apache Kafka and S3
- Explore Apache Kafka for real-time data streaming and Amazon S3 for scalable storage solutions.
6. Viewing Kafka Messages in the Data Lake
- Learn how to view and analyze Kafka messages stored in a data lake for better insights.
7. What is Prometheus?
- Get to know Prometheus, an open-source monitoring and alerting toolkit, and its application in monitoring edge devices.
8. Monitoring Application Metrics with Prometheus
- Detailed instructions on setting up Prometheus to monitor the performance and health of your anomaly detection system.
9. What is Camel K?
- Introduction to Camel K, a lightweight integration framework built on Apache Camel, designed for Kubernetes.
10. Configuring Camel K Integrations for Data Pipelines
- Learn how to configure Camel K for seamless data pipeline integrations in your anomaly detection workflow.
11. What is a Jupyter Notebook?
- Overview of Jupyter Notebooks, an open-source web application for creating and sharing documents with live code, equations, visualizations, and narrative text.
12. Jupyter Notebooks with Code Examples
- Hands-on examples and code snippets in Jupyter Notebooks to help you implement and test anomaly detection models.
Best 20 SEO Techniques To Improve Website Visibility In SERPPixlogix Infotech
Boost your website's visibility with proven SEO techniques! Our latest blog dives into essential strategies to enhance your online presence, increase traffic, and rank higher on search engines. From keyword optimization to quality content creation, learn how to make your site stand out in the crowded digital landscape. Discover actionable tips and expert insights to elevate your SEO game.
CAKE: Sharing Slices of Confidential Data on BlockchainClaudio Di Ciccio
Presented at the CAiSE 2024 Forum, Intelligent Information Systems, June 6th, Limassol, Cyprus.
Synopsis: Cooperative information systems typically involve various entities in a collaborative process within a distributed environment. Blockchain technology offers a mechanism for automating such processes, even when only partial trust exists among participants. The data stored on the blockchain is replicated across all nodes in the network, ensuring accessibility to all participants. While this aspect facilitates traceability, integrity, and persistence, it poses challenges for adopting public blockchains in enterprise settings due to confidentiality issues. In this paper, we present a software tool named Control Access via Key Encryption (CAKE), designed to ensure data confidentiality in scenarios involving public blockchains. After outlining its core components and functionalities, we showcase the application of CAKE in the context of a real-world cyber-security project within the logistics domain.
Paper: https://doi.org/10.1007/978-3-031-61000-4_16
Full-RAG: A modern architecture for hyper-personalizationZilliz
Mike Del Balso, CEO & Co-Founder at Tecton, presents "Full RAG," a novel approach to AI recommendation systems, aiming to push beyond the limitations of traditional models through a deep integration of contextual insights and real-time data, leveraging the Retrieval-Augmented Generation architecture. This talk will outline Full RAG's potential to significantly enhance personalization, address engineering challenges such as data management and model training, and introduce data enrichment with reranking as a key solution. Attendees will gain crucial insights into the importance of hyperpersonalization in AI, the capabilities of Full RAG for advanced personalization, and strategies for managing complex data integrations for deploying cutting-edge AI solutions.
Ocean lotus Threat actors project by John Sitima 2024 (1).pptxSitimaJohn
Ocean Lotus cyber threat actors represent a sophisticated, persistent, and politically motivated group that poses a significant risk to organizations and individuals in the Southeast Asian region. Their continuous evolution and adaptability underscore the need for robust cybersecurity measures and international cooperation to identify and mitigate the threats posed by such advanced persistent threat groups.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
1. RIT Computer Science • Capstone Report • 20171
Deep Learning Innovations in Facial Analysis:
Streamlined Approaches for Expressions and
Gender Detection
Goparapu Krishna Margali
Department of Computer Science
Golisano College of Computing and Information Sciences
Rochester Institute of Technology
Rochester, NY 14586
kg4060@cs.rit.edu
Abstract—This study describes the creation of a sophisticated
facial recognition system that reliably identifies and categorizes
human emotions and gender using a convolutional neural net-
work (CNN) constructed with Keras. In a variety of settings, the
system operates in real-time and demonstrates an outstanding
93% accuracy for emotion recognition and nearly 90% for gender
determination. This is a major advancement in the application
of theoretical ideas related to emotion and gender recognition to
real-world business settings.
The system’s core capability is its ability to interpret grayscale
photographs of faces with 48 by 48 pixels in order to identify
the gender and range of human emotions. The use of the UTK
dataset, which is well-known for its diversity in age, gender, and
ethnicity, was a crucial component of this study and ensured a
well-rounded and effective training regimen. High accuracy levels
have been mostly attained through careful dataset curation and
skillful learning rate adjustment by the Adam optimizer.
Beyond the advances in technology, the study investigates
how this dual-recognition system can be integrated with CRM
software, with an emphasis on enhancing AI chatbots to provide
more meaningful customer care interactions. By tailoring ser-
vices through the real-time assessment of emotional and gender
indicators, such an integration has the potential to completely
change how customers engage with businesses. This will increase
customer satisfaction and provide deeper business insights.
This work not only demonstrates that deep learning is a viable
technological approach for gender and emotion recognition, but
it also lays the groundwork for the practical application of these
technologies in a variety of business contexts. It emphasizes
how urgent ethical practices in the application of AI, ensuring
responsible use of powerful, insight-driven recognition systems.
Index Terms—CNN, Keras, Flask, Haar Cascade Classifier,
Deep Learning, Python, Data Augmentation
I. INTRODUCTION
The capacity to precisely identify and comprehend human
emotions marks a substantial advancement in artificial intelli-
gence and machine learning, improving technology’s capacity
to comprehend and engage with people. This research presents
a novel facial recognition system that uses a convolutional
neural network (CNN) in Keras to accurately recognize and
categorize human emotions based on facial expressions. This
system not only shows how far AI has come, but it also shows
how these technologies can be used in real-world situations.
This research has its roots in the growing need for automated
systems that can engage with empathy, especially in customer-
oriented sectors[10]. This emotion recognition technology,
therefore, stands at the forefront of bridging the gap between
human emotional expression and machine interpretation. The
implications of this technology are vast and varied, offering
transformative potential in areas such as security, mental health
assessment, and, most prominently, in enhancing customer
experience and engagement.
This work was motivated by the increasing demand for
automated systems with human emotion empathy and respon-
siveness, particularly in customer-focused corporate settings.
There is a lot of promise for emotion detection technology
in a lot of areas, like security and healthcare, and especially
in improving customer service and interaction. Through the
utilization of a dataset consisting of grayscale photographs
with 48 by 48 pixels, the system has been trained to identify
and evaluate a range of human emotions. This will enable
more complex and compassionate interactions between robots
and people. This research goes beyond developing an algo-
rithm that can theoretically recognize emotions. It looks at
the potential commercial applications, particularly in terms
of enhancing customer happiness and engagement through
emotion-responsive technology[3]. firms can obtain deeper
insights into client behaviors and preferences by incorporating
this technology into AI chatbots and customer relationship
management (CRM) systems. This allows firms to provide
more effective and tailored services.
The technique, development process, performance assess-
ment, and possible business and customer service applications
of the emotion recognition system are all covered in this
article. The objective is to draw attention to the technological
advancements as well as the moral issues and real-world
difficulties associated with implementing emotion-sensitive AI
systems in a variety of contexts.
II. LITERATURE REVIEW
Over the past two decades, there have been considerable
breakthroughs in the field of facial expression and gender
Rochester Institute of Technology 1 | P a g e
2. RIT Computer Science • Capstone Report • 20171
detection, mostly due to the development of machine learning
and computer vision technology.
A. Evolution of Facial and Gender Detection
The precision and usefulness of early systems were con-
strained, and they mostly depended on simple pattern recog-
nition. A significant advancement in the discipline was the
introduction of increasingly complex algorithms, particularly
Convolutional Neural Networks (CNNs)[4]. One of the main
components of our study, the precise and thorough real-time
analysis of facial traits, has been made possible by these
breakthroughs. CNNs are perfect for in-depth facial analysis
because of their deep layered design, which makes them excel-
lent at extracting and learning complicated characteristics from
picture data. This ability is essential for correctly identifying
minute differences in face expressions and identifying gender
traits.
The Haar Cascade Classifier has become an essential
technique for real-time face feature recognition alongside
CNNs[22]. This method, which was first created for face
detection, effectively recognizes facial features, which is a
necessary step for gender categorization and expression recog-
nition. In real-time image analysis applications, such as the
ones we’re working on, Haar Cascade Classifiers are a com-
mon fixture because of their quick processing speed and high
precision feature detection.
The application of facial recognition has expanded dramat-
ically as a result of the combination of various technologies.
The capacity to precisely identify and evaluate gender and
facial expressions in real time has created new opportunities in
a variety of fields, including interactive marketing and security
systems. Utilizing these developments, our project applies
them to the novel context of consumer behavior analysis and
business effect[11], demonstrating the impact and versatility
of these technologies.
B. Contribution of Deep Learning Frameworks
The spread of deep learning technology has been greatly
aided by TensorFlow’s emergence as an extensive, open-source
platform. TensorFlow has made it possible to create more
advanced and precise face recognition systems by providing
a stable and adaptable framework for creating complicated
models[8]. As a high-level TensorFlow API, Keras makes
building and training deep neural networks easier. The success
of our project is largely due to its user-friendly interface, which
makes model development more accessible and effective.
Data augmentation has been essential in addressing the
problems caused by dataset limits. With this method, the
training dataset is artificially expanded by applying different
image modifications, such as flipping, rotating, or scaling.
It improves the model’s capacity to generalize and function
correctly on untested data by doing this. In the context of
facial recognition, where differences in lighting, orientation,
and facial features can have a major impact on performance,
this approach is especially helpful.
According to the paper[23], the Adam optimizer is a step
forward in the field of deep learning. Through the adaptation
of the learning rate for each parameter, it optimizes the training
process, resulting in a faster and more efficient convergence of
the model. With the help of this optimizer, our model has been
able to quickly adapt to the subtleties of gender and facial
expression recognition, resulting in high levels of efficiency
and accuracy in real-time applications.
C. Evolution of Data Augmentation
In the realm of machine learning, data augmentation has
become an essential approach, especially for image processing
applications like facial recognition. The first difficulty with
picture classification jobs was the small amount of data that
could be used to efficiently train models. In order to over-
come this, data augmentation created artificial enhancements
to datasets using several transformations, including picture
flipping, scaling, and rotation[19]. This method increases the
amount of training data while simultaneously adding variety,
which improves the models’ ability to generalize to new,
unobserved data.
Data augmentation is essential for addressing the difficulties
presented by a wide range of facial expressions and attributes
in facial identification. The way photographs are lit, oriented,
and backgrounded can have a big impact on how well facial
recognition models work. Data augmentation guarantees that
the models are strong enough to manage real-world complex-
ities and are not overfitted to the limited scenarios provided in
the training dataset by performing transformations that mirror
these variances found in real-world situations.
Different lighting conditions, angles, and partial obstruc-
tions can affect the appearance of human faces. A crucial
prerequisite for the dependability and security of such applica-
tions is the use of data augmentation approaches to guarantee
that our model continues to be correct and successful in a
variety of unpredictable real-world situations.
D. Emergence of Web-Based AI Applications
A major step forward in the accessibility and usability of
machine learning applications is the integration of AI models
with web interfaces. With the development of lightweight web
frameworks like as Flask, which made it easier to deploy
complex models to web environments, this integration became
more and more possible[16]. Python-based Flask is a micro-
framework that provides the simplicity and flexibility required
for rapid deployment without requiring deep knowledge of
web programming. This method has made it easier for aca-
demics and developers to share their models with a wider
audience, democratizing the usage of AI.
Real-time interaction with AI models is made possible by
their deployment on web interfaces. This is an essential feature
for applications that need to get feedback instantly, including
gender and facial expression recognition. The practical utility
of these models is greatly increased when they can analyze
and present the findings in real-time on a web interface. This
Rochester Institute of Technology 2 | P a g e
3. RIT Computer Science • Capstone Report • 20171
makes them more suitable for dynamic contexts such as retail
shop customer behavior tracking.
Remote access and monitoring are made possible by the
usage of web interfaces in AI systems. This means that,
in terms of road safety, different stakeholders—like urban
planners, traffic monitors, and road safety officials—can access
the model from any location by having it deployed on a central
server. Its remote accessibility is necessary for a broad and
adaptable use of the technology in diverse urban settings.
III. ANALYSIS
The rapid rate at which artificial intelligence is developing
raises concerns about society’s technological capability to
integrate these technologies into daily life. The adoption of
these technologies by the general public is essential to their
successful use.
A. Societal Impact and Legal Considerations
There are wider social issues that are raised by the use
of gender and facial expression recognition technology in
automobiles. The effect on privacy and the possibility of
spying are two such worries. Our project must work around
these changing legal frameworks, which are still debating
the ramifications of using such biometric data globally[7].
Strong data protection and anonymization mechanisms inside
the system are justified when weighed against the possible
advantages of increased traffic safety.
Furthermore, concerns about accountability and liability
arise when AI is used in delicate settings like retail estab-
lishments. Setting up distinct lines of accountability is crucial
in the event that an accident is caused by a system malfunction
or misunderstanding. This necessitates both legal clarity and
technical safeguards. These legal factors are taken into account
during the project’s development to make sure the system
complies with the strictest moral guidelines.
B. Integration Challenges
In our initiative, we take into account how prepared different
stakeholders are to work with AI and rely on it for important
jobs like road safety. These stakeholders include corporations,
law enforcement, and urban planners. There are integration
problems with the existing infrastructure as well[6]. The uni-
form deployment of AI-based safety systems is hampered by
the variability of vehicles on the road in terms of technology
and the quality of the road conditions. In order to address
these issues, the project suggests a staged integration strategy
that starts with high-risk regions or more recent car models
that have better compatibility with available technologies. In
our analysis, we look at ways to keep consumers involved
and accountable, like user feedback platforms and awareness
campaigns.
IV. HYPOTHESIS
We postulate that a sophisticated system for detecting gen-
der and facial expressions can reliably determine a person’s
gender in real time by using a multi-layered Convolutional
Neural Network (CNN). Because it provides data that can
guide decisions about traffic management and road construc-
tion, it is anticipated that this system will have a major positive
impact on road safety[4]. With a CNN architecture designed
to handle the intricacies of real-time video data and a wide
range of human characteristics, the model ought to perform
well in a variety of scenarios and demographics.
We also hypothesize that the accuracy and dependability
of the system in a variety of uncertain real-world contexts
will be improved by merging this CNN with real-time data
augmentation and adaptive optimization approaches, like the
Adam optimizer. A more comprehensive understanding of
client behavior will be possible with the ability to interpret
live video feeds and identify minute details in face expressions,
especially in less-than-ideal lighting and weather conditions.
The deployment of the trained model via a Flask web
interface, which will allow stakeholders in road safety and
urban planning to remotely monitor and evaluate consumer
behavior data, is the last component of our hypothesis.
V. IMPLEMENTATION
The system’s basic architecture is based on fostering a
cordial relationship between the user and the AI in order
to provide a smooth and natural user experience. Accuracy,
efficiency, and scalability were the three main design ideas that
guided the system’s conceptualization in order to accomplish
this.
A. System Design
The way the system is designed demonstrates how effi-
ciency and accuracy may coexist when artificial intelligence is
approached from the perspective of the user. Fundamentally,
the convolutional neural network (CNN) was designed to
precisely examine and decipher the nuances of gender and
human emotions from facial expressions, guaranteeing great
processing speed and accuracy. This was made possible by
a strong preprocessing pipeline that creates uniform input
data standards and consistent model performance[14]. The
architecture of the system is not just built for the present,
but it is also scalable, allowing for future improvements like
new demographic features or emotion categories to be added
without the need for a complete redesign.
Our dedication to developing an ethical AI system was
fundamental to our design philosophy. Because of this, the
design complies with strict data protection regulations and
includes strong security mechanisms to safeguard user data.
This proactive attitude to ethical and privacy concerns estab-
lishes a new benchmark for responsible AI development[20].
Additionally, the Flask-powered backend infrastructure of the
system is designed to manage large volumes of data, guaran-
teeing that it can grow to accommodate the needs of diverse
deployment contexts, such as interactive digital signage and
customer support platforms, while upholding user confidence
and system integrity.
Rochester Institute of Technology 3 | P a g e
4. RIT Computer Science • Capstone Report • 20171
Fig. 1. Solution Architecture of Customer Profiling
B. System Architecture
1) Data Input and Preprocessing: In order to lower com-
puting demands, 48x48 pixel facial photos are transformed to
grayscale before being input into the pipeline. Preprocessing
involves scaling the pixel values in an image between 0 and
1, known as image normalization, and applying augmentation
techniques like rotation and width shift to strengthen the
model’s resistance to overfitting and enhance its capacity to
generalize across different face orientations and dimensions.
2) Convolutional Neural Networks (CNN): Four
convolutional blocks are stacked one after the other to
form the CNN. The convolutional layer uses a series of
learnable filters to capture spatial hierarchies, and each
block is meant to execute feature extraction. To standardize
the inputs to the following layer, batch normalization is
used. This speeds up training and lessens the sensitivity
to network initialization[23]. Complex pattern learning is
made possible by the non-linearity introduced by the ReLU
activation function. By reducing the spatial dimensions of the
convolutional layer’s output, max pooling lowers the number
of parameters. By randomly changing a portion of input
units to zero at each training update, dropout is positioned to
prevent overfitting.
3) Fully Connected Layers: The network architecture
changes into two tightly connected layers after convolutional
blocks[20], which act as a classifier by interpreting the
features that the convolutions and pooling layers have
extracted.
4) Output Layer: A softmax activation function is used
by the CNN’s last layer to classify emotions into one of
many predetermined classifications. Since gender detection is
a binary classification, a sigmoid activation function is used.
5) Training and Validation: The Adam optimizer is utilized
to train the model due to its ability to automatically adjust
the learning rate, resulting in a fast and effective convergence
of the model. In order to capture the best-performing model
state, model checkpoints are deliberately incorporated during
Fig. 2. Architectural Overview of CNN
training to save the model weights at times when the
validation accuracy increases.
6) Integration with Flask: The trained model is served
by a lightweight web application built with the Flask
framework[16]. The program is made to take in live video
feeds or uploaded images, process the input using CNN, and
output gender and emotion predictions.
7) Model Deployment: After a picture is received, the
system preprocesses it to make sure it meets the training
requirements before sending it to CNN. After processing the
image, the model outputs the binary gender categorization and
the probability of each emotion. After that, these forecasts are
prepared into a JSON framework so that client apps can easily
use them.
The system is designed to manage real-time data streams,
it may be used for a variety of purposes, including targeted
content distribution in marketing campaigns, surveillance
with emotion and gender analysis, and interactive digital signs.
8) Gender Detection Enhancement: The UTK dataset,
which is renowned for its demographic diversity and contains
gender labels, was added to the system to enable gender
detection. Using this dataset, CNN was modified to learn
gender traits without sacrificing its capacity to identify differ-
ent emotions[11]. A meticulous assessment was conducted to
maintain equilibrium in the dual-task learning process, guaran-
teeing that the incorporation of gender detection did not hinder
Rochester Institute of Technology 4 | P a g e
5. RIT Computer Science • Capstone Report • 20171
Fig. 3. End-to-End Model Architecture for Prediction
the initial emotion recognition abilities. The effectiveness of
the extended model in a multi-task learning situation was
validated by a battery of tests that confirmed the incorporation
of gender detection maintained a high degree of accuracy.
VI. PERFORMANCE EVALUATION
A set of experiments was used to evaluate the accuracy and
loss metrics of the model across a predetermined number of
training epochs in order to assess the performance of the facial
expression and gender detection system. Accuracy and loss
curves, which plot these metrics for the training and validation
datasets across epochs, were used to visualize the learning
process of the model[6]. As a direct measure of the predictive
performance of the model, accuracy estimates the percentage
of all predictions that are accurate, while loss quantifies the
difference between the values that were predicted and the
actual values.
A. Evaluation Methodology
1) Loss Function for Classification Tasks: A thorough
technique that prioritized both qualitative and quantitative
metrics was used to objectively assess the emotion and gender
detection model’s effectiveness. The model was first trained
using the FER 2013 and UTK datasets, where it was trained
to identify different patterns linked to different emotions
and genders, respectively. In order to make sure that the
learning was applied and not just memorization, the model
was evaluated in the validation phase using a different set of
data that it had not seen during training.
After every epoch, the accuracy and loss on the training and
validation datasets were computed to perform a quantitative
evaluation[14]. The ratio of accurate forecasts to total
predictions made served as a measure of accuracy. For the
multi-class emotion detection task, categorical cross-entropy
was used to calculate the loss, which represents the model’s
prediction mistakes; for the binary gender classification task,
binary cross-entropy was employed. These loss functions,
which measure the difference between the expected results
and the model’s predictions, are especially well-suited for
classification tasks. These loss functions are particularly suited
for classification problems as they quantify the difference
between the expected outcomes and the predictions made by
the model.
2) Cross-Validation for Generalization: A further analysis
of the model’s generalizability and performance under various
settings was conducted using k-fold cross-validation, a
technique that repeats the training and validation phases k
times using distinct subsets of the data. The performance of
the model was then estimated more reliably by averaging
the results for each fold. This approach lessens variability
and offers a more thorough comprehension of the predictive
capacity of the model.
3) Real-Time Performance Testing: Additionally, real-time
testing were carried out to evaluate the model’s practical appli-
cability. This entailed putting the model to use in a virtualized
setting where it analyzed real-time video feeds and instantly
predicted gender and mood[11]. In order to show the model’s
resilience and efficacy in a real-time setting—a critical step
for applications like interactive systems and surveillance—this
step was necessary.
B. Performance Metrics
The accuracy and loss measures were the two main
metrics utilized to assess the model. The number of accurate
Rochester Institute of Technology 5 | P a g e
6. RIT Computer Science • Capstone Report • 20171
predictions divided by the total number of inputs assessed
was used to measure accuracy, which gave a clear indicator of
how well the model classified the data. The prediction error
of the model was measured using the loss function, namely
binary cross-entropy for gender detection and categorical
cross-entropy for emotion detection. In order to minimize
loss during the training process, a lower value of loss denotes
greater performance.
1) Accuracy: The most logical performance indicator is
accuracy, which measures how well the model classifies
emotions and gender. The ratio of accurately anticipated
observations to the total number of observations is its
definition[9]. Accuracy is a key performance parameter
for the emotion and gender detection tasks, demonstrating
the model’s capacity to identify and decipher the intricate
patterns present in facial data. High accuracy means that
the model can accurately read gender and facial expressions
from the datasets that were utilized. This is important for
real-world applications because inaccurate predictions could
cause misunderstandings.
2) Loss Function: Loss functions play a crucial role in
neural network training by giving an indication of the model’s
error and, consequently, the efficiency of the learning process.
Two different kinds of loss functions were used for this model.
3) Categorical Cross-Entropy: This loss function evaluates
the effectiveness of a classification model whose output is a
probability value between 0 and 1, and it is employed for
the model’s emotion detection component. Cross-entropy loss
is perfect for multi-class classification, when the result can
fall into any of several categories. It grows as the projected
likelihood deviates from the actual label.
4) Binary Cross-Entropy: Binary cross-entropy is a loss
function used in binary classification models that is utilized
for the gender detection problem. It works especially well in
models where every class is autonomous, like gender, which
is usually classified as either male or female.
The model’s capacity to gain knowledge from the training
set of data depends on both loss functions[13]. During training,
the model is guided toward more accurate predictions by
minimizing these values. An further measure of the model’s
stability and maturity over training epochs is the convergence
of loss values.
VII. RESULTS
A. Real-time Classification Results
Extensive real-time testing was conducted to validate the
emotion and gender identification model’s practical effective-
ness. The model’s high degree of accuracy—93% for emotion
recognition and 90% for gender detection—was evident in the
sample photographs. These figures demonstrate how well the
model can decipher intricate facial expressions and accurately
determine gender.
Fig. 4. Loss of the Trained Model
In the above Fig 4, the model properly recognized the
gender as ”Female” and correctly identified the emotion as
”Disgust.” This degree of precision held true for a range
of emotional states, as demonstrated by the later successful
identifications of ’Surprise’ and ’Sad’ facial expressions in
distinct persons, all while correctly determining the gender.
The dependability of the model was demonstrated in a more
dynamic context, where it correctly classified the gender of
two people while concurrently identifying the emotions of
”Happy” and ”Sad” in the same frame.
These outcomes came from the system’s processing of
images, which involved detecting each subject’s face and
drawing a bounding box to enclose the facial region. The
model utilized its trained CNN to extract features related
to gender and emotion within these limitations, yielding the
presented classifications. The model achieved a consistent
reduction in loss and a plateau in accuracy at an advanced
training stage, indicating strong performance during these real-
time testing, as seen by the accuracy and loss curves shown
in Fig. 5.
These real-world test examples show especially encouraging
performance, demonstrating the model’s capacity to work
reliably and effectively in a variety of unpredictable contexts.
This is crucial for use in applications like retail customer ana-
lytics, where knowing demographic information and customer
sentiment may greatly improve the customer experience.
Rochester Institute of Technology 6 | P a g e
7. RIT Computer Science • Capstone Report • 20171
Fig. 5. Accuracy of the Trained Model
B. Performance Results
1) Accuracy: The metrics of accuracy and loss, which are
both essential markers of predicted success in classification
tasks, were used to assess the model’s performance. During
training, the model recognized and classified emotions with a
93% accuracy, and it detected gender with a noteworthy 90%
accuracy. These findings came from analyzing the learning
curves, as shown in Fig. 5, which shows the model’s perfor-
mance over a period of 15 epochs in both tasks.
Performance on the training data shows a steady improve-
ment in the accuracy plots, and the validation accuracy shows
a similar rising trend, albeit with some expected volatility.
This variance shows how well the model generalizes to fresh,
untested data, which is essential for practical use. Interestingly,
the validation accuracy of the model peaks in later epochs,
indicating that the biggest learning gains happen in the early
stages of training.
2) Loss Functions: The training and validation datasets’
loss curves shown in the Fig. 6 a consistent downward trend,
which indicates that the model becomes better over time at
minimizing error. Nonetheless, a noted rise in validation loss in
subsequent epochs indicates that close observation is required
to avoid overfitting. This is where the use of model checkpoints
comes in very handy, enabling the best model state in terms
of validation accuracy to be restored.
These outcomes highlight the model’s ability to classify
gender and emotions with high fidelity, which is further
corroborated by testing scenarios conducted in real time. The
system proved its reliability and accuracy in real-world uses,
like interactive digital signage and surveillance, offering quick
and precise classifications that can improve user experience
and inform.
Fig. 6. Loss of the Trained Model
It has been demonstrated that integrating binary cross-
entropy for gender detection and categorical cross-entropy
loss for emotion detection is a useful strategy for reducing
the model’s sensitivity to each class. The model’s capacity
to sustain high accuracy on both tasks at the same time is
evidence of how well the dual-task learning technique used in
the training phase worked.
VIII. FUTURE WORK
The current model’s success in detecting emotion and
gender paves the way for a host of improvements and in-
vestigations in further research. The dataset’s diversity and
volume should be increased as soon as possible, since this will
probably enhance the model’s functionality and generalizabil-
ity across a range of environmental factors and demographic
groupings. Furthermore, the use of multimodal input, such text
and audio, may result in a more thorough comprehension of
emotional states, opening the door for a multifaceted method
of emotion recognition.
Subsequent advancements could employ sequential neural
network models to investigate the temporal dimensions of
emotion. These models provide a dynamic viewpoint on
emotional shifts by capturing the changes in facial expressions
across time. A more comprehensive and modern concept of
gender is reflected in the improvement of gender classification
to include a range of gender identities. Working together with
interdisciplinary groups made up of technologists, sociologists,
ethicists, and psychologists could improve the development
process even further and guarantee that the technology de-
velops in a way that is morally and socially acceptable in
addition to advancing in capabilities. By working together, it
will be easier to navigate the intricate social environments that
Rochester Institute of Technology 7 | P a g e
8. RIT Computer Science • Capstone Report • 20171
these technologies are introduced into and make sure that they
improve social dynamics rather than cause them to change.
IX. CONCLUSION
The development and implementation of this facial recogni-
tion system’s creation and application mark a substantial ad-
vancement in artificial intelligence, especially in the complex
areas of gender and emotion detection. The model highlights
the potential of deep learning algorithms in interpreting com-
plex human expressions and their practicality in real-world
circumstances, all while achieving high accuracy levels. The
model’s ability to successfully classify emotions and gender
in real-time highlights its transformative potential as a tool for
improving user experiences in a variety of businesses.
The FER 2013 and UTK datasets, the system performs
robustly, demonstrating the value of extensive and varied
training data in the development of objective and efficient AI
models. In addition, the implementation of key checkpoints
and the Adam optimizer during training demonstrates the
depth of contemporary machine learning techniques while
maintaining the accuracy and efficiency of the system.
The potential uses of this technology are enormous as we
look to the future; they range from better security and mental
health evaluation to targeted advertising. By giving businesses
insights into customer emotions and preferences, this system’s
integration with CRM software and AI chatbots has the po-
tential to completely transform customer care by empowering
them to respond to customers in a more customized and
sympathetic manner.
However, the immense power also entails great responsibil-
ity. One cannot stress the ethical issues that surround the use
of such technology. To prevent abuse and safeguard individual
rights, it is essential that the system’s development be directed
by strict ethical guidelines and privacy laws as it continues to
evolve.
To sum up, this initiative establishes the foundation for
future advances in AI that are both technologically and socially
responsible. It acts as a springboard for more advanced, per-
ceptive, moral AI systems that can relate to and comprehend
the intricacies of human behavior.
REFERENCES
[1] Chavali, T., Kandavalli, C. T., Sugash, T. M., Subramani, R. (2023).
Smart Facial Emotion Recognition With Gender and Age Factor Estima-
tion. Procedia Computer Science, 218, 113-123.
[2] Mellouk, W., Handouzi, W. (2020). Facial emotion recognition using
deep learning: review and insights. Procedia Computer Science, 175, 689-
694.
[3] S. Pandey, S. Handoo and Yogesh, ”Facial Emotion Recognition us-
ing Deep Learning,” 2022 International Mobile and Embedded Tech-
nology Conference (MECON), Noida, India, 2022, pp. 248-252, doi:
10.1109/MECON53876.2022.9752189.
[4] Happy, S. L. et al. “A real time facial expression classification system
using Local Binary Patterns.” 2012 4th International Conference on
Intelligent Human Computer Interaction (IHCI) (2015): 1-5.
[5] C. Szegedy et al., ”Going deeper with convolutions,” 2015 IEEE Confer-
ence on Computer Vision and Pattern Recognition (CVPR), Boston, MA,
USA, 2015, pp. 1-9, doi: 10.1109/CVPR.2015.7298594.
[6] Trochidis, Konstantinos Tsoumakas, Grigorios Kalliris, George Vla-
havas, I.. (2008). Multilabel classification of music into emotions. Proc.
9th International Conference on Music Information Retrieval (ISMIR
2008).
[7] Pham, Luan et al. “Facial Expression Recognition Using Residual Mask-
ing Network.” 2020 25th International Conference on Pattern Recognition
(ICPR) (2021): 4513-4519.
[8] Pramerdorfer, Christopher and M. Kampel. “Facial Expression Recog-
nition using Convolutional Neural Networks: State of the Art.” ArXiv
abs/1612.02903 (2016): n. pag.
[9] K. He, X. Zhang, S. Ren and J. Sun, ”Deep Residual Learning for Image
Recognition,” 2016 IEEE Conference on Computer Vision and Pattern
Recognition (CVPR), Las Vegas, NV, USA, 2016, pp. 770-778, doi:
10.1109/CVPR.2016.90.
[10] Kim, Bo-Kyeong et al. “Fusing Aligned and Non-aligned Face Infor-
mation for Automatic Affect Recognition in the Wild: A Deep Learning
Approach.” 2016 IEEE Conference on Computer Vision and Pattern
Recognition Workshops (CVPRW) (2016): 1499-1508.
[11] Xiong, Xuehan and Fernando De la Torre. “Supervised Descent Method
and Its Applications to Face Alignment.” 2013 IEEE Conference on
Computer Vision and Pattern Recognition (2013): 532-539.
[12] Tong Zhang, Zhulin Liu, Xue-Han Wang, Xiao-Fen Xing, C. L.
Philip Chen, and Enhong Chen. 2018. Facial Expression Recognition
via Broad Learning System. In 2018 IEEE International Conference
on Systems, Man, and Cybernetics (SMC). IEEE Press, 1898–1902.
https://doi.org/10.1109/SMC.2018.00328
[13] L. A. Jeni, J. M. Girard, J. F. Cohn and T. Kanade, ”Real-time dense 3D
face alignment from 2D video with automatic facial action unit coding,”
2015 11th IEEE International Conference and Workshops on Automatic
Face and Gesture Recognition (FG), Ljubljana, Slovenia, 2015, pp. 1-1,
doi: 10.1109/FG.2015.7163165.
[14] Cohn, Jeffrey F. and Michael A. Sayette. “Spontaneous facial expression
in a small group can be automatically measured: An initial demonstra-
tion.” Behavior Research Methods 42 (2010): 1079-1086.
[15] Kim, Bo-Kyeong et al. “Hierarchical committee of deep convolutional
neural networks for robust facial expression recognition.” Journal on
Multimodal User Interfaces 10 (2016): 173-189.
[16] Wang, Haopeng et al. “Deep Learning (DL)-Enabled System for Emo-
tional Big Data.” IEEE Access 9 (2021): 116073-116082.
[17] Anand, M and Dr. S. Babu. “A Comprehensive Investigation on Emo-
tional Detection in Deep Learning.” International Journal of Scientific
Research in Computer Science, Engineering and Information Technology
(2022): n. pag.
[18] Subramanian, R. Raja et al. “Design and Evaluation of a Deep Learning
Algorithm for Emotion Recognition.” 2021 5th International Conference
on Intelligent Computing and Control Systems (ICICCS) (2021): 984-
988.
[19] Shit, Sahadeb et al. “Real-time emotion recognition using end-to-end
attention-based fusion network.” Journal of Electronic Imaging 32 (2023):
013050 - 013050.
[20] Debnath, Tanoy et al. “Four-layer ConvNet to facial emotion recognition
with minimal epochs and the significance of data diversity.” Scientific
Reports 12 (2021): n. pag.
[21] Chauhan, Kartik et al. “BhavnaNet: A Deep Convolutional Neural
Network for Facial Emotion Recognition.” 2022 International Conference
on Computational Intelligence and Sustainable Engineering Solutions
(CISES) (2022): 576-581.
[22] Ko, ByoungChul. “A Brief Review of Facial Emotion Recognition Based
on Visual Information.” Sensors (Basel, Switzerland) 18 (2018): n. pag.
[23] Prabaswera, Dwi Redjeki and Haryono Soeparno. “FACIAL EMOTION
RECOGNITION USING CONVOLUTIONAL NEURAL NETWORK
BASED ON THE VISUAL GEOMETRY GROUP-19.” Jurnal TAM
(Technology Acceptance Model) (2023): n. pag.
Rochester Institute of Technology 8 | P a g e