Explainable AI - making ML and DL models more interpretableAditya Bhattacharya
Abstract –
Although industries have started to adopt AI and Machine Learning in almost every sector to solve complex business problems, but are these models always trustworthy? Machine Learning models are not any oracle but rather are scientific methods and mathematical models which best describes the data. But science is all about explaining complex natural phenomena in the simplest way possible! So, can we make ML and DL models more interpretable, so that any business user can understand these models and trust the results of these models?
In order to find out the answer, please join me in this session, in which I will take about concepts of Explainable AI and discuss its necessity and principles which help us demystify black-box AI models. I will be discussing about popular approaches like Feature Importance, Key Influencers, Decomposition trees used in classical Machine Learning interpretable. We will discuss about various techniques used for Deep Learning model interpretations like Saliency Maps, Grad-CAMs, Visual Attention Maps and finally go through more details about frameworks like LIME, SHAP, ELI5, SKATER, TCAV which helps us to make Machine Learning and Deep Learning models more interpretable, trustworthy and useful!
Exploration of the Hidden Influential Factors on Crime.pptxShivaprasad787526
a briefing on Exploration of the Hidden Influential Factors on Crime with different models used in exploring the crime factors and with a cased study on certain location with the use of dataset of 1515 features pertaining the influence to crime and few stats on features.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
Human activity recognition with self-attentionIJECEIAES
In this paper, a self-attention based neural network architecture to address human activity recognition is proposed. The dataset used was collected using smartphone. The contribution of this paper is using a multi-layer multi-head self-attention neural network architecture for human activity recognition and compared to two strong baseline architectures, which are convolutional neural network (CNN) and long-short term network (LSTM). The dropout rate, positional encoding and scaling factor are also been investigated to find the best model. The results show that proposed model achieves a test accuracy of 91.75%, which is a comparable result when compared to both the baseline models.
Explainable AI - making ML and DL models more interpretableAditya Bhattacharya
Abstract –
Although industries have started to adopt AI and Machine Learning in almost every sector to solve complex business problems, but are these models always trustworthy? Machine Learning models are not any oracle but rather are scientific methods and mathematical models which best describes the data. But science is all about explaining complex natural phenomena in the simplest way possible! So, can we make ML and DL models more interpretable, so that any business user can understand these models and trust the results of these models?
In order to find out the answer, please join me in this session, in which I will take about concepts of Explainable AI and discuss its necessity and principles which help us demystify black-box AI models. I will be discussing about popular approaches like Feature Importance, Key Influencers, Decomposition trees used in classical Machine Learning interpretable. We will discuss about various techniques used for Deep Learning model interpretations like Saliency Maps, Grad-CAMs, Visual Attention Maps and finally go through more details about frameworks like LIME, SHAP, ELI5, SKATER, TCAV which helps us to make Machine Learning and Deep Learning models more interpretable, trustworthy and useful!
Exploration of the Hidden Influential Factors on Crime.pptxShivaprasad787526
a briefing on Exploration of the Hidden Influential Factors on Crime with different models used in exploring the crime factors and with a cased study on certain location with the use of dataset of 1515 features pertaining the influence to crime and few stats on features.
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
The papers for publication in The International Journal of Engineering& Science are selected through rigorous peer reviews to ensure originality, timeliness, relevance, and readability.
Human activity recognition with self-attentionIJECEIAES
In this paper, a self-attention based neural network architecture to address human activity recognition is proposed. The dataset used was collected using smartphone. The contribution of this paper is using a multi-layer multi-head self-attention neural network architecture for human activity recognition and compared to two strong baseline architectures, which are convolutional neural network (CNN) and long-short term network (LSTM). The dropout rate, positional encoding and scaling factor are also been investigated to find the best model. The results show that proposed model achieves a test accuracy of 91.75%, which is a comparable result when compared to both the baseline models.
A neural is a system hardware or software that is patterned to function and was named after the neurons in the brains of humans. A neural network is known to involve several huge processors that are arranged and work in the parallel format for effectiveness. Applications include: handwriting recognition, oil data analysis, weather forecast prediction and face recognition.
Human Emotion Recognition using Machine Learningijtsrd
It is quite interesting to recognize the human emotions in the field of machine learning. Using a person's facial expression one can know his emotions or what the person wants to express. But at the same time it's not easy to recognize one's emotion easily its quite challenging at times. Facial expression consist of various human emotions such as sad, happy , excited, angry, frustrated and surprise. Few years back Natural language processing was used to detect the sentiment from the text and then it took a step forward towards emotion detection. Sentiments can be positive, negative or neutral where as emotions are more refined categories. There are many techniques used to recognize emotions. This paper provides a review of research work carried out and published in the field of human emotion recognition and various techniques used for human emotions recognition. Prof. Mrs. Dhanamma Jagli | Ms. Pooja Shetty "Human Emotion Recognition using Machine Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd25217.pdfPaper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/25217/human-emotion-recognition-using-machine-learning/prof-mrs-dhanamma-jagli
call for papers, research paper publishing, where to publish research paper, journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJEI, call for papers 2012,journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, research and review articles, engineering journal, International Journal of Engineering Inventions, hard copy of journal, hard copy of certificates, journal of engineering, online Submission, where to publish research paper, journal publishing, international journal, publishing a paper, hard copy journal, engineering journal
The Smart Way to Invest in Artificial Intelligence and Machine Learning: Lisha Li, Amplify Partners
AI and ML are seeping into every startup, at least into every pitch deck. But what does it mean to build an AI/ML company? Some startups do require a closet filled with five PhD’s in data science, but that doesn’t necessarily mean yours does. Building intelligently with AI and ML.
Handwriting identification using deep convolutional neural network methodTELKOMNIKA JOURNAL
Handwriting is a unique thing that produced differently for each person. Handwriting has a characteristic that remain the same with single writer, so a handwriting can be used as a variable in biometric systems. Each person have a different form of handwriting style but with a small possibility that same characters have something commons. We propose a handwriting identification method using sentence segmented handwriting forms. Sentence form is used to get more complete handwriting characteristics than using a single characters or words. Dataset used is divided into three categories of images, binary, grayscale, and inverted binary. All datasets have same image with different in color and consist of 100 class. Transfer learning used in this paper are pre-trained model VGG19. Training was conducted in 100 epochs. Highest result is grayscale images with genuince acceptance rate of 92.3% and equal error rate of 7.7%.
Explains how deep learning creates howlers using commonly used annotation tools for images. We have identified several such howlers. Essentially, this presentation outlines the deficiencies of deep learning networks. We also explain the theoretical reasoning for these, building on Bengio's recent paper. The presentation also contains solutions which address these gaps, such as capsule networks, transfer learning, meta-learning and federated learning.
This document will examine issues pertaining to feature extraction, classification and prediction. It will
consider the application of these techniques to unlabelled Electroencephalogram (E.E.G.) data in an
attempt to discriminate between left and right hand imagery movements
A neural is a system hardware or software that is patterned to function and was named after the neurons in the brains of humans. A neural network is known to involve several huge processors that are arranged and work in the parallel format for effectiveness. Applications include: handwriting recognition, oil data analysis, weather forecast prediction and face recognition.
Human Emotion Recognition using Machine Learningijtsrd
It is quite interesting to recognize the human emotions in the field of machine learning. Using a person's facial expression one can know his emotions or what the person wants to express. But at the same time it's not easy to recognize one's emotion easily its quite challenging at times. Facial expression consist of various human emotions such as sad, happy , excited, angry, frustrated and surprise. Few years back Natural language processing was used to detect the sentiment from the text and then it took a step forward towards emotion detection. Sentiments can be positive, negative or neutral where as emotions are more refined categories. There are many techniques used to recognize emotions. This paper provides a review of research work carried out and published in the field of human emotion recognition and various techniques used for human emotions recognition. Prof. Mrs. Dhanamma Jagli | Ms. Pooja Shetty "Human Emotion Recognition using Machine Learning" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd25217.pdfPaper URL: https://www.ijtsrd.com/computer-science/artificial-intelligence/25217/human-emotion-recognition-using-machine-learning/prof-mrs-dhanamma-jagli
call for papers, research paper publishing, where to publish research paper, journal publishing, how to publish research paper, Call For research paper, international journal, publishing a paper, IJEI, call for papers 2012,journal of science and technology, how to get a research paper published, publishing a paper, publishing of journal, research and review articles, engineering journal, International Journal of Engineering Inventions, hard copy of journal, hard copy of certificates, journal of engineering, online Submission, where to publish research paper, journal publishing, international journal, publishing a paper, hard copy journal, engineering journal
The Smart Way to Invest in Artificial Intelligence and Machine Learning: Lisha Li, Amplify Partners
AI and ML are seeping into every startup, at least into every pitch deck. But what does it mean to build an AI/ML company? Some startups do require a closet filled with five PhD’s in data science, but that doesn’t necessarily mean yours does. Building intelligently with AI and ML.
Handwriting identification using deep convolutional neural network methodTELKOMNIKA JOURNAL
Handwriting is a unique thing that produced differently for each person. Handwriting has a characteristic that remain the same with single writer, so a handwriting can be used as a variable in biometric systems. Each person have a different form of handwriting style but with a small possibility that same characters have something commons. We propose a handwriting identification method using sentence segmented handwriting forms. Sentence form is used to get more complete handwriting characteristics than using a single characters or words. Dataset used is divided into three categories of images, binary, grayscale, and inverted binary. All datasets have same image with different in color and consist of 100 class. Transfer learning used in this paper are pre-trained model VGG19. Training was conducted in 100 epochs. Highest result is grayscale images with genuince acceptance rate of 92.3% and equal error rate of 7.7%.
Explains how deep learning creates howlers using commonly used annotation tools for images. We have identified several such howlers. Essentially, this presentation outlines the deficiencies of deep learning networks. We also explain the theoretical reasoning for these, building on Bengio's recent paper. The presentation also contains solutions which address these gaps, such as capsule networks, transfer learning, meta-learning and federated learning.
This document will examine issues pertaining to feature extraction, classification and prediction. It will
consider the application of these techniques to unlabelled Electroencephalogram (E.E.G.) data in an
attempt to discriminate between left and right hand imagery movements
Seminar of U.V. Spectroscopy by SAMIR PANDASAMIR PANDA
Spectroscopy is a branch of science dealing the study of interaction of electromagnetic radiation with matter.
Ultraviolet-visible spectroscopy refers to absorption spectroscopy or reflect spectroscopy in the UV-VIS spectral region.
Ultraviolet-visible spectroscopy is an analytical method that can measure the amount of light received by the analyte.
Comparing Evolved Extractive Text Summary Scores of Bidirectional Encoder Rep...University of Maribor
Slides from:
11th International Conference on Electrical, Electronics and Computer Engineering (IcETRAN), Niš, 3-6 June 2024
Track: Artificial Intelligence
https://www.etran.rs/2024/en/home-english/
This pdf is about the Schizophrenia.
For more details visit on YouTube; @SELF-EXPLANATORY;
https://www.youtube.com/channel/UCAiarMZDNhe1A3Rnpr_WkzA/videos
Thanks...!
Richard's entangled aventures in wonderlandRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Professional air quality monitoring systems provide immediate, on-site data for analysis, compliance, and decision-making.
Monitor common gases, weather parameters, particulates.
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Sérgio Sacani
Since volcanic activity was first discovered on Io from Voyager images in 1979, changes
on Io’s surface have been monitored from both spacecraft and ground-based telescopes.
Here, we present the highest spatial resolution images of Io ever obtained from a groundbased telescope. These images, acquired by the SHARK-VIS instrument on the Large
Binocular Telescope, show evidence of a major resurfacing event on Io’s trailing hemisphere. When compared to the most recent spacecraft images, the SHARK-VIS images
show that a plume deposit from a powerful eruption at Pillan Patera has covered part
of the long-lived Pele plume deposit. Although this type of resurfacing event may be common on Io, few have been detected due to the rarity of spacecraft visits and the previously low spatial resolution available from Earth-based telescopes. The SHARK-VIS instrument ushers in a new era of high resolution imaging of Io’s surface using adaptive
optics at visible wavelengths.
5. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
6. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
7. ABSTRACT
• In the recent years deep learning have showed effectiveness in detecting features from large amount of data.
However, within the field of electromyography-based gesture recognition it was seldom employed as they
require an unreasonable amount of effort from a single person, to generate tens of thousands of examples.
• This work assumes that we can generate general/informative features can be generated when aggregating
signals from multiple users. thus, reducing the recording burden while enhancing gesture recognition. Using
transfer learning
• Datasets: Two datasets comprised of 19 and 17 able-bodied participants respectively (the first one is
employed for pre-training) were recorded for this work, using the Myo Armband. A third Myo Armband
dataset was taken from the NinaPro database and is comprised of 10 able-bodied participants.
• Three different deep learning networks employing three different modalities as input (raw EMG,
Spectrograms and Continuous Wavelet Transform (CWT)) are tested on the second and third dataset.
• achieving an offline accuracy of 98.31% for 7 gestures over 17 participants for the CWT-based ConvNet and
68.98% for 18 gestures over 10 participants for the raw EMG-based ConvNet.
• Finally, a use-case study employing eight able-bodied participants suggests that real-time feedback allows
users to adapt their muscle activation strategy which reduces the degradation in accuracy normally
experienced over time.
8. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
9. I. INTRODUCTION
• Importance of the prosthetics, what is EMG signal, and we can record them then use AI to control bionics.
• The literature review showed that the papers used to get features from the signals to predict the gesture,
recently deep learning was introduced. shifting the paradigm from feature engineering to feature learning.
One of the main important parameters for successful prediction using AI is the amount of data. The
traditional approach was to collect very large amount of data from a single user to allow the deep learning
model predict his gesture, but the paper propose aggregating the readings from multiple users to generalize
the prediction for everyone. Consequently, deep learning offers a particularly attractive context from which
to develop a Transfer Learning (TL) algorithm to leverage inter-user data by pre-training a model on
multiple subjects before training it on a new participant.
• As such, the main contribution of this work is to present a new TL scheme employing a convolutional
network (ConvNet) to leverage inter-user data within the context of sEMG-based gesture recognition.
• there is another paprt [7], followed that approach before. This paper continue the work, improving the TL
algorithm to reduce its computational load and improving its performance. Additionally, three new ConvNet
architectures, employing three different input modalities, specifically designed for the robust and efficient
classification of sEMG signals are presented. The raw signal, short-time Fourier transform-based
spectrogram and Continuous Wavelet Transform (CWT) are considered for the characterization of the sEMG
signals to be fed to these ConvNets. Another major contribution of this article is the publication of a new
sEMG-based gesture classification dataset comprised of 36 able-bodied participants.
10. I. INTRODUCTION
• This dataset and the implementation of the ConvNets along with their TL augmented version are made
readily available. Finally, this paper further expands the aforementioned conference paper by proposing a
use-case experiment on the effect of real-time feedback on the online performance of a classifier without
recalibration over a period of fourteen days.
• The paper organization:
• Overview of related work for gesture recognition using deep learning and transfer learning.
• Presenting the new data-sets being used, with the data acquisition and the preprocessing. Also the NinaPro DB5
dataset.
• Presentation of the different state-of-the-art feature sets employed in this work is given
• The networks’ architecture.
• Presenting the TL model used
• Comparison between the state of the art in gesture recognition
• A real-time use case.
• Results and discussion>
11. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
12. II. RELATED WORK
• EMG signals vary from subjects even with precious placement of the electrodes. That is why classifiers
trained from a user performs slightly better than random when someone else use it. But not anyway close to
precision. Therefore, many complicated techniques have been acquired to get the inter-user information.
For example, research has been one on the common features between the original subject and a new user.
Another research on finding a pretrained model to remove the need to work with data from multiple
subjects. These non deep learning approaches showed way better results than the non augmented versions.
• Short time Fourier transform (STFT) was not used a lot in the past decades for the classification of the EMG
data. Maybe the reason why it isn’t being used is the large number of features it produce that would be
computationally expensive. Moreover, STFT have also been shown to be less accurate than wavelet transform
on their own to classify EMG signals. Recently anyway the STFT features in the form of spectrograms was
applied as input feature space for classification of sEMG data by leveraging ConvNets.
• Continuous wavelet transform (CWT) features have been used for electroencephalography and EMG signal
analysis but mainly for lower limbs. Also wavelet based features have been used in the past for sEMG based
hand gesture classification. The wavelet based however that was used was on discrete waverlet transform
and the wavelet packet transform not continuous wavelet transform. Then mentions some reasons why they
used them not CWT. Anyway this paper for the first time introduces the CWT for utilizing the sEMG gesture
recognition.
• Recently the convNets have started to be used for the gesture recognition using single array and matrix of
electrodes. To the best of our knowledge, this paper, which is an extension of [7], is the first time inter-user
data is leveraged through TL for training deep learning algorithms on sEMG data.
13. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
14. III. sEMG Datasets
• A) Myo Dataset:
• One of the main contribution of this paper is to provide a new publicity available sEMG-based hand gesture
recognition dataset referred to Myo dataset. The dataset is formed from two sub-datasets one for the
training and one for the evaluation. The first subset (the former) is composed of 19 abled bodied
participents, is used to build, validate and optimize. While the second subset (the latter) is composed of 17
abled bodied data and it is used for the final testing. This is the largest myoelectric data set available.
• The data acquisition protocol was approved by the Comit´es d’ ´ Ethique de la Recherche avec des ˆetres
humains de l’Universit´e Laval (approbation number: 2017-026/21-02- 2016) and informed consent was
obtained from all participants.
1. Recording hardware:
• the dataset was recorded using Myo armband , 8 channels, dry electrode, low sampling rate (200 Hz). Low
consumer grade armband.
• the armband is simply slipped to the forearm which is way easier than gel-based electrodes which requires
shaving the skin to obtain optimal contact between the skin’s subject and electrodes.
• The limitations of the armband comes still. The dry electrodes are less accurate and robust to motion artifact
than gel based ones.
• The recommended EMG frequency is in the range of 5-500Hz. Which requires a sampling frequency minimum of
1000Hz. The myo armband is limited to 200Hz. This loss was a main reason to impact the ability of various
classifiers. As such, robust and adequate classification techniques are needed to process the collected signals
accurately.
15. III. sEMG Datasets
• 2) Time Window Length:
• For real time control in a closed loop, time latency is important factor to consider. A paper recommended a time of
300ms. Even another papers recommended time between 100-250 ms. And as the performance of the classifier is
more important than time, the window size was taken equal to 260ms was selected to achieve a reasonable numbers of
samples between each prediction due to the low frequency of myoelectric.
• 3)Labeled Data Acquisition Protocol:
• The seven hand gestures considered in the work are shown here, the labeled data were recorded by letting the user to
hold each gesture for five seconds. The data recording was only applied when the user was familiar with the four
gestures. A time of five seconds were given for the user between each gesture and this time wasn’t recorded. The
recording of the seven gestures each for five seconds is called a cycle (35 sec). With four cycles is formed a round. In
case of the pre-trained data, a single round is available per subject (140 sec per participant). For the evaluation dataset
three rounds are available with the first round utilized for training (i.e., 140s per participant) and the last two for
testing (i.e., 240s per participant).
• During the recordings, the subjects were asked to stand and hold their
forearm parallel to the floor. And the armband was tightened to avoid bias in
the same orientation shown. The dataset is formed from the row EMG
signals.
• Signal processing must be applied to efficiently train a classifier on the data
recorded by the Myo armband. The data is first separated by applying sliding
windows of 52 samples (260ms) with an overlap of 235ms (i.e. 7x190
samples for one cycle (5s of data)). Employing windows of 260ms allows
40ms for the pre-processing and classification process, while still staying
within the 300ms target.
16. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
17. IV. Classis EMG Classification
• Before using deep learning for classification, they used feature engineering where they manually finds trends
in EMG signals to perform each gesture and over years there were many different efficient ways in time and
frequency domain.
• Another thing that the paper will do is to test four of the proposed feature sets (features for manually
identifying hand gestures) from the literature review on five different classifiers (SVM, ANN, RF, K-NN and
LDA) then compare them with the proposed TL deep learning. The hyper parameters for each classifier were
selected by employing three-fold cross-validation alongside random search, testing 50 different
combinations of hyperparameters for each participant’s dataset for each classifier. The hyperparameters
considered for each classifier are presented in Appendix D.
• He made something called dimensionality reduction and showed that for most of the cases it reduced the
computational power and increased the performance.
• The four features that were selected for the comparison are:
• Time Domain Features (TD)
• Enhanced TD
• Nina Pro Features
• SampleEn pipeline
18. IV. Classis EMG Classification
• Before using deep learning for classification, they used feature engineering where they manually finds trends
in EMG signals to perform each gesture and over years there were many different efficient ways in time and
frequency domain.
• Another thing that the paper will do is to test four of the proposed feature sets (features for manually
identifying hand gestures) from the literature review on five different classifiers (SVM, ANN, RF, K-NN and
LDA) then compare them with the proposed TL deep learning. The hyper parameters for each classifier were
selected by employing three-fold cross-validation alongside random search, testing 50 different
combinations of hyperparameters for each participant’s dataset for each classifier. The hyperparameters
considered for each classifier are presented in Appendix D.
• He made something called dimensionality reduction and showed that for most of the cases it reduced the
computational power and increased the performance.
• The four features that were selected for the comparison are:
• Time Domain Features (TD)
• Enhanced TD
• Nina Pro Features
• SampleEn pipeline
19. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
20. V. Deep Learning Classifiers Overview
• As Mentioned before, due to the limitations of data from a single individual, to address the overfitting
problem, he used: Monte carlo dropout, Batch normalization and early stopping.
• A)Patch normalization, each patch of examples was normalized separately using the mean and the variance.
• B)Proposed CNN architecture
• Here he will be using a slow-fusion architecture based ConvNet in this work is due to the similarities between videos data
and the proposed characterization of sEMG signals, as both representations have analogous structures (i.e. Time x Spatial x
Spatial for videos) and can describe non-stationary information. Three architectures will be used
21. V. Deep Learning Classifiers Overview
• 1-ConvNet for Spectrograms: the spectrograms which are fed to the CNN were calculated using Hann
windows with the length 28 and an overlap of 20 yielding a matrix of 4x15. The first frequency band was
removed in an effort to reduce baseline drift and motion artifact. As the armband features eight channels,
eight such spectrograms were calculated, yielding a final matrix of 4x8x14 (Time x Channel x Frequency).
• The CNN was performed on a python framework called “Theano”
• Adam optimizer, Alpha = 0.00681, MC Dropout = 0.5, batch size= 128, early stop = 10%
22. V. Deep Learning Classifiers Overview
• 2-ConvNet for Continuous Wavelet Transforms:
• Has similar fashion as the previous one. Both the Morlet and Mexican Hat wavelet were considered for this
work and the Mexican hat performed better.
• The CWTs were calculated with 32 scales yielding a 32x52 matrix. Down sampling is firstly applied. Similar to spectrogram,
the last row of the calculated CWT was removed to reduce the motion artifacts. Additionally, the last column of the
calculated CWT was also removed as to provide an even number of time-columns from which to perform the slow-fusion
process. The final matrix shape is thus 12x8x7 (i.e. Time x Channel x Scale).
• All hyperparameters stayed the same except for Alpha = 0.08799
23. V. Deep Learning Classifiers Overview
• 3-ConvNet for raw EMG:
• This network will help assess if employing time-frequency features lead to sufficient gains in accuracy
performance to justify the increase in computational cost.
• The architecture was taken from another paper performing the same classification using PyTorch v.0.4.1.
with a change in learning rate obtained from performing cross-validation.
• Finally he made some enhancements to make the raw CNN like that:
24. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
25. VI. Transfer Learning
• Starts by saying that TL is the solution when you have few data. And he will perform something called
automatic alignment to solve the sensor’s misalignment problem from a subject to another.
• A) Progressive Neural Networks Fine-tuning is the most used TL technique which train a model source
domain, and when you have a new task, you use these trained weights as the initial weights for new tasks.
However, it suffers from many things.
• PNN is used to address these problems, it pretrain the model on a source domain and freeze its weights,
When a new task appears, a new network, with random initialization, is created and connected in a layer-
wise fashion to the original network.
• B) Adaptive Batch NormalizationIn opposition to the PNN architecture, which uses a different network for the
source and the target, AdaBatch employs the same network for both tasks. The TL occurs by freezing all the
network’s weights (learned during pre-training) when training on the target, except for the parameters
associated with BN. The hypothesis behind this technique is that the label-related
• C) Proposed Transfer Learning Architecture d
26. VI. Transfer Learning
• C) Proposed Transfer Learning Architecture
The main tenet behind TL is that similar tasks
can be completed in similar ways. The
difficulty in this paper’s context is then to
learn a mapping between the source and
target task as to leverage information learned
during pre-training.
• How he will use TL is he will train the CNN in
section v on the pre-training data (Named the
source network), then freeze the weights we
have except for BN. Then have another CNN
with nearly identical architecture (Named the
secondary network) with a different
initialization. The way the two networks are
connected is shown in the below figure. Then
we use the training dataset to train the whole
system together.
• The whole system is referred to as the target
network.
27. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
33. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
34. VIII. Case Study
• This last experiment section proposes a use-case study of real-time performance of the classifier over a
period of 14 days for eight able-bodied participants. The main goal of this use-case experiment is to
evaluate if users can self-adapt and improve the way they perform gestures based on visual feedback from
complex classifiers.
• The eight participants were randomly separated into two equal groups. With and without feedback
• The details of the test (period, duration, accuracy …) is being told.
• The cases reported muscle fatigue that is why the test was function in time.
35. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
36. IX. Discussion
• the TL augmented ConvNets significantly outperformed their nonaugmented versions regardless of the
number of training cycles.
• Increasing the number of cycles increased the performance
• Overall, the proposed TL-augmented ConvNets were competitive with the current state-of-the-art.
• Furthermore, the TL method outperformed the non-augmented ConvNets on the out-of-sample experiment. This suggests
that the proposed TL algorithm enables the network to learn features that can generalize not only across participants but
also for never-seen before gestures. As such, the weights learned from the pretraining dataset can easily be re-used for
other work that employs the Myo Armband with different gestures.
•
• This paper proposed the architecture to be two identical networks each perform different task. As such further
differentiation of both networks might lead to increased performance. Thus, a future work might be change the second layer
architecture.
• the real-time experiment accuracy was less than the Evaluation Dataset accuracy (95.42% vs 98.31% respectively). This is
likely due to the reaction delay of the participants, but more importantly to the transition between gestures. These
transitions are not part of the training dataset, because they are too time consuming to record as the number of possible
transitions …... Consequently, it is expected that the classifiers predictive power on transition data is poor in these
circumstances. As such, being able to accurately detect such transitions in an unsupervised way might have a greater impact
on the system’s responsiveness than simply reducing the window size. This and the aforementioned point will be
investigated in future works.
37. IX. Discussion
• The main limitation of this study is the absence of tests with amputees. Additionally, the issue of electrode
shifts has not been explicitly studied and the variability introduced by various limb positions was not
considered when recording the dataset.
• A limitation of the proposed TL scheme is its difficulty to adapt when the new user cannot wear the same amount of
electrodes as the group used for pre-training. This is because changing the number of channels changes the representation
of the phenomena (i.e. muscle contraction) being fed to the algorithm. The most straightforward way of addressing this
would be to numerically remove the relevant channels from the dataset used for pre-training. Then re-running the proposed
TL algorithm on an architecture adapted to the new representation fed as input. Another solution is to consider the EMG
channels in a similar way as color channels in image. This type of architecture seems, however, to perform worse than the
ones presented in this paper
38. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . s E M G D a t a s e t s
I V . C l a s s i s E M G C l a s s i f i c a t i o n
V . D e e p L e a r n i n g C l a s s i f i e r s O v e r v i e w
V I . T r a n s f e r L e a r n i n g
V I I . C l a s s i f i e r C o m p a r i s o n
V I I I . R e a l - T i m e C l a s s i f i c a t i o n a n d
M e d i u m - T e r m P e r f o r m a n c e
I X . D i s c u s s i o n
X . C o n c l u s i o n
X I . A c k n o w l e d g e m e n t
39. VIII. Conclusion
• This paper presented three novel covNets that showed competitiveness with the current classifier.
• On the newly proposed evaluation dataset, the TL augmented ConvNet achieves an average accuracy of
98.31% over 17 participants.
• On the nina-pro, DB5(18 hand/wrist gestures), the proposed classifier achieved an average accuracy of
68.98% over 10 participants on a single Myo Armband.
• This dataset showed that the proposed TL algorithm learns sufficiently general features to significantly
enhance the performance of ConvNets on out-ofsample gestures.
• Future works will focus on adapting and testing the proposed TL algorithm on upper-extremity amputees. This will provide
additional challenges due to the greater muscle variability across amputees and the decrease in classification accuracy
compared to able-bodied participants [35].
43. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
44. Abstract
• Urban intelligence is an emerging concept which guides a series of infrastructure developments in modern
smart cities. Human-computer interaction (HCI) is the interface between residents and the smart cities, it
plays a key role in bridging the gap in applicating information technologies in modern cities.
• Hand gestures have been widely acknowledged as a promising HCI method.
• state-of-the-art signal processing technologies are not robust in feature extraction and pattern recognition
with sEMG signals
• In this paper, linear discriminant analysis (LDA) and extreme learning machine (ELM) are implemented in
hand gesture recognition system.
• The characteristic map slope (CMS) is extracted by using the feature re-extraction method because CMS can
strengthen the relationship of features cross time domain and enhance the feasibility of cross-time
identification.
• This study is focusing on optimizing the time differences in sEMG pattern recognition.
45. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
46. I. Introduction
• HCI is important could be used for IoT, IT, Cloud Computing…..
• Talks about EMG signals and that we can use it for gesture recognition.
• At present, the research on sEMG signal mainly focuses on the spatial non-stationarity of sEMG, but the time
non-stationarity of sEMG has always been a big challenge.
• In this paper, LDA is used to reduce the dimensionality of high-dimensional signals and eliminate the
redundant information in sEMG.
• The method of feature re-extraction is adopted to extract the characteristic slope value, and the ELM is
optimized by genetic algorithm to establish a gesture recognition system across time dimensions. The signal
is studied in time dimension.
• Paper Overview:
• Section II: reviews the basic methods of gesture recognition, including feature extraction, feature reduction and pattern
recognition based on neural network.
• Section III: elaborates the whole experimental process, including the sEMG signal col- lection plans. We proposed a
feature extraction method for reducing time non-stationarity, establishes GA-ELM optimization recognition network,
and analyzes the algorithm process of each step.
• Section IV: Results and conclusion
47. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
48. II. Related Work
• Read the generation of EMG. “interesting”.
• Pattern recognition is formed from three stages: Signal detection, signal representation and signal classification. The
accuracy is usually from 50-70%.
• There has been wide research on feature extraction, feature dimensionality reduction and pattern recognition as: time-
domain processing method “then he explain it RMS, Zero crossing, wavelength…”. The next paragraph he discuss the
“Frequency domain processing method”.
• These two methods yet ignores the chaotic and instability features of the EMG. In exercise the EMG is non-linear, thus Non-
liner dynamic method can construct multi-dimensional dynamic model based on one-dimensional time series to extract
more hidden information. The main non-liner characteristics include correlation dimension, entropy, complexity and
Lyapunov exponent.
• Mention some reasons, Therefore, when performing pattern recognition, it is necessary to adopt the method of feature
dimension reduction.
• Explains two ways for feature dimension reduction. feature selection and feature transformation.
• At present, the widely used pattern recognition methods of sEMG mainly include support vector machine (SVM), radial basis
function (RBF) algorithm, artificial neural net- work (ANN), hidden Markov model (HMM) and linear discriminant classifier.
ANN is the best. But here anyway he will use simpler extreme learning machine ( genetic algorithm).
49. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
50. II. Method
• A) ACQUISITION: the nine gestures that will be selected depends on moving
fingers and wrist.
• Didn’t talk about the data size, just mentioned it was within three days.
• B) DIMENSION REDUCTION AND FEATURE FUSION:
1) PRELIMINARY FEATURE EXTRACTION: window functions are used to segment these
continuous signals into appropriate sizes. Mainly three features were captured. Two
related to the time domain: RMS and waveform length. And one related to frequency
domain: median amplitude spectrum. These three parameters are the input for the
identification network.
ACQUISITION PREPROCESSING
CLASSIFIER
2) FEATURE FUSION AND DIMENSION REDUCTION: The direct extracted features, RMS, WL and MAS, are in high dimensional
feature spaces, which is not suitable for classification. In order to improve the accuracy of gesture recognition and generalize
the classier, reducing the dimension of the feature space is very important. (not sure but kind of normalizing). The output is
matrix X3
3) FEATURE RE-EXTRACTION:
Here we’ll find the features that will be used as input for the ANN. Just the time domain signal won’t be enough, and he
mentions why.
Then he mentions that they had to map the matrix X3 to a new eigen vector matrix X4 (feature matrix). The purpose of the
feature matrix it maps the values of each gesture to a new separate then the other gestures (uses an equation to map).
• C) Classifier design and parameter optimization: genetic algorithm in GA-ELM is designed to optimize the initial
weights and thresholds of the network.
• Extreme learning machine and genetic algorithm:
It’s a single hidden layer feedforward neural network.
51. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
52. III. RESULTS
• In order to illustrate the advantages of this method, the data before and after dimensionality reduction are used
as input of ELM classifier. And the results are compared.
• ANALYSIS OF TWO EXPERIMENTAL RESULTS BEFORE GA OPTIMIZATION
• The average recognition accuracy of ELM network after LDA dimension reduction is 67.18%, and that of ELM
network after feature extraction is 75.74%.
• EXPERIMENTAL RESULTS ANALYSIS OF GA OPTIMIZATION
• The results show that the accuracy is 79.3%
53. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . R e l a t e d W o r k
I I I . M e t h o d
I V . R e s u l t A n a l y s i s o f
G e s t u r e R e c o g n i t i o n
V . C o n c l u s i o n
54. IV. CONCLUSION
• In this paper, the second feature extraction of static gesture based on sEMG is studied. Theoretical and
experimental results have been obtained, however there are still problems need to be further explored. The
extraction of eigenvalue slope improves the recognition accuracy in our work, to define new features or feature
selection methods are promising research directions in the future.
58. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
59. Abstract
• This study presents a comprehensive and systematic investigation of the feasibility of hand gesture recognition
using EMG signals recorded at the wrist. A direct comparison of signal and information quality is conducted
between concurrently recorded wrist and forearm signals. Both signals were collected simultaneously from 21
subjects while they performed a selection of 17 different single-finger gestures, multi-finger gestures and wrist
gestures.
• Wrist EMG signals yielded consistently higher (p < 0:05) signal quality metrics than forearm signals for gestures
that involved fine finger movements, while maintaining comparable quality for wrist gestures.
• Classifiers trained and tested using wrist EMG signals achieved average accuracy levels of 92.1% for single-
finger gestures, 91.2% for multi-finger gestures and 94.7% for the conventional wrist gestures.
60. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
61. I. Introduction
• Applications for sEMG gesture recognition for IoT.
• Pattern recognition is widely used using forearm due to its use in prosthetics. The most commercial use is even
in prosthetics
• Early attempts for commercializing the PR was with Myo and gForce and both was for amputees. Here we say
there are more things we can do using PR than amputees. And the PR using forearm has been just limited in the
labs because customers aren’t used to wear bands around their forearm. That is why we won’t be focusing on
reading the data from forearm, instead from the wrist.
• Other technologies has also been used for PR like pressure sensing, ultrasound, data gloves……. But just kept
in labs and there isn’t any commercial products and mention some reasons why
• IMU was being effective when combined with EMG to detect hand positions and air gestures. But IMU failed to
detect salient hand gestures that don’t involve the movement of the wrist joint.
• Another reason for wrist instead of hand that it should be comfy to be worn for long time and socially accepted.
• Few studies used the wrist EMG. Some used it combined with forearm EMG. Some with the muscles inside the
hand (very high accuracy).
• Another study compared between the EMG at the wrist and at the forearm and didn’t find a difference in
strength or snr
62. I. Introduction
• Some studies focused only on the wrest but in a very limited scope. Consequently, for the first time, this study
encompasses a comprehensive and systematic investigation of the feasibility of using wrist EMG signals for
hand gesture recognition with a broad set of finger and hand gestures. Using many subjects.
• The first objective of the paper is to show the importance of wrist EMG PR.
• The second objective of this work is therefore to explore how the body of knowledge from the forearm EMG
literature, such as feature engineering and dimensionality reduction, can be applied to wrist EMG signals.
• Finally, the paper is concluded with a set of recommendations for future research studies and the
commercialization of wrist-based EMG wearables.
63. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
64. II. Methodology
• A) Data Collection: For this study, EMG data were collected from 21 ablebodied participants (25:247:56 years,
14 males, 7 females, 19 right-handed) . Each participant wore 8 electrodes, 4 on the wrist and 4 on the
forearem, sampling rate was set to 1kHz. And talk more about the electrodes position.
• Participents were asked to perform 18 gesture. These gestures included common contractions consistent with 5
single-finger extensions, 6 multi-finger gestures and 6 wrist gestures.
• EMG data were collected from each participant while performing 4 repetitions of each gesture for 2 seconds
with breaks of 2 seconds between gestures. Hence, a balanced dataset of 72 repetitions per participant was
collected.
• B)Evaluating the signal quality: here he compares the signals from forearm with signals from wrist. The
comparison had Four signal quality indices were evaluated to quantify the amount of useful information
contained in the signals and to evaluate signal noise.: 1) Forearm-to-Wrist Signal Strength Ratio (FWR). 2)
Signal-to-Noise Ratio (SNR). 3) Signal-to-Motion Artifact Ratio (SMR). 4) Power Spectrum Deformation ().
• Each of them is detailly discussed.
• C) Feature extraction and engineering: Feature engineering techniques were applied to assess the performance
of individual state-of-the-art EMG features and a feature set previously identified in the forearm literature.
• EMG signals were segmented to 150ms sliding windows with 50% overlap. A total of 23 time-domain and
frequency-domain features were extracted from each channel as listed in Table I.
65. II. Methodology
• This resulted in a feature vector of length 280 per gesture repetition, considering all 8 channels and given that
HIST and AR features each produce 10 and 4 values, respectively. This feature set has proven to provide good
hand gestures recognition rates using forearm EMG signals.
• Class separation and feature repeatability were evaluated in multi-dimensional feature space using two
statistical indices: Davies-Bouldin index (DBI) and repeatability index (RI).
• D) Dimensionality Reduction:
• 1- feature selection: The sequential floating feature selection (SFFS) algorithm was applied to identify the most
important features for gesture recognition from the wrist and forearm EMG signals.
• 2- Feature projection: To visualize differences in the underlying information content, principle component
analysis (PCA) was applied to all wrist and forearm EMG features listed in Table I.
• E)Classification Algorithm: LDA and SVM (with a linear kernel) classifiers were employed in this study using the
leave-one-repetition-out cross-validation technique to estimate classification accuracies.
• F) Statistical Analysis: t-test was performed
66. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
70. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
71. II. Discussion
• A) Wrist electrode configuration:
• some of the forearm muscles that control extension/flexion of the medial four fingers, such as the extensor
digiti minimi (EDM), extensor digitorum (ED) and flexor digitorum superficialis (FDS), extend down to the wrist.
Also, the extensor carpi ulnaris (ECU) muscle, which aids in hand extension and abduction, extends from the
proximal forearm down to the wrist [36]. Thus, the electrical activity of all these muscles can still be recorded at
the wrist level. Furthermore, there is a unique group of muscles that are only present at the wrist, near the
distal end of the forearm, proximal to the ulnar styloid process, and control fine finger and thumb movements.
• Hence, the electrodes placed more distally, near the wrist joint, are likely to have measured EMG activity
relevant to fine finger movements.
• B) Feasibility of Wrist-Based EMG Gesture Recognition:
• Results showed that wrist EMG signals had higher signal quality compared to the forearm signals for multiple
gestures involving fine finger movements. Indeed, results showed that gestures that involved the movement of
the thumb and index fingers had significantly higher wrist signal strength (FWR< 1) and SNR values when
compared to forearm signals.
• in the current study, the results found that wrist EMG had higher amplitudes and SNR values than forearm
signals for gestures that involved finger motions.
• READ THE REST….
72. II. Discussion
• Future studies should thus explore the effect of dynamic and confounding factors on the performance of wrist-
based EMG pattern recognition systems. These factors, as have been outlined in the forearm literature, may
include limb position, electrode shift, and the effect of contraction intensity on the pattern recognition
performance. Another point to consider in future studies is the difference between the continuous classification
approach in prosthetics versus the discrete gesture approach in HCI (e.g. snapping fingers). Other directions for
researchers to pursue are the performance of wrist-worn EMG systems over time and how to maintain
performance over many days without the need to retrain the system.
73. 1.Contents
A b s t r a c t
I . I n t r o d u c t i o n
I I . M e t h o d o l o g y
I I I . R e s u l t s
I V . D i s c u s s i o n
V . C o n c l u s i o n
74. III. Conclusion
• Wrist-based EMG classifiers achieved average accuracy levels of more than 91.2% for fine finger gestures and
94.7% for the conventional wrist gestures. This study highlights the great potential of wrist EMG signals and
paves the way for reliable wrist-based EMG wearables.
76. 4.
MASTERS THESIS
MYOELECTRIC HUMAN COMPUTER
INTERACTION USING LSTM-CNN
NEURAL NETWORK FOR DYNAMIC
HAND GESTURES RECOGNITION
I E E E X p lor e
2 0 2 1
U S A
78. ABSTRACT
• There are two goals for this research:
• 1- Decease the effect of arm positions when recognizing a gestures. To tackle this issue, a CNN-LSTM neural
network is introduced. Compared to Dr. Shin’s work, the new model is able to classify more gestures with more
positions.
• 2- Apply the new model to a human-computer interaction system. A 7-DoF Kinova robot arm.
79. 1.Contents
1 . I N T R O D U C T I O N A N D L I T E R A T U R E R E V I E W
1.1 Literature Review
1.1.1 ElectroMyoGraphic Signal
1.1.2 Myoelectric Control
‘1.2 Observation
1.3 Thesis Objective
2 . P R O P O S E D M E T H O D F O R G E S T U R E R E C O G N I T I O N
2.1 Introduction
2.2 Proposed Method.
2.2.1 EMG Gestures Dataset
2.2.2 CNN-LSTM network Structure
2.2.3 Training Results
2.3 Real Time Gestures Recognition
2.3.1 Real Time Recognition System
2.3.2 Data Collection
2.4 Results
2.4.1 Real-time Classification Accuracy.
2.4.2 Results Analysis
2.4.3 Discussion
3 . R E A L - T I M E H C I S Y S T E M F O R A 7 - D O F R O B O T A R M
C O N T R O L
4 . S U M M A R Y A N D C O N C L U S I O N S
80. 1.Contents
1 . I N T R O D U C T I O N A N D L I T E R A T U R E R E V I E W
1.1 Literature Review
1.1.1 ElectroMyoGraphic Signal
1.1.2 Myoelectric Control
‘1.2 Observation
1.3 Thesis Objective
2 . P R O P O S E D M E T H O D F O R G E S T U R E R E C O G N I T I O N
2.1 Introduction
2.2 Proposed Method.
2.2.1 EMG Gestures Dataset
2.2.2 CNN-LSTM network Structure
2.2.3 Training Results
2.3 Real Time Gestures Recognition
2.3.1 Real Time Recognition System
2.3.2 Data Collection
2.4 Results
2.4.1 Real-time Classification Accuracy.
2.4.2 Results Analysis
2.4.3 Discussion
3 . R E A L - T I M E H C I S Y S T E M F O R A 7 - D O F R O B O T A R M
C O N T R O L
4 . S U M M A R Y A N D C O N C L U S I O N S
81. Introduction
• 1. Literature review: Skip, Not important
• 2. ElectroMyoGraphic Signal: Pattern recognition for EMG classification usually consist of data pre-processing,
data segmentation, feature extraction, dimensionality reduction, and classification stages.
• He then explain breafly each of them.
• Keeps mentioning Dr shin’s work. And put a diagram od LSTM that Dr shin used during his work.
82. Introduction
• 1.1.2 Myoelectric Control: Not important
• 1.2 Observation: Gestures include dynamic and static motions recognition is a time-series problem. Therefore,
recurrent neural network(RNN) is proper for this.
• 1.3)Thesis Objective: This project aims to design a system that enables people to use gestures with an MYO
armband to manipulate a robot arm. This system has two main parts: 1) recognition part: using neural networks
to recognize dynamic gestures. 2) control part: implementing a relative functions on a robot arm based on
different gestures.
This system should also ensure:
• Position Independent: Ensure that gestures could be recognized while limb positions change.
• Respond Promptly: Ensure that the time of completing a task by using myoelectric control is not much longer
than other control methods.
• Two devices are used in this research. First is the Myo armband. Seocnd one is a 7 DoF Kinova robot arm
83. 1.Contents
1 . I N T R O D U C T I O N A N D L I T E R A T U R E R E V I E W
1.1 Literature Review
1.1.1 ElectroMyoGraphic Signal
1.1.2 Myoelectric Control
‘1.2 Observation
1.3 Thesis Objective
2 . P R O P O S E D M E T H O D F O R G E S T U R E R E C O G N I T I O N
2.1 Introduction
2.2 Proposed Method.
2.2.1 EMG Gestures Dataset
2.2.2 CNN-LSTM network Structure
2.2.3 Training Results
2.3 Real Time Gestures Recognition
2.3.1 Real Time Recognition System
2.3.2 Data Collection
2.4 Results
2.4.1 Real-time Classification Accuracy.
2.4.2 Results Analysis
2.4.3 Discussion
3 . R E A L - T I M E H C I S Y S T E M F O R A 7 - D O F R O B O T A R M
C O N T R O L
4 . S U M M A R Y A N D C O N C L U S I O N S
84. 2.1Introduction
• The four Challenges being faced: limb position factor, constrain intensity factor, electrode shift factor and
within/between day factor.
• Mentions a paper that found explained how the limb positions influence in detail: the muscular activity that
maintains limb positions against gravitational force is dependent on the position of the limb[16].
• Dr shin’s work shows that dynamic gestures are in principle more reliable indicators of intent.
85. 2.2 Proposed Method
• 2.2.1 EMG Gestures Dataset: There are five dynamic gestures with five arm positions in the dataset. (dynamic =
includes motion like waving).
• The training data is from seven human subjects. Every gesture at each position would be repeated for 10 times.
After finishing seven experiments, there are 7*5*5*10 = 1750 samples. For each sample, there is a 8*680
matrix where 8 is the the amount of channels and 680 is the time steps (The matrix would be pad zeros to
make every matrix same length). Tn original dataset is in below.
• Due to the small dataset. Augmentation as in case of images
Is added.
• After augmentation
86. 2.2 Proposed Method
• 2.2.2CNN-LSTM network Structure: Explains in detail the CNN and LSTM.
• 2.2.3 Training Result: The dataset is split into training dataset and validation dataset, which are 80% and 20%
respectively. Therefore, for the original dataset, there are 1750*0.8 = 1400 samples for training and 350
samples for validation. And for dataset after the augmentation, there are 10500*0.8 = 8400 samples for
training and 2100 smaples for validation.
• The validation accuracy of original dataset is approximate 90% which is lower than the model trained by new
dataset. The training accuracy is 95.8% and validation accuracy is 92.7%.
87. 2.3 Real Time Gestures Recognition
• 2.3.1 Real Time Recognition System: useless diagram shows the steps of reading a gesture.
• 2.3.2 Data Collection: there are 3000 sample for testing.
• 2.4.1 Real-time Classification Accuracy: overall 84.2%
• Then he add some confusion matrixes.
• 2.4.3 discussion:
• From the previous analysis, there are three observations:
1. The model is relatively position independent. The effect of arm positions still exists but is small.
2. 2. Different gestures impact recognition accuracy. In this model, influence on gestures may be bigger than
positions for classification accuracy. The gestures share less common would get better results.
3. The model is affected by subject dependency. The best and worst results show it clearly. Due to the different
performing habits for different human subjects, more data from variouspeople may help.
2.5 conclusion:
The training and validation accuracy are both over 90%.
Real-time accuracy = The overall accuracy is 84.2%.
88. 1.Contents
1 . I N T R O D U C T I O N A N D L I T E R A T U R E R E V I E W
1.1 Literature Review
1.1.1 ElectroMyoGraphic Signal
1.1.2 Myoelectric Control
‘1.2 Observation
1.3 Thesis Objective
2 . P R O P O S E D M E T H O D F O R G E S T U R E R E C O G N I T I O N
2.1 Introduction
2.2 Proposed Method.
2.2.1 EMG Gestures Dataset
2.2.2 CNN-LSTM network Structure
2.2.3 Training Results
2.3 Real Time Gestures Recognition
2.3.1 Real Time Recognition System
2.3.2 Data Collection
2.4 Results
2.4.1 Real-time Classification Accuracy.
2.4.2 Results Analysis
2.4.3 Discussion
3 . R E A L - T I M E H C I S Y S T E M F O R A 7 - D O F R O B O T A R M
C O N T R O L
4 . S U M M A R Y A N D C O N C L U S I O N S
89. 4. SUMMARY AND CONCLUSIONS
• Using CNN to generate the features instead of calculating time or frequency domain features manually before
input.
• Compared to many researches which focus on gestures recognition at the same limb position and have bad
performance when the limb position changes, the model in this research is relatively position-invariant.
• Compared to Dr. Shin’s work[3] which involves 3 gestures and 4 limb positions in the real time recognition, this
research includes 5 gestures and 5 limb positions. The model is more robust on gestures and limb positions for
real-time classification.
• Future Work:
1. Removing the trigger(fist) in the real-time system.
2. Speeding up processing time. This could be done by using a better GPU
3. Improving accuracy for the classification. Other advanced methods could be tried such as transfer learning.
Also more data from various types of people my help since the amount of gestures and arm positions is not
small and there are considerable differences among different human beings.
4. Increasing recognizable gestures
5. Reducing the arm position’s effect to gestures recognition. In other words, the gestures performing at the
random arm position could be classified successfully.
6. Decreasing the effect of human subject dependency.
7. Applying the proposed myoelectric control system on other applications like a drone control.
8. Adding feedback in the myoelectric control system.