SlideShare a Scribd company logo
1 of 12
Project Echo and
Speech Recognition Software
Matthew N. Gray
NIFS Student Intern – Summer 2016
Airspace Operations and Safety Program
Crew Systems and Aviation Operations Branch
Mentors: Dr. Angela Harrivel, Chad Stephens
1
Crew State Monitoring (CSM) – Motivation
• National airspace is becoming increasingly
busy and more complex.
• If we could stop just one fatal commercial
aviation accident, we could save hundreds
of lives and a billion dollars.
• Loss of Control – Inflight (LOC-I)
o Largest Category of aircraft-related Fatal
Events
• CSM aims to improve pilot operational
efficiency during safety-critical operations
by:
o Improving the human-machine interface in
aircraft
o Aiding in pilot attention training 3 Boeing Statistical Summary of Commercial Worldwide Jet
Transport Accidents, 2011. Includes only accidents involving
turbofan or turbojet airplanes with max takeoff weight > 60,000
lbs., referenced in the CAST Airplane State Awareness Joint
Safety Analysis Team Final Report, June 17, 2014
2
Background – Crew State Monitoring (CSM)
• Data Collection
o fNIRS, EEG, EKG, Resp., GSR, Eye-
tracking, etc.
• Data Synchronization
o MAPPS Software
• Signal Pre-Processing
o Filtering, Feature Extraction
• Machine Learning
o Classification Algorithm
o Model Evaluation
• Real-time State Indicator
Display
o High/Low Workload, Distraction,
Inattentional Blindness, Confirmation
Bias, etc.
Figure courtesy of Charles Liles
and the LaRC Big Data and Machine Information Team3
Background – AFDC 2.0 and SHARP 1.0
• Augmented Flight Deck
Countermeasures (AFDC
2.0)
o Seeks to improve human-
machine interaction among
safety critical operations of
airplanes
• Scenarios for Human
Attention Recovery using
Psychophysiology (SHARP
1.0)
o Supports CAST goals by
measuring crew cognitive
states in simulators during
safety critical operations SHARP Display
4
Muse
SmartEye
Spire
Empatica E4
BIOPAC fNIR100B
EEG
fNIRS
Heart Rate (PPG)
GSR
Temp.
Accel.
Respiratory Rate
Eye-tracking
Project Echo
Sensors
Data
Acquisition
Data
Synchronization
MAPPS or MAPPS Equivalent:
• Lab Streaming Layer (LSL)
• iMotions
• XTObserver
Signal
Processing
Classification
Algorithm
Hidden Markov
Models (HMMs)
Neural Netowrks (NNs)
Machine
Learning
Display
Indicate Cognitive State:
• High/Nominal Workload
• Channelized Attention
• Diverted Attention
• Confirmation Bias
• Inattentional Blindness
Project Echo – SmartEye® Eye-Tracking System
• Hardware Setup
• Camera Calibration
• Building World Model
• Head Profile Creation
• Gaze Calibration
• Real-time data transfer
(future)
SmartEye® Setup with corresponding
World Model
6
Project Echo – Empatica E4 Wristband
• Measures
o Heart Rate – HR (PPG)
o Galvanic Skin Response – GSR
o Skin Temperature
o Acceleration/Movement
• Lightweight, non-invasive,
portable
• Real-time data streaming
through Matlab
o TCP/IP Connection
o Stored in text file
7
Speech Recognition Software – Motivation
• Talking could affect
classification accuracy during
runs
oMore/varying cognitive
activation
oIrregular breathing
oMovement
• Automatic labeling of speech
vs. other noises for future
analysis
Talking
8
Speech Recognition – MFCCs
• Mel Frequency Cepstral
Coefficients (MFCCs)
oFeature derived from audio
signal
oRepresents shape of vocal tract
through short-time frequency
analysis
oApplies filter to frequency
response of signal to emulate
human hearing
 Humans can distinguish low
frequencies easier than high
frequencies
Cross-sectional shapes of vocal tracts
Cochlear frequency map showing logarithmic
frequency resolution of human hearing 9
Speech Recognition – MFCCs (cont.)
Steps:
• Calculate moving window PSD
o Assume ‘stationarity’ at 25ms window
sizes
o Apply Hamming filter to windowed signal
• Create Mel filter bank (shown to the
right)
• Multiply each windowed PSD by each
filter in filter bank
• Sum powers in each binned and filtered
frequency for each frame
• Take log and discrete cosine transform
of summed powers
• Produces array of 12 coefficients for
each windowed time series
×
=
10
Future Work
• Acquire data real-time from SmartEye®
system
• Parse out data stream from Empatica in
Matlab into separate text files
• Synchronize data streams from all
devices with MAPPS or MAPPS
equivalent (LSL, etc.)
• Incorporate in-house preprocessing and
machine learning scripts
• Input MFCCs from audio signals into
Hidden Markov Model machine learning
classifier
11
Acknowledgements
• Angela Harrivel, PhD
• Chad Stephens, PhD Candidate
• Kyle Ellis, PhD
• Ray Comstock, PhD
• Kellie Kennedy, PhD Candidate
• Nick Napoli
• Katrina Colucci-Chang
• Will Hollingsworth
• Alex Liang
12

More Related Content

Similar to Matthew Gray Exit Presentation Summer 2016 Full

Iisrt saksham sood (cs)
Iisrt saksham sood (cs)Iisrt saksham sood (cs)
Iisrt saksham sood (cs)
IISRT
 
Book of abstract volume 8 no 9 ijcsis december 2010
Book of abstract volume 8 no 9 ijcsis december 2010Book of abstract volume 8 no 9 ijcsis december 2010
Book of abstract volume 8 no 9 ijcsis december 2010
Oladokun Sulaiman
 
A Document on Artificial Passenger
A Document on Artificial PassengerA Document on Artificial Passenger
A Document on Artificial Passenger
Mohammed Iqbal
 

Similar to Matthew Gray Exit Presentation Summer 2016 Full (20)

2016 davis-plantbio
2016 davis-plantbio2016 davis-plantbio
2016 davis-plantbio
 
Iit kgp workshop
Iit kgp workshopIit kgp workshop
Iit kgp workshop
 
Deep Learning using OpenPOWER
Deep Learning using OpenPOWERDeep Learning using OpenPOWER
Deep Learning using OpenPOWER
 
Iisrt saksham sood (cs)
Iisrt saksham sood (cs)Iisrt saksham sood (cs)
Iisrt saksham sood (cs)
 
Book of abstract volume 8 no 9 ijcsis december 2010
Book of abstract volume 8 no 9 ijcsis december 2010Book of abstract volume 8 no 9 ijcsis december 2010
Book of abstract volume 8 no 9 ijcsis december 2010
 
Accelerating GWAS epistatic interaction analysis methods
Accelerating GWAS epistatic interaction analysis methodsAccelerating GWAS epistatic interaction analysis methods
Accelerating GWAS epistatic interaction analysis methods
 
Speech Recognition By Hardik Mistry(Laxmi Institute Of Technology)
Speech Recognition By Hardik Mistry(Laxmi Institute Of Technology)Speech Recognition By Hardik Mistry(Laxmi Institute Of Technology)
Speech Recognition By Hardik Mistry(Laxmi Institute Of Technology)
 
sleep apnea detection
sleep apnea detectionsleep apnea detection
sleep apnea detection
 
Mobility collector: Battery Conscious Mobile Tracking
Mobility collector: Battery Conscious Mobile TrackingMobility collector: Battery Conscious Mobile Tracking
Mobility collector: Battery Conscious Mobile Tracking
 
State-Of-The Art Machine Learning Algorithms and How They Are Affected By Nea...
State-Of-The Art Machine Learning Algorithms and How They Are Affected By Nea...State-Of-The Art Machine Learning Algorithms and How They Are Affected By Nea...
State-Of-The Art Machine Learning Algorithms and How They Are Affected By Nea...
 
Speech Recognition Technology
Speech Recognition TechnologySpeech Recognition Technology
Speech Recognition Technology
 
A Document on Artificial Passenger
A Document on Artificial PassengerA Document on Artificial Passenger
A Document on Artificial Passenger
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
 
Utterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANNUtterance Based Speaker Identification Using ANN
Utterance Based Speaker Identification Using ANN
 
Human factors design review - ROV piloting - presentation
Human factors design review - ROV piloting - presentationHuman factors design review - ROV piloting - presentation
Human factors design review - ROV piloting - presentation
 
Autonomous Learning for Autonomous Systems, by Prof. Plamen Angelov
Autonomous Learning for Autonomous Systems, by Prof. Plamen AngelovAutonomous Learning for Autonomous Systems, by Prof. Plamen Angelov
Autonomous Learning for Autonomous Systems, by Prof. Plamen Angelov
 
SPEECH RECOGNITION USING NEURAL NETWORK
SPEECH RECOGNITION USING NEURAL NETWORK SPEECH RECOGNITION USING NEURAL NETWORK
SPEECH RECOGNITION USING NEURAL NETWORK
 
Psychophysiological aspects of Human-System Integration i C4 and operation sa...
Psychophysiological aspects of Human-System Integration i C4 and operation sa...Psychophysiological aspects of Human-System Integration i C4 and operation sa...
Psychophysiological aspects of Human-System Integration i C4 and operation sa...
 
Human Evaluation: Why do we need it? - Dr. Sheila Castilho
Human Evaluation: Why do we need it? - Dr. Sheila CastilhoHuman Evaluation: Why do we need it? - Dr. Sheila Castilho
Human Evaluation: Why do we need it? - Dr. Sheila Castilho
 
50120140502007
5012014050200750120140502007
50120140502007
 

Matthew Gray Exit Presentation Summer 2016 Full

  • 1. Project Echo and Speech Recognition Software Matthew N. Gray NIFS Student Intern – Summer 2016 Airspace Operations and Safety Program Crew Systems and Aviation Operations Branch Mentors: Dr. Angela Harrivel, Chad Stephens 1
  • 2. Crew State Monitoring (CSM) – Motivation • National airspace is becoming increasingly busy and more complex. • If we could stop just one fatal commercial aviation accident, we could save hundreds of lives and a billion dollars. • Loss of Control – Inflight (LOC-I) o Largest Category of aircraft-related Fatal Events • CSM aims to improve pilot operational efficiency during safety-critical operations by: o Improving the human-machine interface in aircraft o Aiding in pilot attention training 3 Boeing Statistical Summary of Commercial Worldwide Jet Transport Accidents, 2011. Includes only accidents involving turbofan or turbojet airplanes with max takeoff weight > 60,000 lbs., referenced in the CAST Airplane State Awareness Joint Safety Analysis Team Final Report, June 17, 2014 2
  • 3. Background – Crew State Monitoring (CSM) • Data Collection o fNIRS, EEG, EKG, Resp., GSR, Eye- tracking, etc. • Data Synchronization o MAPPS Software • Signal Pre-Processing o Filtering, Feature Extraction • Machine Learning o Classification Algorithm o Model Evaluation • Real-time State Indicator Display o High/Low Workload, Distraction, Inattentional Blindness, Confirmation Bias, etc. Figure courtesy of Charles Liles and the LaRC Big Data and Machine Information Team3
  • 4. Background – AFDC 2.0 and SHARP 1.0 • Augmented Flight Deck Countermeasures (AFDC 2.0) o Seeks to improve human- machine interaction among safety critical operations of airplanes • Scenarios for Human Attention Recovery using Psychophysiology (SHARP 1.0) o Supports CAST goals by measuring crew cognitive states in simulators during safety critical operations SHARP Display 4
  • 5. Muse SmartEye Spire Empatica E4 BIOPAC fNIR100B EEG fNIRS Heart Rate (PPG) GSR Temp. Accel. Respiratory Rate Eye-tracking Project Echo Sensors Data Acquisition Data Synchronization MAPPS or MAPPS Equivalent: • Lab Streaming Layer (LSL) • iMotions • XTObserver Signal Processing Classification Algorithm Hidden Markov Models (HMMs) Neural Netowrks (NNs) Machine Learning Display Indicate Cognitive State: • High/Nominal Workload • Channelized Attention • Diverted Attention • Confirmation Bias • Inattentional Blindness
  • 6. Project Echo – SmartEye® Eye-Tracking System • Hardware Setup • Camera Calibration • Building World Model • Head Profile Creation • Gaze Calibration • Real-time data transfer (future) SmartEye® Setup with corresponding World Model 6
  • 7. Project Echo – Empatica E4 Wristband • Measures o Heart Rate – HR (PPG) o Galvanic Skin Response – GSR o Skin Temperature o Acceleration/Movement • Lightweight, non-invasive, portable • Real-time data streaming through Matlab o TCP/IP Connection o Stored in text file 7
  • 8. Speech Recognition Software – Motivation • Talking could affect classification accuracy during runs oMore/varying cognitive activation oIrregular breathing oMovement • Automatic labeling of speech vs. other noises for future analysis Talking 8
  • 9. Speech Recognition – MFCCs • Mel Frequency Cepstral Coefficients (MFCCs) oFeature derived from audio signal oRepresents shape of vocal tract through short-time frequency analysis oApplies filter to frequency response of signal to emulate human hearing  Humans can distinguish low frequencies easier than high frequencies Cross-sectional shapes of vocal tracts Cochlear frequency map showing logarithmic frequency resolution of human hearing 9
  • 10. Speech Recognition – MFCCs (cont.) Steps: • Calculate moving window PSD o Assume ‘stationarity’ at 25ms window sizes o Apply Hamming filter to windowed signal • Create Mel filter bank (shown to the right) • Multiply each windowed PSD by each filter in filter bank • Sum powers in each binned and filtered frequency for each frame • Take log and discrete cosine transform of summed powers • Produces array of 12 coefficients for each windowed time series × = 10
  • 11. Future Work • Acquire data real-time from SmartEye® system • Parse out data stream from Empatica in Matlab into separate text files • Synchronize data streams from all devices with MAPPS or MAPPS equivalent (LSL, etc.) • Incorporate in-house preprocessing and machine learning scripts • Input MFCCs from audio signals into Hidden Markov Model machine learning classifier 11
  • 12. Acknowledgements • Angela Harrivel, PhD • Chad Stephens, PhD Candidate • Kyle Ellis, PhD • Ray Comstock, PhD • Kellie Kennedy, PhD Candidate • Nick Napoli • Katrina Colucci-Chang • Will Hollingsworth • Alex Liang 12

Editor's Notes

  1. LOC-I: Loss of Control - Inflight