This document discusses a new system to convert visual-speech to text for deaf individuals using telephone communication. The system aims to automatically recognize Cued Speech gestures and convert them to text. Researchers extracted lip shapes and hand coordinates from video recordings of individuals performing Cued Speech. They used Hidden Markov Models and feature fusion to integrate lip shapes and hand gestures for isolated word and continuous phoneme recognition. The system achieved 86-94% accuracy for isolated words and 82-89% for continuous phonemes, indicating it can effectively convert visual-speech to text.
Speech to text conversion for visually impaired person using µ law compandingiosrjce
The paper represents the overall design and implementation of DSP based speech recognition and
text conversion system. Speech is usually taken as a preferred mode of operation for human being, This paper
represent voice oriented command for converting into text. We intended to compute the entire speech processing
in real time. This involves simultaneously accepting the input from the user and using software filters to analyse
the data. The comparison was then to be established by using correlation and µ law companding techniques. In
this paper, voice recognition is carried out using MATLAB. The voice command is a person independent. The
voice command is stored in the data base with the help of the function keys. The real time input speech received
is then processed in the speech recognition system where the required feature of the speech words are extracted,
filtered out and matched with the existing sample stored in the database. Then the required MATLAB processes
are done to convert the received data and into text form.
Complete power point presentation on SPEECH RECOGNITION TECHNOLOGY.
Very helpful for final year students for their seminar.
One can use this presentation as their final year seminar.
Speech Recognition is a very interesting topic for seminar.
Speech to text conversion for visually impaired person using µ law compandingiosrjce
The paper represents the overall design and implementation of DSP based speech recognition and
text conversion system. Speech is usually taken as a preferred mode of operation for human being, This paper
represent voice oriented command for converting into text. We intended to compute the entire speech processing
in real time. This involves simultaneously accepting the input from the user and using software filters to analyse
the data. The comparison was then to be established by using correlation and µ law companding techniques. In
this paper, voice recognition is carried out using MATLAB. The voice command is a person independent. The
voice command is stored in the data base with the help of the function keys. The real time input speech received
is then processed in the speech recognition system where the required feature of the speech words are extracted,
filtered out and matched with the existing sample stored in the database. Then the required MATLAB processes
are done to convert the received data and into text form.
Complete power point presentation on SPEECH RECOGNITION TECHNOLOGY.
Very helpful for final year students for their seminar.
One can use this presentation as their final year seminar.
Speech Recognition is a very interesting topic for seminar.
Also known as automatic speech recognition or computer speech recognition which means understanding voice by the computer and performing any required task.
This presentation was delivered to a "Web Enabled Business" class at Simon Fraser University in Vancouver. The topic is speech recognition technology, and the presentation covers its origins, how it works, issues, latest trends and future opportunities.
speech processing and recognition basic in data miningJimit Rupani
Basic presentation about speech processing
Name of the paper i read is :"An educational platform to demonstrate speech processing techniques on Android based smart phones and tablets" on Elsevier
This is a ppt on speech recognition system or automated speech recognition system. I hope that it would be helpful for all the people searching for a presentation on this technology
YouTube Link: https://youtu.be/sHeJgKBaiAI
** Python Certification Training: https://www.edureka.co/python **
This Edureka video on 'Speech Recognition in Python' will cover the concepts of speech recognition module in python with a program using speech recognition to translate speech into text. Following are the topics discussed:
How Speech Recognition Works?
How To Install SpeechRecognition In Python?
Working With Microphones
How To Install Pyaudio In Python?
Use case
speech recognition,History of speech recognition,what is speech recognition,Voice recognition software , Advantages and Disadvantages speech recognition, voice recognition,Voice recognition in operating systems ,Types of speech recognition
Description:
This project is an online Chat Room System which is built in Java Swing. The system adopts 3 tier architecture in which the client request server to register itself to the room, server register’s the user to the room and finally the client can share content with other registered client via server.
The system allows the user to connect from different location in a common chat-room where they can broadcast text messages and share drawing with other registered users.
Features:
• Apart from sharing content, the server broadcast entry and exist of the user.
• The chat window of every user is registered with client specific Title.
• Messages from client are pre-pend with their name.
• Cosmetic color change is implemented on-hover, on-active on the chat window, its borders and buttons.
• System allows tracking the user who is currently present in the chat room.
• System has the feature to save text messages/history in a text file from the chat window to the client local system.
Also known as automatic speech recognition or computer speech recognition which means understanding voice by the computer and performing any required task.
This presentation was delivered to a "Web Enabled Business" class at Simon Fraser University in Vancouver. The topic is speech recognition technology, and the presentation covers its origins, how it works, issues, latest trends and future opportunities.
speech processing and recognition basic in data miningJimit Rupani
Basic presentation about speech processing
Name of the paper i read is :"An educational platform to demonstrate speech processing techniques on Android based smart phones and tablets" on Elsevier
This is a ppt on speech recognition system or automated speech recognition system. I hope that it would be helpful for all the people searching for a presentation on this technology
YouTube Link: https://youtu.be/sHeJgKBaiAI
** Python Certification Training: https://www.edureka.co/python **
This Edureka video on 'Speech Recognition in Python' will cover the concepts of speech recognition module in python with a program using speech recognition to translate speech into text. Following are the topics discussed:
How Speech Recognition Works?
How To Install SpeechRecognition In Python?
Working With Microphones
How To Install Pyaudio In Python?
Use case
speech recognition,History of speech recognition,what is speech recognition,Voice recognition software , Advantages and Disadvantages speech recognition, voice recognition,Voice recognition in operating systems ,Types of speech recognition
Description:
This project is an online Chat Room System which is built in Java Swing. The system adopts 3 tier architecture in which the client request server to register itself to the room, server register’s the user to the room and finally the client can share content with other registered client via server.
The system allows the user to connect from different location in a common chat-room where they can broadcast text messages and share drawing with other registered users.
Features:
• Apart from sharing content, the server broadcast entry and exist of the user.
• The chat window of every user is registered with client specific Title.
• Messages from client are pre-pend with their name.
• Cosmetic color change is implemented on-hover, on-active on the chat window, its borders and buttons.
• System allows tracking the user who is currently present in the chat room.
• System has the feature to save text messages/history in a text file from the chat window to the client local system.
Hoja de Ruta (Propuesta de 7 puntos)
Sobre Ley Electoral – Pedimos una nueva ley electoral (Restauración de la Soberanía Nacional)
Foro por los Derecho y Libertades
The Future of News, Publishing, and Media (INMA 2010 Presentation)Gerd Leonhard
Gerd Leonhard's presentation at the 2010 INMA World Congress in New York, April 27, 2010 (International News Marketing Association) http://www.inma.org/modules/event/2010WorldCongress/
Better email response time using Microsoft Exchange 2013 with the Dell PowerE...Principled Technologies
In a market where servers can seem the same at a glance, look for the differences. Your email infrastructure choices will directly affect end-user experience for your UC&C applications. Equipped with more drives in its extra drive slots, the Dell PowerEdge R730xd delivered 31.7 percent better Exchange 2013 response times than a similarly configured, current-generation Supermicro server did. With better Microsoft Exchange Server 2013 response times, the PowerEdge R730xd can help deliver an improved experience for users in your organization.
Effect of MFCC Based Features for Speech Signal Alignmentskevig
The fundamental techniques used for man-machine communication include Speech synthesis, speech
recognition, and speech transformation. Feature extraction techniques provide a compressed
representation of the speech signals. The HNM analyses and synthesis provides high quality speech with
less number of parameters. Dynamic time warping is well known technique used for aligning two given
multidimensional sequences. It locates an optimal match between the given sequences. The improvement in
the alignment is estimated from the corresponding distances. The objective of this research is to investigate
the effect of dynamic time warping on phrases, words, and phonemes based alignments. The speech signals
in the form of twenty five phrases were recorded. The recorded material was segmented manually and
aligned at sentence, word, and phoneme level. The Mahalanobis distance (MD) was computed between the
aligned frames. The investigation has shown better alignment in case of HNM parametric domain. It has
been seen that effective speech alignment can be carried out even at phrase level
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTSijnlc
The fundamental techniques used for man-machine communication include Speech synthesis, speech
recognition, and speech transformation. Feature extraction techniques provide a compressed
representation of the speech signals. The HNM analyses and synthesis provides high quality speech with
less number of parameters. Dynamic time warping is well known technique used for aligning two given
multidimensional sequences. It locates an optimal match between the given sequences. The improvement in
the alignment is estimated from the corresponding distances. The objective of this research is to investigate
the effect of dynamic time warping on phrases, words, and phonemes based alignments. The speech signals
in the form of twenty five phrases were recorded. The recorded material was segmented manually and
aligned at sentence, word, and phoneme level. The Mahalanobis distance (MD) was computed between the
aligned frames. The investigation has shown better alignment in case of HNM parametric domain. It has
been seen that effective speech alignment can be carried out even at phrase level.
Speech Recognition Application for the Speech Impaired using the Android-base...TELKOMNIKA JOURNAL
Those who are speech impaired (tunawicara in the Indonesian language) suffer from
abnormalities in their delivery (articulation) of the language as well their voice in normal speech, resulting
in difficulty in communicating verbally within their environment. Therefore, an application is required that
can help and facilitate conversations for communication. In this research, the authors have developed a
speech recognition application that can recognise speech of the speech impaired, and can translate into
text form with input in the form of sound detected on a smartphone. By using the Google Cloud Speech
Application Programming Interface (API), this allows converting audio to text, and it is also user-friendly to
use such APIs. The Google Cloud Speech API integrates with Google Cloud Storage for data storage.
Although research into speech recognition to text has been widely practiced, this research try to develop
speech recognition, specially for speech impaired's speech, as well as perform a likelihood calculation to
see the factor of tone, pronunciation, and speech speed in speech recognition. The test was conducted by
mentioning the digits 1 through 10. The experimental results showed that the recognition rate for the
speech impaired is about 80%, while the recognition rate for normal speech is 100%.
Speech recognition converts spoken words to text. The term "speech recognition" is used to refer to recognition systems that must be trained to any speaker—as is the case for most desktop recognition software.
For Deaf people, ASL is a visual language (i.e.,) mobile video phones have the potential to give Deaf people access to real-time mobile communication in their preferred language.
Effect of Dynamic Time Warping on Alignment of Phrases and Phonemeskevig
Speech synthesis and recognition are the basic techniques used for man-machine communication. This type
of communication is valuable when our hands and eyes are busy in some other task such as driving a
vehicle, performing surgery, or firing weapons at the enemy. Dynamic time warping (DTW) is mostly used
for aligning two given multidimensional sequences. It finds an optimal match between the given sequences.
The distance between the aligned sequences should be relatively lesser as compared to unaligned
sequences. The improvement in the alignment may be estimated from the corresponding distances. This
technique has applications in speech recognition, speech synthesis, and speaker transformation. The
objective of this research is to investigate the amount of improvement in the alignment corresponding to the
sentence based and phoneme based manually aligned phrases. The speech signals in the form of twenty five
phrases were recorded from each of six speakers (3 males and 3 females). The recorded material was
segmented manually and aligned at sentence and phoneme level. The aligned sentences of different speaker
pairs were analyzed using HNM and the HNM parameters were further aligned at frame level using DTW.
Mahalanobis distances were computed for each pair of sentences. The investigations have shown more than
20 % reduction in the average Mahalanobis distances.
EFFECT OF DYNAMIC TIME WARPING ON ALIGNMENT OF PHRASES AND PHONEMESkevig
Speech synthesis and recognition are the basic techniques used for man-machine communication. This type
of communication is valuable when our hands and eyes are busy in some other task such as driving a
vehicle, performing surgery, or firing weapons at the enemy. Dynamic time warping (DTW) is mostly used
for aligning two given multidimensional sequences. It finds an optimal match between the given sequences.
The distance between the aligned sequences should be relatively lesser as compared to unaligned
sequences. The improvement in the alignment may be estimated from the corresponding distances. This
technique has applications in speech recognition, speech synthesis, and speaker transformation. The
objective of this research is to investigate the amount of improvement in the alignment corresponding to the
sentence based and phoneme based manually aligned phrases. The speech signals in the form of twenty five
phrases were recorded from each of six speakers (3 males and 3 females). The recorded material was
segmented manually and aligned at sentence and phoneme level. The aligned sentences of different speaker
pairs were analyzed using HNM and the HNM parameters were further aligned at frame level using DTW.
Mahalanobis distances were computed for each pair of sentences. The investigations have shown more than
20 % reduction in the average Mahalanobis distances.
Incremental Difference as Feature for LipreadingIDES Editor
This paper represents a method of computing
incremental dif ference f eatures on the basis of scan line
projection and scan converting lines for the lipreading problem
on a set of isolated word utterances. These features are affine
invariants and found to be eff ective in identification of
similarity between utterances by the speaker in spatial domain
MULTILINGUAL SPEECH TO TEXT USING DEEP LEARNING BASED ON MFCC FEATURESmlaij
The proposed methodology presented in the paper deals with solving the problem of multilingual speech
recognition. Current text and speech recognition and translation methods have a very low accuracy in
translating sentences which contain a mixture of two or more different languages. The paper proposes a
novel approach to tackling this problem and highlights some of the drawbacks of current recognition and
translation methods.
Democratizing Fuzzing at Scale by Abhishek Aryaabh.arya
Presented at NUS: Fuzzing and Software Security Summer School 2024
This keynote talks about the democratization of fuzzing at scale, highlighting the collaboration between open source communities, academia, and industry to advance the field of fuzzing. It delves into the history of fuzzing, the development of scalable fuzzing platforms, and the empowerment of community-driven research. The talk will further discuss recent advancements leveraging AI/ML and offer insights into the future evolution of the fuzzing landscape.
Quality defects in TMT Bars, Possible causes and Potential Solutions.PrashantGoswami42
Maintaining high-quality standards in the production of TMT bars is crucial for ensuring structural integrity in construction. Addressing common defects through careful monitoring, standardized processes, and advanced technology can significantly improve the quality of TMT bars. Continuous training and adherence to quality control measures will also play a pivotal role in minimizing these defects.
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Water scarcity is the lack of fresh water resources to meet the standard water demand. There are two type of water scarcity. One is physical. The other is economic water scarcity.
Courier management system project report.pdfKamal Acharya
It is now-a-days very important for the people to send or receive articles like imported furniture, electronic items, gifts, business goods and the like. People depend vastly on different transport systems which mostly use the manual way of receiving and delivering the articles. There is no way to track the articles till they are received and there is no way to let the customer know what happened in transit, once he booked some articles. In such a situation, we need a system which completely computerizes the cargo activities including time to time tracking of the articles sent. This need is fulfilled by Courier Management System software which is online software for the cargo management people that enables them to receive the goods from a source and send them to a required destination and track their status from time to time.
About
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Technical Specifications
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
Key Features
Indigenized remote control interface card suitable for MAFI system CCR equipment. Compatible for IDM8000 CCR. Backplane mounted serial and TCP/Ethernet communication module for CCR remote access. IDM 8000 CCR remote control on serial and TCP protocol.
• Remote control: Parallel or serial interface
• Compatible with MAFI CCR system
• Copatiable with IDM8000 CCR
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
Application
• Remote control: Parallel or serial interface.
• Compatible with MAFI CCR system.
• Compatible with IDM8000 CCR.
• Compatible with Backplane mount serial communication.
• Compatible with commercial and Defence aviation CCR system.
• Remote control system for accessing CCR and allied system over serial or TCP.
• Indigenized local Support/presence in India.
• Easy in configuration using DIP switches.
Forklift Classes Overview by Intella PartsIntella Parts
Discover the different forklift classes and their specific applications. Learn how to choose the right forklift for your needs to ensure safety, efficiency, and compliance in your operations.
For more technical information, visit our website https://intellaparts.com
Vaccine management system project report documentation..pdfKamal Acharya
The Division of Vaccine and Immunization is facing increasing difficulty monitoring vaccines and other commodities distribution once they have been distributed from the national stores. With the introduction of new vaccines, more challenges have been anticipated with this additions posing serious threat to the already over strained vaccine supply chain system in Kenya.
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfKamal Acharya
The College Bus Management system is completely developed by Visual Basic .NET Version. The application is connect with most secured database language MS SQL Server. The application is develop by using best combination of front-end and back-end languages. The application is totally design like flat user interface. This flat user interface is more attractive user interface in 2017. The application is gives more important to the system functionality. The application is to manage the student’s details, driver’s details, bus details, bus route details, bus fees details and more. The application has only one unit for admin. The admin can manage the entire application. The admin can login into the application by using username and password of the admin. The application is develop for big and small colleges. It is more user friendly for non-computer person. Even they can easily learn how to manage the application within hours. The application is more secure by the admin. The system will give an effective output for the VB.Net and SQL Server given as input to the system. The compiled java program given as input to the system, after scanning the program will generate different reports. The application generates the report for users. The admin can view and download the report of the data. The application deliver the excel format reports. Because, excel formatted reports is very easy to understand the income and expense of the college bus. This application is mainly develop for windows operating system users. In 2017, 73% of people enterprises are using windows operating system. So the application will easily install for all the windows operating system users. The application-developed size is very low. The application consumes very low space in disk. Therefore, the user can allocate very minimum local disk space for this application.
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdffxintegritypublishin
Advancements in technology unveil a myriad of electrical and electronic breakthroughs geared towards efficiently harnessing limited resources to meet human energy demands. The optimization of hybrid solar PV panels and pumped hydro energy supply systems plays a pivotal role in utilizing natural resources effectively. This initiative not only benefits humanity but also fosters environmental sustainability. The study investigated the design optimization of these hybrid systems, focusing on understanding solar radiation patterns, identifying geographical influences on solar radiation, formulating a mathematical model for system optimization, and determining the optimal configuration of PV panels and pumped hydro storage. Through a comparative analysis approach and eight weeks of data collection, the study addressed key research questions related to solar radiation patterns and optimal system design. The findings highlighted regions with heightened solar radiation levels, showcasing substantial potential for power generation and emphasizing the system's efficiency. Optimizing system design significantly boosted power generation, promoted renewable energy utilization, and enhanced energy storage capacity. The study underscored the benefits of optimizing hybrid solar PV panels and pumped hydro energy supply systems for sustainable energy usage. Optimizing the design of solar PV panels and pumped hydro energy supply systems as examined across diverse climatic conditions in a developing country, not only enhances power generation but also improves the integration of renewable energy sources and boosts energy storage capacities, particularly beneficial for less economically prosperous regions. Additionally, the study provides valuable insights for advancing energy research in economically viable areas. Recommendations included conducting site-specific assessments, utilizing advanced modeling tools, implementing regular maintenance protocols, and enhancing communication among system components.
Event Management System Vb Net Project Report.pdfKamal Acharya
In present era, the scopes of information technology growing with a very fast .We do not see any are untouched from this industry. The scope of information technology has become wider includes: Business and industry. Household Business, Communication, Education, Entertainment, Science, Medicine, Engineering, Distance Learning, Weather Forecasting. Carrier Searching and so on.
My project named “Event Management System” is software that store and maintained all events coordinated in college. It also helpful to print related reports. My project will help to record the events coordinated by faculties with their Name, Event subject, date & details in an efficient & effective ways.
In my system we have to make a system by which a user can record all events coordinated by a particular faculty. In our proposed system some more featured are added which differs it from the existing system such as security.
Visual speech to text conversion applicable to telephone communication
1. Visual-speech to text
conversion applicable
to telephone
communication for deaf
individuals
30TH APRIL 2013
2. Visual-speech to text conversion applicable to telephone communication for deaf individuals
INTRODUCTION
Lip-reading technique,
speech can be understood by interpreting
movements of lips, face and tongue.
not one-to-one
Impossible to distinguish phonemes using
visual information alone
3. Visual-speech to text conversion applicable to telephone communication for deaf individuals
the Cued Speech system
developed by Cornett
contains two components:
the hand shape the hand position relative to the
face.
Hand shapes- consonant phonemes
hand positions -vowel phonemes.
improves speech perception to a large extent
4. Visual-speech to text conversion applicable to telephone communication for deaf individuals
the Cued Speech system
5. Visual-speech to text conversion applicable to telephone communication for deaf individuals
AIM OF NEW SYSTEM
To investigate the designing of a system able to
automatically recognize Cued Speech and convert it
to text.
Possible for deaf or speech-impaired individuals to
communicate with each other and also with normal-hearing
persons
Using gestures
captured by devices equipped by a camera
6. Visual-speech to text conversion applicable to telephone communication for deaf individuals
METHODS
Corpus, feature extraction, and
statistical modeling
The speakers’ lips were painted blue, and color
marks were placed on the speakers’ fingers. .
The data were derived from a video recording of
the cuers pronouncing and coding in Cued
Speech
landmarks with different colors were placed on
the fingers
7. Visual-speech to text conversion applicable to telephone communication for deaf individuals
faster and more accurate image processing
stage.
The audio part of the video recording was
synchronized with the image.
An automatic image processing method was
appliedli pt ow idththe ( Av)i,d eo
lip aperture (B),
lip area (S).
pinching of the upper lip (Bsup)
lower (Binf) lip
8. Visual-speech to text conversion applicable to telephone communication for deaf individuals
Concatenative feature fusion
Tracks and extracts the xy coordinates
each time frame,
uses those values as features in the
HMM modeling.
uses the concatenation of the
synchronous lip shape and hand features
as the joint feature vector given by,
9. Visual-speech to text conversion applicable to telephone communication for deaf individuals
Joint lip hand
feature vector,
Lip shape
feature vector,
Hand feature
vector,
Dimensionality of the
joint feature vector
Parameters used for lip
shape modeling.
10. Visual-speech to text conversion applicable to telephone communication for deaf individuals
RESULTS
Isolated word recognition
1. Recognition in normal-hearing subject
11. Visual-speech to text conversion applicable to telephone communication for deaf individuals
2. Recognition in deaf subject
12. Visual-speech to text conversion applicable to telephone communication for deaf individuals
3. Multi-speaker isolated word recognition:
investigate whether it is possible to train speaker-independent
HMMs for Cued Speech recognition.
The training data consisted of 750 words from the
normal-hearing subject, and 750 words from the
deaf subject.
For testing 700 words from normal-hearing subject
and 700 words from the deaf subject were used,
respectively.
Each state was modeled with a mixture of 4
Gaussian distributions.
For lip shape and hand shape integration,
concatenative feature fusion was used.
13. Visual-speech to text conversion applicable to telephone communication for deaf individuals
14. Visual-speech to text conversion applicable to telephone communication for deaf individuals
4. Continuous phoneme recognition
Phoneme correct for continuous phoneme word
recognition in the case of a normal-hearing subject.
15. Visual-speech to text conversion applicable to telephone communication for deaf individuals
Phoneme correct for continuous phoneme word
recognition in the case of a deaf subject.
16. Visual-speech to text conversion applicable to telephone communication for deaf individuals
CONCLUSION
Hand shapes and lips shape were integrated
using concatenative feature fusion and HMM-based
automatic recognition was conducted.
For continuous phoneme recognition, a 86%
phoneme correct was achieved for the normal-hearing
cuer and a 82.7% phoneme correct for
the dead cuer were achieved, respectively.
Speech in both normal-hearing and deaf
subjects were also conducted obtaining a
94.9% and a 89% accuracy, respectively.
.
17. Visual-speech to text conversion applicable to telephone communication for deaf individuals
CONCLUSION
A multi-speaker experiment using data
from both normal-hearing and deaf subject
showed a 89.6% word accuracy, on
average.
This result indicates that training speaker-independent
HMMs for Cued Speech using
a large number of subjects should not face
particular difficulties
18. Visual-speech to text conversion applicable to telephone communication for deaf individuals
REFERENCES
G. Potamianos, C. Neti, G. Gravier, A. Garg, and A.W. Senior,
“recent Advances in the automatic recognition of audiovisual
speech,” in Proceedings of the IEEE, vol. 91, issue 9, pp.
1306–1326, 2003.
S. Nakamura, K. Kumatani, and S. Tamura, “Multi-modal
temporal asynchronicity modeling by product hmms for
robust audio-visual speech recognition,” in Proceedings of
Fourth IEEE International Conference on Multimodal
Interfaces (ICMI’02), p. 305, 2002.
R. O. Cornett, “Cued speech,” American Annals of the Deaf,
vol. 112, pp. 3–13, 1967.
J. Leybaert, “Phonology acquired through the eyes and
spelling in deaf children,”Journal of Experimental Child
Psychology, vol. 75, pp. 291– 318, 2000