SlideShare a Scribd company logo
1 of 7
Download to read offline
IOSR Journal of Computer Engineering (IOSRJCE)
ISSN: 2278-0661 Volume 4, Issue 4 (Sep.-Oct. 2012), PP 24-30
www.iosrjournals.org
www.iosrjournals.org 24 | Page
Speech Based Search Engine System Control and User
Interaction
Prof.R.Vijayasarathi1
, Prof.Dr.CManoharan2
1
Asst.prof, Department of Computer Application, Annai Mathammal Sheela Engg College, Namakkal.
2
principal, Annai Mathammal Sheela Engg College, Namakkal.
Abstract: To make a solution to the users who cannot able to interact with the system directly without the
keyboard and mouse. There is no such a need of typing by keyboard. This research will helps the users to type
and work with the system on the basis of voice control using the speech synthesis and speech recognition. With
this any user can access the computer without the help of mouse and keyboard. The user will have the ability to
use the computer without touching it. This will be done by using the features of Artificial Intelligence and the
Natural Language Processor.
Also this system will be very much useful to the differently abled persons and the visually challenged
persons who are difficult to read and know the details in their education and general issues in their real world.
Normally, the visually challenged persons acquire knowledge and exchange with others mainly through the
speech and writing. This system will help those persons in their education and the computer will be made user-
friendly. The process of speech based search engine system control and user interaction will be helpful to the
persons who are lacking in the computer knowledge and to interact with the computer without other person’s
help. This research is based on the fact that the computer will be able to interact with the user and fulfill their
needs in the computer world. In this research we have introduced a new approach of using Speech recognition
for educating visually challenged persons. A visually challenged person can easily interact with the computer
systems without the help of other persons.
Keywords: speech recognition, visually challenged persons, Artificial Intelligence.
I. Introduction
Speech processing is the study of speech signals and the processing methods of these signals. The
signals are usually processed in a digital presentation, so that speech processing can be regarded as a special
case of digital signal processing, applied to speech signal. It is also closely tied to natural language processing
(NLP)[2]
. E.g. text-to-speech synthesis may use a syntactic parser on its input text and speech recognition's
output may be used by e.g. information extraction techniques. Speech processing technology enables natural
interaction with all kinds of computers, from cell phones, PDAs, and PCs [7]
.
Speech processing can be divided into following categories:
 Speech Recognition - This deals with analysis of the linguistic content of a speech signal.
 Speaker Recognition - where the aim is to recognize the identity of the speaker.
 Speech Coding - a specialized form of data compression is important in the Telecommunication area.
 Voice Analysis for medical purposes, such as analysis of vocal loading and dysfunction of the vocal cords.
 Speech Synthesis - the artificial synthesis of speech, which usually means computer-generated speech.
 Speech Enhancement - Enhancing the intelligibility and/or perceptual quality of a speech signal, like audio
noise reduction for audio signals.
Benefits of speech interaction include:
 Interact with disabilities people (e.g. Visually Challenged)
 Users can move around their homes or offices freely and still interact with computers without having to sit
down and use a mouse or keyboard.
 Writing documents through speech.
 Hearing documents through speech.
This system will be very much useful for physically challenged persons and visually challenged
persons who are difficult to read and know the details about their education and general issues in their real world
[2]
. Normally, the visually challenged persons acquire knowledge and exchange with others mainly through the
speech and writing. This system will help those persons in their education and the computer will be made user-
friendly. The process of speech based search engine, system control and user interaction will be helpful to the
persons who are lacking in the computer knowledge and to interact with the computer without other person‟s
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 25 | Page
help. This research is based on the fact that the computer will be able to interact with the user and fulfill their
needs in the computer world [5]
.
In this Research we have introduced a new approach of using Speech recognition for educating visually
challenged persons. A visually challenged person can easily interact with the computer systems without any
help of other persons. India is now home to the world's largest number of blind people. The 37 million people
across the globe who are blind, over 15 million are from India and 15 million blind people in India, three
million, 26% of whom are children, suffer due to corneal disorders with 34% of the illiterate population in the
world, India has the largest number of illiterates by far.
1.1 OBJECTIVE OF THE RESEARCH
The Objective of this Research to make an intelligent environment in the speech processing with the
computer system by getting the Input from the user as speech input and it works for the request by the user and
the process will be made easily. This is also interacts with the users who are working with the system. Also to
make an innovative approach of establishing a communication with the visually challenged persons who can
able to work with the system directly by typing and operating using keyboard and mouse.
1.2 PROPOSED SYSTEM
The proposed system mainly focused on the process of making an interaction between the computer
systems with their own speaking language. This system can directly access the Google API to recognize the
human speaking sentence quickly. To the use of knowledge engineering to perform understand the human
statement and perform the action. This system will be very useful to the persons who can directly speak with
their computer [1]
. It helps the visually challenged students to get details about their education and other details
in the internet without typing in the keyboard or by other means. It will enable a person to get the basic details in
any website that is displayed in the screen through voice. It will dictate the details on the website or speaks with
the person who needs the details on any particular file or website or anywhere in the system. Our ultimate goal
is making Robotic Intelligence operating System (RIOS) with the help of this library and Natural Language
Processing for Visually challenged person. An innovative approach of developing a specific interface for the
users.
1.3 BENEFITS IN THE PROPOSED SYSTEM
 Users can work without the use of keyboard and mouse.
 Visually challenged persons will be able to access the computer system without other person‟s help.
 Users can access the system using their own voice itself.
 Also they can able to interact with the system. (i.e.) Any user can synthesize and recognize at a time.
1.4 FEASIBILITY STUDY
When a request is made to develop a system, a study is undertaken to ascertain if the research is
feasible whether it is worth the cost that the development of the system will involve, whether the necessary
hardware and software are available if not whether it is possible to buy them and finally whether the system will
be accepted and accommodated by those for whom the system is developed [3]
.
After the problem is clearly understood and the solution is proposed the next step is to make the feasibility
analysis. The existing system is more time consuming and inefficient. Typing all the data using speech is
inconsistent. Hence to access each application using the keyboard and mouse which is time consuming and
difficult to the visually challenged persons.
A detailed study is carried out to check the work ability of the proposed system. The feasibility study is
a test of system proposal regarding its work ability, impact on the organization, ability to meet user needs and
effective use of resources [5][9]
During the feasibility analysis for this research, the following areas are considered preliminary investigations
examine the research feasibility that is whether system is useful to the organization in many ways. Accurate
information can be obtained at right time. The four tests of feasibility are
 Economic Feasibility
 Feasibility Analysis
 Technical Feasibility
 Operational Feasibility
 Social Feasibility
1.4.1 ECONOMIC FEASIBILITY
Economic justification includes abroad range of concerns that includes cost benefit analysis. This
system will be developed and operated in the existing hardware and software infrastructure. Hence there is no
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 26 | Page
need of procuring additional hardware and software for the proposed system. The proposed will give the
information within minutes, hence the performance is improved.
Economic feasibility is checked by whether the financial benefits are exceeds the cost. This system uses only
the Open source software which is economically feasible.
The proposed system will minimize the time and efforts involved in processing, hence it is
economically feasible.
1.4.2 FEASIBILITY ANALYSIS
During system analysis the feasibility study of the proposed system was carried out to see, how far it
would be beneficial to the organization [9]
. A feasibility analysis is to test the system proposal according to its
workability, impact on the organization, ability to meet user needs and effectively use of resources for the
present day online messaging reliable resources of data transfer is not available and instant message is possible
[8]
.
1.4.3 TECHNICAL FEASIBILITY
The proposed system is said to be technically feasible if the necessary hardware and software are
available for the system. The technical feasibility issues during the feasibility stage of the investigation include,
 The apt technology is adopted to develop the system
 The system can be expanded and the organization should provide the sufficient equipment to develop the
system.
The proposed system is technically feasible if the following basic requirements are available,
 Windows 2000 and above
 New library RIOS.DSP will be used. It will integrates the Synthesis and Recognition process easily,
 Advanced domain such as Natural Language Processing will be used.
7
1.4.4 OPERATIONAL FEASIBILITY
Operational Feasibility is the ability and desire of the management, users and others to use and support
the proposed system.
The proposed system offers greater levels of user friendliness combined with effective services to the
users. After feasibility study the analyst has to find if there any faults or error still existing. When a problem is
defined clear it can be easily solved using appropriate solutions. Each module is developed using the appropriate
technology and the system is developed [10]
. Thus the proposed system increases the efficiency and provides
more benefits to the organization.
1.4.5 SOCIAL FEASIBILITY
Social Feasibility is the ability to make changes in the society or the country that the research will be
very helpful in its development of the education to the differently abled persons who cannot able to complete
their basic education and their knowledge from the internet without other persons help for their needs. Thus the
proposed system increases the benefits to the people and provides more benefits to the differently abled persons.
2.1 PROBLEM DEFINITION
Speech recognition applications include voice user interfaces such as voice dialing (e.g., "Call home"),
call routing (e.g., "I would like to make a collect call"), domotic appliance control, search (e.g., find a podcast
where particular words were spoken), simple data entry (e.g., entering a credit card number), preparation of
structured documents (e.g., a radiology report), speech-to-text processing (e.g., word processors or emails), and
aircraft (usually termed Direct Voice Input).
Synthesized speech is created by concatenating pieces of recorded speech that are stored in a database.
Systems differ in the size of the stored speech units; a system that stores phones or diaphones provides the
largest output range, but may lack clarity. For specific usage of domains, the storage of entire words or
sentences allows high-quality output. Alternatively, a synthesizer can incorporate a model of the vocal tract and
other human voice characteristics to create a completely "synthetic" voice output. These recognition and
synthesis process can be combined in a single program and this library provides a thin layer to combine these
processes together to form both interaction and recognition process.
2.2 OVERVIEW OF THE RESEARCH
As the Web transforms from a text only medium into a more multimedia rich medium the need arises to
perform searches based on the multimedia content. In this paper we present an audio and video search engine to
tackle this problem. The engine uses speech recognition technology to index spoken audio and video files from
the World Wide Web, when no transcriptions are available. If transcriptions (even imperfect ones) are available
we can also take advantage of them to improve the indexing process.
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 27 | Page
Our engine indexes several thousand talks and news radio shows covering a wide range of topics and
speaking styles from a selection of public web sites with multimedia archives. Our Web site is similar in spirit to
normal web search sites; it contains an index, not the actual multimedia content. The audio from these shows
suffers in acoustic quality due to bandwidth limitations, coding, compression, and poor acoustic conditions. Our
word-error rate results using appropriately trained acoustic models which shows remarkable resilience to the
high compression, though many factors combine to increase the average word-error rates over standard
broadcast news benchmarks. We show that, even if the transcription is inaccurate, we can still achieve good
retrieval performance for typical user queries (77.5%).
Speech Recognition is technology that can translate spoken words into text. Some SR systems use
"training" where an individual speaker reads sections of text into the SR system. These systems analyze the
person's specific voice and use it to fine tune the recognition of that person's speech, resulting in more accurate
transcription. Systems that do not use training are called "Speaker Independent" systems. Systems that use
training are called "Speaker Dependent" systems [2][6]
.
The term voice recognition refers to finding the identity of "who" is speaking, rather than what they are
saying. Recognizing the speaker voice recognition can simplify the task of translating speech in systems that
have been trained on specific person's voices or it can be used to authenticate or verify the identity of a speaker
as part of a security process. "Voice recognition" means "recognizing by voice", something humans do all the
time over the phone. As soon as someone familiar says "hello" the listener can identify them by the sound of
their voice alone.
Speech synthesis is the artificial production of human speech. A computer system used for this purpose
is called a speech synthesizer, and can be implemented in software or hardware. A text-to-speech (TTS) system
converts normal language text into speech; other systems render symbolic linguistic representations like
phonetic transcriptions into speech. These features of the speech recognition and synthesis combined together to
form a library for perform both recognition and synthesis process in a single machine. The user can search the
content by giving the work likewise the human can give the keyword through the speech sequence the system
can analysis and understand the human need and search the content and parse the raw content to extract the
exact knowledge data. After those process completed the synthesis can speak the data to the user. The data flow
diagram of this research is given in the figure 2.2 as data flow diagram.
Fig 2.2 Data Flow Diagram
2.3 MODULE DESCRIPTION
2.3.1 Speech Synthesis
2.3.2 Speech Recognition
2.3.3 Data Extraction and Knowledge Understand
2.3.4 Command Processing
2.3.1 SPEECH SYNTHESIS MODULE
 The text in any file or a source can then converted to the voice through synthesizer.
 It will convert the digital string to the respective voice through the US or UK Dictionary which we were
used.
 Then it will dictate the related word in the dictionary.
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 28 | Page
Fig .2.3.1 Speech Synthesis
A speech synthesizer converts written text into spoken language. Speech synthesis is also referred to as text-to-
speech (TTS) conversion. The proposed form of Speech Synthesis is shown in the Fig 2.3.1
.
2.3.2 SPEECH RECOGNITION MODULE
 Speech recognition is the process of converting spoken language to written text or some similar form.
 To get the command or voice from the user through microphone and recognize the acoustics.
 Then it will check the clarity of voice and pass to the Extraction part.
 We use the globalized grammar database using “Google API” via internet to perform Speech Recognition.
 Speech recognition is the process of converting an acoustic signal, captured by a microphone or a
telephone, to a set of words.
 The recognized words can be the final results for linguistic processing in order to reply the persons.
Fig .2.3.2 Speech Recognition
Speech recognition is the process of converting spoken language to written text or some similar form. Fig. 2.3.2
shows the process of Recognition.
The basic characteristics of a speech recognizer supporting the Java Speech API are:
1. It is mono-lingual: it supports a single specified language.
2. It processes a single input audio stream.
3. It can optionally adapt to the voice of its users.
4. Its grammars can be dynamically updated.
5. It has a small, defined set of application-controllable properties.
The major steps of a typical speech recognizer are:
1. Grammar
2. Signal processing
3. Phoneme recognition
4. Word recognition.
2.3.3 DATA EXTRACTION AND KNOWLEDGE UNDERSTAND MODULE
 The voice signal can be then converted to the digital format based on the Parser.
 It will convert the digital string to word.
 Then it will process the command given by the user and sent to the synthesizer.
 It redirects to synthesizer.
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 29 | Page
Fig .2.3.3 Data Extraction and Knowledge Understand
 Data Extraction performs the role of getting the raw factor of the user input from the internet. These raw
factor (XML Based data) could be mined and get the exact knowledge. Knowledge Understand can
understand the user keyword to perform both searching process as well as command processing.
Knowledge Understand module will understand the user‟s keyword to perform both searching process as
well as command processing methods.
2.3.4 COMMAND PROCESSING
 After understand the knowledge it can do the correct task, what the user need.
 Then the interaction processes simultaneously speak with user for effective communication and verify the
tasks.
Fig 2.3.4 Command Processing
After understanding the knowledge, it does the correct task, what the user needs. Then the interaction processes
simultaneously speak with user for effective communication and for verifying the tasks.
II. Input Design
The input design is the link that types the information system to the world of its user. It consists of
developing specifications and procedures for data preparations and to input into the computer for processing.
Simply we can Input Design is the process of converting the user-generated inputs into computer-based format.
THE IMPORTANT FEATURES ARE AS FOLLOWS
 The input screen is not over crowded, as the user can understand a piece of information from the screen.
Instead more screens are provided to understand the data entry process.
 The input validation is being done at program level to check errors and help messages to be provided to
effectively enter data into data entry screen in the software such as password checking, input checking etc.,
 As human is prone for errors and mistakes, confirmations for critical data entries are being provided. In the
research scheduler confirm messages are provide at such a critical areas like deletion or existing without
saving etc.
Speech Based Search Engine System Control And User Interaction
www.iosrjournals.org 30 | Page
3.2 OUTPUT DESIGN
All the output screens are information and interactive in such a way that the user can fulfill his/her
requirements. Output is information delivered to users through the information system. All the process of
quoting of tender, acquiring business, packing of goods, scheduling the transportation and follow up till delivery
etc., are shown in the output screen. Therefore an effective output design is an important feature of design
specification. Since useful output is essential for gaining user acceptance of the system, the system analysis
should try and follow the objectives, which are useful for designing acceptable output.
Most commonly outputs are printed on paper or displayed on screens. Outputs often have to be distributed to the
user. The increase in online, screen displayed output, that is personally accessible, has cut down on the problem
of distribution. Appropriate distribution is still an important objective for the system analyst.
III. CONCLUSION
This research represents an interactive communication between human and computer. The speech
synthesis and recognition process combines together and form an interactive system to search and control the
computer via speech processing. This could be done by using the concept of Natural Language Processing in
Artificial Intelligence. These processes can help the human quickly to access the content form the internet and
makes to listen easily. So the human no need search and read the content from the internet. And he/she can
easily control the system by using the system without using the keyboard and mouse.
3.1 FUTURE ENHANCEMENTS
The future goal is make it up to simple compare to this current system. And additionally will be
implement Voice analysis, Speech enhancement, Speech coding, Speech audio encoding/decoding, Speaker
analysis. The following features will be under implementing our ultimate goal to make Robotic Intelligence
operating System (RIOS) with the help of this library and Natural Language Processing for physically
challenged person.
 Adding new synthesized voices.
 Creating new library for create this synthesize voice.
 Automatic grammar generator for a large JSGF.
 To avoid time delay for fetching grammar word.
Portably imported to any application for interaction environment.3
Knowledgements
The author prof.R.vijayasarathi, wish to thank our chairman Mr.M.Mani, principal
prof.Dr.C.Manoharan, Hod Prof.M.Geetha and staff members of “Annai Mathammal Sheela
Engineering College Namakkal, TamilNadu” for his continuous encouragement for carrying
over my research works.
References
[1]. Alani, H. Southampton Univ., UK Sanghee Kim ; Millard, D.E. ; Weal, M.J. ; Hall, W. Lewis, P.H. ; Shadbolt, N.R. (2011),
Automatic ontology-based knowledge extraction from Web documents‟, Vol. 18, pp. 14 – 21.
[2]. Dragon Naturally Speaking http://www.scansoft.com/naturallyspeaking
[3]. Fetch Labs., Fetch Technol., El Segundo, CA, USA(2011), „Automatic Speech Recognition Based on Non-Uniform Error Criteria‟
Vol. 20, pp. 780 – 793.
[4]. FreeTTS http://freetts.sourceforge.net
[5]. JavaTM Speech API Programmer's Guide http://java.sun.com/products/ java- media/speech/forDevelopers/jsapi-guide/
[6]. Language development and everyday functioning of children with hearing loss assessedat 3 years of age, Teresa Y. C.
Ching, Kathryn Crowe, Vivienne Martin, Julia Day, Nicole Mahler, Samantha Youn, Laura Street, Cassandra Cook, Julia
Orsini,International Journal of Speech-Language Pathology Apr 2010, Vol. 12, No. 2: 124–131.
[7]. Roe, David B., and Jay G. Wilpon. Whither speech recognition: the next 25 years. IEEE communications magazine, v. 31, Nov.
1993: 54-62.
[8]. Raitio,T., „ HMM-Based Speech Synthesis Utilizing Glottal Inverse Filtering‟ Vol. 19, pp. 153 – 165.
[9]. Toth,J. Inst. of Telecommun., Slovak Univ. of Technol., Bratislava, Slovakia Kondelova,A. Rozinaj,G. (2011) „Natural language
processing of abbreviations‟ pp. 225 – 228.
[10]. Kenneth Thomas Schutte “Parts-based Models and Local Features for Automatic Speech Recognition” B.S., University of Illinois
at Urbana-Champaign (2001) S.M., Massachusetts Institute of Technology (2003).

More Related Content

What's hot

Developing a hands-free interface to operate a Computer using voice command
Developing a hands-free interface to operate a Computer using voice commandDeveloping a hands-free interface to operate a Computer using voice command
Developing a hands-free interface to operate a Computer using voice commandMohammad Liton Hossain
 
IRJET- My Buddy App: Communications between Smart Devices through Voice A...
IRJET-  	  My Buddy App: Communications between Smart Devices through Voice A...IRJET-  	  My Buddy App: Communications between Smart Devices through Voice A...
IRJET- My Buddy App: Communications between Smart Devices through Voice A...IRJET Journal
 
Human Computer Interaction - Interaction Design
Human Computer Interaction - Interaction DesignHuman Computer Interaction - Interaction Design
Human Computer Interaction - Interaction DesignVrushali Dhanokar
 
Automated UI & UX Framework
Automated UI & UX FrameworkAutomated UI & UX Framework
Automated UI & UX FrameworkIJARIIT
 
IRJET- Home Automation using Natural Language Interaction(NLI)
IRJET- Home Automation using Natural Language Interaction(NLI)IRJET- Home Automation using Natural Language Interaction(NLI)
IRJET- Home Automation using Natural Language Interaction(NLI)IRJET Journal
 
Touch Research 2: HCI Details [Handouts]
Touch Research 2: HCI Details [Handouts]Touch Research 2: HCI Details [Handouts]
Touch Research 2: HCI Details [Handouts]Harald Felgner, PhD
 
Affective computing by- Sandeep Jadhav
Affective computing by- Sandeep JadhavAffective computing by- Sandeep Jadhav
Affective computing by- Sandeep JadhavSandep Jadhav
 
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...Cemal Ardil
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET Journal
 
Predicting depression using deep learning and ensemble algorithms on raw twit...
Predicting depression using deep learning and ensemble algorithms on raw twit...Predicting depression using deep learning and ensemble algorithms on raw twit...
Predicting depression using deep learning and ensemble algorithms on raw twit...IJECEIAES
 
Human Computer Interface
Human Computer InterfaceHuman Computer Interface
Human Computer Interfacesanathkumar
 
Application for Iraqi sign language translation on Android system
Application for Iraqi sign language translation  on Android system Application for Iraqi sign language translation  on Android system
Application for Iraqi sign language translation on Android system IJECEIAES
 

What's hot (19)

Developing a hands-free interface to operate a Computer using voice command
Developing a hands-free interface to operate a Computer using voice commandDeveloping a hands-free interface to operate a Computer using voice command
Developing a hands-free interface to operate a Computer using voice command
 
IRJET- My Buddy App: Communications between Smart Devices through Voice A...
IRJET-  	  My Buddy App: Communications between Smart Devices through Voice A...IRJET-  	  My Buddy App: Communications between Smart Devices through Voice A...
IRJET- My Buddy App: Communications between Smart Devices through Voice A...
 
Human Computer Interaction - Interaction Design
Human Computer Interaction - Interaction DesignHuman Computer Interaction - Interaction Design
Human Computer Interaction - Interaction Design
 
2 4-10
2 4-102 4-10
2 4-10
 
Automated UI & UX Framework
Automated UI & UX FrameworkAutomated UI & UX Framework
Automated UI & UX Framework
 
IRJET- Home Automation using Natural Language Interaction(NLI)
IRJET- Home Automation using Natural Language Interaction(NLI)IRJET- Home Automation using Natural Language Interaction(NLI)
IRJET- Home Automation using Natural Language Interaction(NLI)
 
Ai applications study
Ai applications  studyAi applications  study
Ai applications study
 
CRC Final Report
CRC Final ReportCRC Final Report
CRC Final Report
 
Touch Research 2: HCI Details [Handouts]
Touch Research 2: HCI Details [Handouts]Touch Research 2: HCI Details [Handouts]
Touch Research 2: HCI Details [Handouts]
 
Affective computing by- Sandeep Jadhav
Affective computing by- Sandeep JadhavAffective computing by- Sandeep Jadhav
Affective computing by- Sandeep Jadhav
 
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...
A black-box-approach-for-response-quality-evaluation-of-conversational-agent-...
 
G1802024651
G1802024651G1802024651
G1802024651
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras ModelIRJET - Mutecom using Tensorflow-Keras Model
IRJET - Mutecom using Tensorflow-Keras Model
 
201500 Cognitive Informatics
201500 Cognitive Informatics201500 Cognitive Informatics
201500 Cognitive Informatics
 
Predicting depression using deep learning and ensemble algorithms on raw twit...
Predicting depression using deep learning and ensemble algorithms on raw twit...Predicting depression using deep learning and ensemble algorithms on raw twit...
Predicting depression using deep learning and ensemble algorithms on raw twit...
 
Human Computer Interface
Human Computer InterfaceHuman Computer Interface
Human Computer Interface
 
IoT [Internet of Things]
IoT [Internet of Things]IoT [Internet of Things]
IoT [Internet of Things]
 
Human Computer Interaction of an Information System
Human Computer Interaction of an Information SystemHuman Computer Interaction of an Information System
Human Computer Interaction of an Information System
 
Application for Iraqi sign language translation on Android system
Application for Iraqi sign language translation  on Android system Application for Iraqi sign language translation  on Android system
Application for Iraqi sign language translation on Android system
 

Viewers also liked

An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...
An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...
An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...IOSR Journals
 
The Analysis of Plyometric Training Program on University Handball Players
The Analysis of Plyometric Training Program on University Handball PlayersThe Analysis of Plyometric Training Program on University Handball Players
The Analysis of Plyometric Training Program on University Handball PlayersIOSR Journals
 
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata model
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata modelAnalyzing Employee’s Heart rate using Nonlinear Cellular Automata model
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata modelIOSR Journals
 
Optimal Estimating Sequence for a Hilbert Space Valued Parameter
Optimal Estimating Sequence for a Hilbert Space Valued ParameterOptimal Estimating Sequence for a Hilbert Space Valued Parameter
Optimal Estimating Sequence for a Hilbert Space Valued ParameterIOSR Journals
 
Levels of Physical Activity Participation of the Staff of Universiti Selangor
Levels of Physical Activity Participation of the Staff of Universiti SelangorLevels of Physical Activity Participation of the Staff of Universiti Selangor
Levels of Physical Activity Participation of the Staff of Universiti SelangorIOSR Journals
 
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...A Comparative Study of Certain Administrative and Academic Aspects of Minorit...
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...IOSR Journals
 
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...IOSR Journals
 
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...IOSR Journals
 
Performance Evaluation of Soft RoCE over 1 Gigabit Ethernet
Performance Evaluation of Soft RoCE over 1 Gigabit EthernetPerformance Evaluation of Soft RoCE over 1 Gigabit Ethernet
Performance Evaluation of Soft RoCE over 1 Gigabit EthernetIOSR Journals
 
Effect of astaxanthin on ethylene glycol induced nephrolithiasis
Effect of astaxanthin on ethylene glycol induced nephrolithiasisEffect of astaxanthin on ethylene glycol induced nephrolithiasis
Effect of astaxanthin on ethylene glycol induced nephrolithiasisIOSR Journals
 
Generalised Statistical Convergence For Double Sequences
Generalised Statistical Convergence For Double SequencesGeneralised Statistical Convergence For Double Sequences
Generalised Statistical Convergence For Double SequencesIOSR Journals
 
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.IOSR Journals
 
Postural Comparison of Dextrous and Ambidextrous Players
Postural Comparison of Dextrous and Ambidextrous PlayersPostural Comparison of Dextrous and Ambidextrous Players
Postural Comparison of Dextrous and Ambidextrous PlayersIOSR Journals
 
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...IOSR Journals
 

Viewers also liked (20)

An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...
An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...
An Analysis of Wushu's Theoretical Concept and its Reflection in Wushu's Prac...
 
The Analysis of Plyometric Training Program on University Handball Players
The Analysis of Plyometric Training Program on University Handball PlayersThe Analysis of Plyometric Training Program on University Handball Players
The Analysis of Plyometric Training Program on University Handball Players
 
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata model
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata modelAnalyzing Employee’s Heart rate using Nonlinear Cellular Automata model
Analyzing Employee’s Heart rate using Nonlinear Cellular Automata model
 
A0740103
A0740103A0740103
A0740103
 
Optimal Estimating Sequence for a Hilbert Space Valued Parameter
Optimal Estimating Sequence for a Hilbert Space Valued ParameterOptimal Estimating Sequence for a Hilbert Space Valued Parameter
Optimal Estimating Sequence for a Hilbert Space Valued Parameter
 
F0733038
F0733038F0733038
F0733038
 
D0952126
D0952126D0952126
D0952126
 
Levels of Physical Activity Participation of the Staff of Universiti Selangor
Levels of Physical Activity Participation of the Staff of Universiti SelangorLevels of Physical Activity Participation of the Staff of Universiti Selangor
Levels of Physical Activity Participation of the Staff of Universiti Selangor
 
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...A Comparative Study of Certain Administrative and Academic Aspects of Minorit...
A Comparative Study of Certain Administrative and Academic Aspects of Minorit...
 
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...
In-Vitro and In-Vivo Assessment of Anti-Asthmatic Activity of Polyherbal Ayur...
 
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...
Analysis of Cow’s Urine for Detection of Lipase Activity and Anti-Microbial P...
 
P0748292
P0748292P0748292
P0748292
 
Performance Evaluation of Soft RoCE over 1 Gigabit Ethernet
Performance Evaluation of Soft RoCE over 1 Gigabit EthernetPerformance Evaluation of Soft RoCE over 1 Gigabit Ethernet
Performance Evaluation of Soft RoCE over 1 Gigabit Ethernet
 
Effect of astaxanthin on ethylene glycol induced nephrolithiasis
Effect of astaxanthin on ethylene glycol induced nephrolithiasisEffect of astaxanthin on ethylene glycol induced nephrolithiasis
Effect of astaxanthin on ethylene glycol induced nephrolithiasis
 
Generalised Statistical Convergence For Double Sequences
Generalised Statistical Convergence For Double SequencesGeneralised Statistical Convergence For Double Sequences
Generalised Statistical Convergence For Double Sequences
 
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.
Design of a Dual-Band Microstrip Patch Antenna for GPS,WiMAX and WLAN.
 
D0321624
D0321624D0321624
D0321624
 
Postural Comparison of Dextrous and Ambidextrous Players
Postural Comparison of Dextrous and Ambidextrous PlayersPostural Comparison of Dextrous and Ambidextrous Players
Postural Comparison of Dextrous and Ambidextrous Players
 
G0544650
G0544650G0544650
G0544650
 
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...
Physicochemical and Bacteriological Analyses of Sachets Water Samples in Kano...
 

Similar to Speech Based Search Engine System Control and User Interaction

Speech enabled interactive voice response system
Speech enabled interactive voice response systemSpeech enabled interactive voice response system
Speech enabled interactive voice response systemeSAT Journals
 
Dialogue management system
Dialogue management systemDialogue management system
Dialogue management systemMayank Agarwal
 
VOCAL- Voice Command Application using Artificial Intelligence
VOCAL- Voice Command Application using Artificial IntelligenceVOCAL- Voice Command Application using Artificial Intelligence
VOCAL- Voice Command Application using Artificial IntelligenceIRJET Journal
 
A Voice Based Assistant Using Google Dialogflow And Machine Learning
A Voice Based Assistant Using Google Dialogflow And Machine LearningA Voice Based Assistant Using Google Dialogflow And Machine Learning
A Voice Based Assistant Using Google Dialogflow And Machine LearningEmily Smith
 
A Chatbot For Medical Purpose Using Deep Learning
A Chatbot For Medical Purpose Using Deep LearningA Chatbot For Medical Purpose Using Deep Learning
A Chatbot For Medical Purpose Using Deep LearningMartha Brown
 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET Journal
 
IRJET - Sign Language Converter
IRJET -  	  Sign Language ConverterIRJET -  	  Sign Language Converter
IRJET - Sign Language ConverterIRJET Journal
 
SURVEY ON SMART VIRTUAL VOICE ASSISTANT
SURVEY ON SMART VIRTUAL VOICE ASSISTANTSURVEY ON SMART VIRTUAL VOICE ASSISTANT
SURVEY ON SMART VIRTUAL VOICE ASSISTANTIRJET Journal
 
SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012Smarcos Eu
 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specificationijtsrd
 
Voicenotes - Android Based Smart Classroom
Voicenotes - Android Based Smart ClassroomVoicenotes - Android Based Smart Classroom
Voicenotes - Android Based Smart ClassroomIRJET Journal
 
Voice Assistant Using Python and AI
Voice Assistant Using Python and AIVoice Assistant Using Python and AI
Voice Assistant Using Python and AIIRJET Journal
 
Automated System Using Speech Recognition
Automated System Using Speech RecognitionAutomated System Using Speech Recognition
Automated System Using Speech RecognitionIRJET Journal
 
Top 5 Machine Learning Tools for Software Development in 2024.pdf
Top 5 Machine Learning Tools for Software Development in 2024.pdfTop 5 Machine Learning Tools for Software Development in 2024.pdf
Top 5 Machine Learning Tools for Software Development in 2024.pdfPolyxer Systems
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET Journal
 
Usability guidelines for usable user interface
Usability guidelines for usable user interfaceUsability guidelines for usable user interface
Usability guidelines for usable user interfaceeSAT Publishing House
 

Similar to Speech Based Search Engine System Control and User Interaction (20)

Speech enabled interactive voice response system
Speech enabled interactive voice response systemSpeech enabled interactive voice response system
Speech enabled interactive voice response system
 
Dialogue management system
Dialogue management systemDialogue management system
Dialogue management system
 
VOCAL- Voice Command Application using Artificial Intelligence
VOCAL- Voice Command Application using Artificial IntelligenceVOCAL- Voice Command Application using Artificial Intelligence
VOCAL- Voice Command Application using Artificial Intelligence
 
A Voice Based Assistant Using Google Dialogflow And Machine Learning
A Voice Based Assistant Using Google Dialogflow And Machine LearningA Voice Based Assistant Using Google Dialogflow And Machine Learning
A Voice Based Assistant Using Google Dialogflow And Machine Learning
 
A Chatbot For Medical Purpose Using Deep Learning
A Chatbot For Medical Purpose Using Deep LearningA Chatbot For Medical Purpose Using Deep Learning
A Chatbot For Medical Purpose Using Deep Learning
 
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
IRJET - A Robust Sign Language and Hand Gesture Recognition System using Conv...
 
IRJET - Sign Language Converter
IRJET -  	  Sign Language ConverterIRJET -  	  Sign Language Converter
IRJET - Sign Language Converter
 
SURVEY ON SMART VIRTUAL VOICE ASSISTANT
SURVEY ON SMART VIRTUAL VOICE ASSISTANTSURVEY ON SMART VIRTUAL VOICE ASSISTANT
SURVEY ON SMART VIRTUAL VOICE ASSISTANT
 
SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012SMARCOS Abstract Paper submitted to ICCHP 2012
SMARCOS Abstract Paper submitted to ICCHP 2012
 
30
3030
30
 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specification
 
sample PPT.pptx
sample PPT.pptxsample PPT.pptx
sample PPT.pptx
 
Voicenotes - Android Based Smart Classroom
Voicenotes - Android Based Smart ClassroomVoicenotes - Android Based Smart Classroom
Voicenotes - Android Based Smart Classroom
 
Voice Assistant Using Python and AI
Voice Assistant Using Python and AIVoice Assistant Using Python and AI
Voice Assistant Using Python and AI
 
Automated System Using Speech Recognition
Automated System Using Speech RecognitionAutomated System Using Speech Recognition
Automated System Using Speech Recognition
 
Machine learning
Machine learningMachine learning
Machine learning
 
Bt35408413
Bt35408413Bt35408413
Bt35408413
 
Top 5 Machine Learning Tools for Software Development in 2024.pdf
Top 5 Machine Learning Tools for Software Development in 2024.pdfTop 5 Machine Learning Tools for Software Development in 2024.pdf
Top 5 Machine Learning Tools for Software Development in 2024.pdf
 
IRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN MethodologyIRJET- Hand Gesture based Recognition using CNN Methodology
IRJET- Hand Gesture based Recognition using CNN Methodology
 
Usability guidelines for usable user interface
Usability guidelines for usable user interfaceUsability guidelines for usable user interface
Usability guidelines for usable user interface
 

More from IOSR Journals (20)

A011140104
A011140104A011140104
A011140104
 
M0111397100
M0111397100M0111397100
M0111397100
 
L011138596
L011138596L011138596
L011138596
 
K011138084
K011138084K011138084
K011138084
 
J011137479
J011137479J011137479
J011137479
 
I011136673
I011136673I011136673
I011136673
 
G011134454
G011134454G011134454
G011134454
 
H011135565
H011135565H011135565
H011135565
 
F011134043
F011134043F011134043
F011134043
 
E011133639
E011133639E011133639
E011133639
 
D011132635
D011132635D011132635
D011132635
 
C011131925
C011131925C011131925
C011131925
 
B011130918
B011130918B011130918
B011130918
 
A011130108
A011130108A011130108
A011130108
 
I011125160
I011125160I011125160
I011125160
 
H011124050
H011124050H011124050
H011124050
 
G011123539
G011123539G011123539
G011123539
 
F011123134
F011123134F011123134
F011123134
 
E011122530
E011122530E011122530
E011122530
 
D011121524
D011121524D011121524
D011121524
 

Recently uploaded

Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfNeo4j
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxLoriGlavin3
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentPim van der Noll
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationKnoldus Inc.
 
Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024TopCSSGallery
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityIES VE
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersNicole Novielli
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesBernd Ruecker
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch TuesdayIvanti
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterMydbops
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024BookNet Canada
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integrationmarketing932765
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxfnnc6jmgwh
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesManik S Magar
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfLoriGlavin3
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPathCommunity
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkPixlogix Infotech
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...itnewsafrica
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsNathaniel Shimoni
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxLoriGlavin3
 

Recently uploaded (20)

Connecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdfConnecting the Dots for Information Discovery.pdf
Connecting the Dots for Information Discovery.pdf
 
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptxPasskey Providers and Enabling Portability: FIDO Paris Seminar.pptx
Passkey Providers and Enabling Portability: FIDO Paris Seminar.pptx
 
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native developmentEmixa Mendix Meetup 11 April 2024 about Mendix Native development
Emixa Mendix Meetup 11 April 2024 about Mendix Native development
 
Data governance with Unity Catalog Presentation
Data governance with Unity Catalog PresentationData governance with Unity Catalog Presentation
Data governance with Unity Catalog Presentation
 
Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024Top 10 Hubspot Development Companies in 2024
Top 10 Hubspot Development Companies in 2024
 
Decarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a realityDecarbonising Buildings: Making a net-zero built environment a reality
Decarbonising Buildings: Making a net-zero built environment a reality
 
A Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software DevelopersA Journey Into the Emotions of Software Developers
A Journey Into the Emotions of Software Developers
 
QCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architecturesQCon London: Mastering long-running processes in modern architectures
QCon London: Mastering long-running processes in modern architectures
 
2024 April Patch Tuesday
2024 April Patch Tuesday2024 April Patch Tuesday
2024 April Patch Tuesday
 
Scale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL RouterScale your database traffic with Read & Write split using MySQL Router
Scale your database traffic with Read & Write split using MySQL Router
 
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
New from BookNet Canada for 2024: Loan Stars - Tech Forum 2024
 
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS:  6 Ways to Automate Your Data IntegrationBridging Between CAD & GIS:  6 Ways to Automate Your Data Integration
Bridging Between CAD & GIS: 6 Ways to Automate Your Data Integration
 
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptxGenerative AI - Gitex v1Generative AI - Gitex v1.pptx
Generative AI - Gitex v1Generative AI - Gitex v1.pptx
 
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotesMuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
MuleSoft Online Meetup Group - B2B Crash Course: Release SparkNotes
 
Moving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdfMoving Beyond Passwords: FIDO Paris Seminar.pdf
Moving Beyond Passwords: FIDO Paris Seminar.pdf
 
UiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to HeroUiPath Community: Communication Mining from Zero to Hero
UiPath Community: Communication Mining from Zero to Hero
 
React Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App FrameworkReact Native vs Ionic - The Best Mobile App Framework
React Native vs Ionic - The Best Mobile App Framework
 
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...Abdul Kader Baba- Managing Cybersecurity Risks  and Compliance Requirements i...
Abdul Kader Baba- Managing Cybersecurity Risks and Compliance Requirements i...
 
Time Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directionsTime Series Foundation Models - current state and future directions
Time Series Foundation Models - current state and future directions
 
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptxMerck Moving Beyond Passwords: FIDO Paris Seminar.pptx
Merck Moving Beyond Passwords: FIDO Paris Seminar.pptx
 

Speech Based Search Engine System Control and User Interaction

  • 1. IOSR Journal of Computer Engineering (IOSRJCE) ISSN: 2278-0661 Volume 4, Issue 4 (Sep.-Oct. 2012), PP 24-30 www.iosrjournals.org www.iosrjournals.org 24 | Page Speech Based Search Engine System Control and User Interaction Prof.R.Vijayasarathi1 , Prof.Dr.CManoharan2 1 Asst.prof, Department of Computer Application, Annai Mathammal Sheela Engg College, Namakkal. 2 principal, Annai Mathammal Sheela Engg College, Namakkal. Abstract: To make a solution to the users who cannot able to interact with the system directly without the keyboard and mouse. There is no such a need of typing by keyboard. This research will helps the users to type and work with the system on the basis of voice control using the speech synthesis and speech recognition. With this any user can access the computer without the help of mouse and keyboard. The user will have the ability to use the computer without touching it. This will be done by using the features of Artificial Intelligence and the Natural Language Processor. Also this system will be very much useful to the differently abled persons and the visually challenged persons who are difficult to read and know the details in their education and general issues in their real world. Normally, the visually challenged persons acquire knowledge and exchange with others mainly through the speech and writing. This system will help those persons in their education and the computer will be made user- friendly. The process of speech based search engine system control and user interaction will be helpful to the persons who are lacking in the computer knowledge and to interact with the computer without other person’s help. This research is based on the fact that the computer will be able to interact with the user and fulfill their needs in the computer world. In this research we have introduced a new approach of using Speech recognition for educating visually challenged persons. A visually challenged person can easily interact with the computer systems without the help of other persons. Keywords: speech recognition, visually challenged persons, Artificial Intelligence. I. Introduction Speech processing is the study of speech signals and the processing methods of these signals. The signals are usually processed in a digital presentation, so that speech processing can be regarded as a special case of digital signal processing, applied to speech signal. It is also closely tied to natural language processing (NLP)[2] . E.g. text-to-speech synthesis may use a syntactic parser on its input text and speech recognition's output may be used by e.g. information extraction techniques. Speech processing technology enables natural interaction with all kinds of computers, from cell phones, PDAs, and PCs [7] . Speech processing can be divided into following categories:  Speech Recognition - This deals with analysis of the linguistic content of a speech signal.  Speaker Recognition - where the aim is to recognize the identity of the speaker.  Speech Coding - a specialized form of data compression is important in the Telecommunication area.  Voice Analysis for medical purposes, such as analysis of vocal loading and dysfunction of the vocal cords.  Speech Synthesis - the artificial synthesis of speech, which usually means computer-generated speech.  Speech Enhancement - Enhancing the intelligibility and/or perceptual quality of a speech signal, like audio noise reduction for audio signals. Benefits of speech interaction include:  Interact with disabilities people (e.g. Visually Challenged)  Users can move around their homes or offices freely and still interact with computers without having to sit down and use a mouse or keyboard.  Writing documents through speech.  Hearing documents through speech. This system will be very much useful for physically challenged persons and visually challenged persons who are difficult to read and know the details about their education and general issues in their real world [2] . Normally, the visually challenged persons acquire knowledge and exchange with others mainly through the speech and writing. This system will help those persons in their education and the computer will be made user- friendly. The process of speech based search engine, system control and user interaction will be helpful to the persons who are lacking in the computer knowledge and to interact with the computer without other person‟s
  • 2. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 25 | Page help. This research is based on the fact that the computer will be able to interact with the user and fulfill their needs in the computer world [5] . In this Research we have introduced a new approach of using Speech recognition for educating visually challenged persons. A visually challenged person can easily interact with the computer systems without any help of other persons. India is now home to the world's largest number of blind people. The 37 million people across the globe who are blind, over 15 million are from India and 15 million blind people in India, three million, 26% of whom are children, suffer due to corneal disorders with 34% of the illiterate population in the world, India has the largest number of illiterates by far. 1.1 OBJECTIVE OF THE RESEARCH The Objective of this Research to make an intelligent environment in the speech processing with the computer system by getting the Input from the user as speech input and it works for the request by the user and the process will be made easily. This is also interacts with the users who are working with the system. Also to make an innovative approach of establishing a communication with the visually challenged persons who can able to work with the system directly by typing and operating using keyboard and mouse. 1.2 PROPOSED SYSTEM The proposed system mainly focused on the process of making an interaction between the computer systems with their own speaking language. This system can directly access the Google API to recognize the human speaking sentence quickly. To the use of knowledge engineering to perform understand the human statement and perform the action. This system will be very useful to the persons who can directly speak with their computer [1] . It helps the visually challenged students to get details about their education and other details in the internet without typing in the keyboard or by other means. It will enable a person to get the basic details in any website that is displayed in the screen through voice. It will dictate the details on the website or speaks with the person who needs the details on any particular file or website or anywhere in the system. Our ultimate goal is making Robotic Intelligence operating System (RIOS) with the help of this library and Natural Language Processing for Visually challenged person. An innovative approach of developing a specific interface for the users. 1.3 BENEFITS IN THE PROPOSED SYSTEM  Users can work without the use of keyboard and mouse.  Visually challenged persons will be able to access the computer system without other person‟s help.  Users can access the system using their own voice itself.  Also they can able to interact with the system. (i.e.) Any user can synthesize and recognize at a time. 1.4 FEASIBILITY STUDY When a request is made to develop a system, a study is undertaken to ascertain if the research is feasible whether it is worth the cost that the development of the system will involve, whether the necessary hardware and software are available if not whether it is possible to buy them and finally whether the system will be accepted and accommodated by those for whom the system is developed [3] . After the problem is clearly understood and the solution is proposed the next step is to make the feasibility analysis. The existing system is more time consuming and inefficient. Typing all the data using speech is inconsistent. Hence to access each application using the keyboard and mouse which is time consuming and difficult to the visually challenged persons. A detailed study is carried out to check the work ability of the proposed system. The feasibility study is a test of system proposal regarding its work ability, impact on the organization, ability to meet user needs and effective use of resources [5][9] During the feasibility analysis for this research, the following areas are considered preliminary investigations examine the research feasibility that is whether system is useful to the organization in many ways. Accurate information can be obtained at right time. The four tests of feasibility are  Economic Feasibility  Feasibility Analysis  Technical Feasibility  Operational Feasibility  Social Feasibility 1.4.1 ECONOMIC FEASIBILITY Economic justification includes abroad range of concerns that includes cost benefit analysis. This system will be developed and operated in the existing hardware and software infrastructure. Hence there is no
  • 3. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 26 | Page need of procuring additional hardware and software for the proposed system. The proposed will give the information within minutes, hence the performance is improved. Economic feasibility is checked by whether the financial benefits are exceeds the cost. This system uses only the Open source software which is economically feasible. The proposed system will minimize the time and efforts involved in processing, hence it is economically feasible. 1.4.2 FEASIBILITY ANALYSIS During system analysis the feasibility study of the proposed system was carried out to see, how far it would be beneficial to the organization [9] . A feasibility analysis is to test the system proposal according to its workability, impact on the organization, ability to meet user needs and effectively use of resources for the present day online messaging reliable resources of data transfer is not available and instant message is possible [8] . 1.4.3 TECHNICAL FEASIBILITY The proposed system is said to be technically feasible if the necessary hardware and software are available for the system. The technical feasibility issues during the feasibility stage of the investigation include,  The apt technology is adopted to develop the system  The system can be expanded and the organization should provide the sufficient equipment to develop the system. The proposed system is technically feasible if the following basic requirements are available,  Windows 2000 and above  New library RIOS.DSP will be used. It will integrates the Synthesis and Recognition process easily,  Advanced domain such as Natural Language Processing will be used. 7 1.4.4 OPERATIONAL FEASIBILITY Operational Feasibility is the ability and desire of the management, users and others to use and support the proposed system. The proposed system offers greater levels of user friendliness combined with effective services to the users. After feasibility study the analyst has to find if there any faults or error still existing. When a problem is defined clear it can be easily solved using appropriate solutions. Each module is developed using the appropriate technology and the system is developed [10] . Thus the proposed system increases the efficiency and provides more benefits to the organization. 1.4.5 SOCIAL FEASIBILITY Social Feasibility is the ability to make changes in the society or the country that the research will be very helpful in its development of the education to the differently abled persons who cannot able to complete their basic education and their knowledge from the internet without other persons help for their needs. Thus the proposed system increases the benefits to the people and provides more benefits to the differently abled persons. 2.1 PROBLEM DEFINITION Speech recognition applications include voice user interfaces such as voice dialing (e.g., "Call home"), call routing (e.g., "I would like to make a collect call"), domotic appliance control, search (e.g., find a podcast where particular words were spoken), simple data entry (e.g., entering a credit card number), preparation of structured documents (e.g., a radiology report), speech-to-text processing (e.g., word processors or emails), and aircraft (usually termed Direct Voice Input). Synthesized speech is created by concatenating pieces of recorded speech that are stored in a database. Systems differ in the size of the stored speech units; a system that stores phones or diaphones provides the largest output range, but may lack clarity. For specific usage of domains, the storage of entire words or sentences allows high-quality output. Alternatively, a synthesizer can incorporate a model of the vocal tract and other human voice characteristics to create a completely "synthetic" voice output. These recognition and synthesis process can be combined in a single program and this library provides a thin layer to combine these processes together to form both interaction and recognition process. 2.2 OVERVIEW OF THE RESEARCH As the Web transforms from a text only medium into a more multimedia rich medium the need arises to perform searches based on the multimedia content. In this paper we present an audio and video search engine to tackle this problem. The engine uses speech recognition technology to index spoken audio and video files from the World Wide Web, when no transcriptions are available. If transcriptions (even imperfect ones) are available we can also take advantage of them to improve the indexing process.
  • 4. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 27 | Page Our engine indexes several thousand talks and news radio shows covering a wide range of topics and speaking styles from a selection of public web sites with multimedia archives. Our Web site is similar in spirit to normal web search sites; it contains an index, not the actual multimedia content. The audio from these shows suffers in acoustic quality due to bandwidth limitations, coding, compression, and poor acoustic conditions. Our word-error rate results using appropriately trained acoustic models which shows remarkable resilience to the high compression, though many factors combine to increase the average word-error rates over standard broadcast news benchmarks. We show that, even if the transcription is inaccurate, we can still achieve good retrieval performance for typical user queries (77.5%). Speech Recognition is technology that can translate spoken words into text. Some SR systems use "training" where an individual speaker reads sections of text into the SR system. These systems analyze the person's specific voice and use it to fine tune the recognition of that person's speech, resulting in more accurate transcription. Systems that do not use training are called "Speaker Independent" systems. Systems that use training are called "Speaker Dependent" systems [2][6] . The term voice recognition refers to finding the identity of "who" is speaking, rather than what they are saying. Recognizing the speaker voice recognition can simplify the task of translating speech in systems that have been trained on specific person's voices or it can be used to authenticate or verify the identity of a speaker as part of a security process. "Voice recognition" means "recognizing by voice", something humans do all the time over the phone. As soon as someone familiar says "hello" the listener can identify them by the sound of their voice alone. Speech synthesis is the artificial production of human speech. A computer system used for this purpose is called a speech synthesizer, and can be implemented in software or hardware. A text-to-speech (TTS) system converts normal language text into speech; other systems render symbolic linguistic representations like phonetic transcriptions into speech. These features of the speech recognition and synthesis combined together to form a library for perform both recognition and synthesis process in a single machine. The user can search the content by giving the work likewise the human can give the keyword through the speech sequence the system can analysis and understand the human need and search the content and parse the raw content to extract the exact knowledge data. After those process completed the synthesis can speak the data to the user. The data flow diagram of this research is given in the figure 2.2 as data flow diagram. Fig 2.2 Data Flow Diagram 2.3 MODULE DESCRIPTION 2.3.1 Speech Synthesis 2.3.2 Speech Recognition 2.3.3 Data Extraction and Knowledge Understand 2.3.4 Command Processing 2.3.1 SPEECH SYNTHESIS MODULE  The text in any file or a source can then converted to the voice through synthesizer.  It will convert the digital string to the respective voice through the US or UK Dictionary which we were used.  Then it will dictate the related word in the dictionary.
  • 5. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 28 | Page Fig .2.3.1 Speech Synthesis A speech synthesizer converts written text into spoken language. Speech synthesis is also referred to as text-to- speech (TTS) conversion. The proposed form of Speech Synthesis is shown in the Fig 2.3.1 . 2.3.2 SPEECH RECOGNITION MODULE  Speech recognition is the process of converting spoken language to written text or some similar form.  To get the command or voice from the user through microphone and recognize the acoustics.  Then it will check the clarity of voice and pass to the Extraction part.  We use the globalized grammar database using “Google API” via internet to perform Speech Recognition.  Speech recognition is the process of converting an acoustic signal, captured by a microphone or a telephone, to a set of words.  The recognized words can be the final results for linguistic processing in order to reply the persons. Fig .2.3.2 Speech Recognition Speech recognition is the process of converting spoken language to written text or some similar form. Fig. 2.3.2 shows the process of Recognition. The basic characteristics of a speech recognizer supporting the Java Speech API are: 1. It is mono-lingual: it supports a single specified language. 2. It processes a single input audio stream. 3. It can optionally adapt to the voice of its users. 4. Its grammars can be dynamically updated. 5. It has a small, defined set of application-controllable properties. The major steps of a typical speech recognizer are: 1. Grammar 2. Signal processing 3. Phoneme recognition 4. Word recognition. 2.3.3 DATA EXTRACTION AND KNOWLEDGE UNDERSTAND MODULE  The voice signal can be then converted to the digital format based on the Parser.  It will convert the digital string to word.  Then it will process the command given by the user and sent to the synthesizer.  It redirects to synthesizer.
  • 6. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 29 | Page Fig .2.3.3 Data Extraction and Knowledge Understand  Data Extraction performs the role of getting the raw factor of the user input from the internet. These raw factor (XML Based data) could be mined and get the exact knowledge. Knowledge Understand can understand the user keyword to perform both searching process as well as command processing. Knowledge Understand module will understand the user‟s keyword to perform both searching process as well as command processing methods. 2.3.4 COMMAND PROCESSING  After understand the knowledge it can do the correct task, what the user need.  Then the interaction processes simultaneously speak with user for effective communication and verify the tasks. Fig 2.3.4 Command Processing After understanding the knowledge, it does the correct task, what the user needs. Then the interaction processes simultaneously speak with user for effective communication and for verifying the tasks. II. Input Design The input design is the link that types the information system to the world of its user. It consists of developing specifications and procedures for data preparations and to input into the computer for processing. Simply we can Input Design is the process of converting the user-generated inputs into computer-based format. THE IMPORTANT FEATURES ARE AS FOLLOWS  The input screen is not over crowded, as the user can understand a piece of information from the screen. Instead more screens are provided to understand the data entry process.  The input validation is being done at program level to check errors and help messages to be provided to effectively enter data into data entry screen in the software such as password checking, input checking etc.,  As human is prone for errors and mistakes, confirmations for critical data entries are being provided. In the research scheduler confirm messages are provide at such a critical areas like deletion or existing without saving etc.
  • 7. Speech Based Search Engine System Control And User Interaction www.iosrjournals.org 30 | Page 3.2 OUTPUT DESIGN All the output screens are information and interactive in such a way that the user can fulfill his/her requirements. Output is information delivered to users through the information system. All the process of quoting of tender, acquiring business, packing of goods, scheduling the transportation and follow up till delivery etc., are shown in the output screen. Therefore an effective output design is an important feature of design specification. Since useful output is essential for gaining user acceptance of the system, the system analysis should try and follow the objectives, which are useful for designing acceptable output. Most commonly outputs are printed on paper or displayed on screens. Outputs often have to be distributed to the user. The increase in online, screen displayed output, that is personally accessible, has cut down on the problem of distribution. Appropriate distribution is still an important objective for the system analyst. III. CONCLUSION This research represents an interactive communication between human and computer. The speech synthesis and recognition process combines together and form an interactive system to search and control the computer via speech processing. This could be done by using the concept of Natural Language Processing in Artificial Intelligence. These processes can help the human quickly to access the content form the internet and makes to listen easily. So the human no need search and read the content from the internet. And he/she can easily control the system by using the system without using the keyboard and mouse. 3.1 FUTURE ENHANCEMENTS The future goal is make it up to simple compare to this current system. And additionally will be implement Voice analysis, Speech enhancement, Speech coding, Speech audio encoding/decoding, Speaker analysis. The following features will be under implementing our ultimate goal to make Robotic Intelligence operating System (RIOS) with the help of this library and Natural Language Processing for physically challenged person.  Adding new synthesized voices.  Creating new library for create this synthesize voice.  Automatic grammar generator for a large JSGF.  To avoid time delay for fetching grammar word. Portably imported to any application for interaction environment.3 Knowledgements The author prof.R.vijayasarathi, wish to thank our chairman Mr.M.Mani, principal prof.Dr.C.Manoharan, Hod Prof.M.Geetha and staff members of “Annai Mathammal Sheela Engineering College Namakkal, TamilNadu” for his continuous encouragement for carrying over my research works. References [1]. Alani, H. Southampton Univ., UK Sanghee Kim ; Millard, D.E. ; Weal, M.J. ; Hall, W. Lewis, P.H. ; Shadbolt, N.R. (2011), Automatic ontology-based knowledge extraction from Web documents‟, Vol. 18, pp. 14 – 21. [2]. Dragon Naturally Speaking http://www.scansoft.com/naturallyspeaking [3]. Fetch Labs., Fetch Technol., El Segundo, CA, USA(2011), „Automatic Speech Recognition Based on Non-Uniform Error Criteria‟ Vol. 20, pp. 780 – 793. [4]. FreeTTS http://freetts.sourceforge.net [5]. JavaTM Speech API Programmer's Guide http://java.sun.com/products/ java- media/speech/forDevelopers/jsapi-guide/ [6]. Language development and everyday functioning of children with hearing loss assessedat 3 years of age, Teresa Y. C. Ching, Kathryn Crowe, Vivienne Martin, Julia Day, Nicole Mahler, Samantha Youn, Laura Street, Cassandra Cook, Julia Orsini,International Journal of Speech-Language Pathology Apr 2010, Vol. 12, No. 2: 124–131. [7]. Roe, David B., and Jay G. Wilpon. Whither speech recognition: the next 25 years. IEEE communications magazine, v. 31, Nov. 1993: 54-62. [8]. Raitio,T., „ HMM-Based Speech Synthesis Utilizing Glottal Inverse Filtering‟ Vol. 19, pp. 153 – 165. [9]. Toth,J. Inst. of Telecommun., Slovak Univ. of Technol., Bratislava, Slovakia Kondelova,A. Rozinaj,G. (2011) „Natural language processing of abbreviations‟ pp. 225 – 228. [10]. Kenneth Thomas Schutte “Parts-based Models and Local Features for Automatic Speech Recognition” B.S., University of Illinois at Urbana-Champaign (2001) S.M., Massachusetts Institute of Technology (2003).