SlideShare a Scribd company logo
1 of 6
Download to read offline
International Journal of Engineering Research and Development
e-ISSN: 2278-067X, p-ISSN : 2278-800X, www.ijerd.com
Volume 5, Issue 3 (December 2012), PP. 01-06

  Opinion Mining Using The Prosodic Information In The
  Spoken Utterance’s For Efficient Dialogue Management
              In A Spoken Dialogue System
                                 Manzoor Ahmad1, Romana Wani2
     1
         University of Kashmir, Scientist ‗D‘, Department of Computer Science, Hazratbal Srinagar 190006
                   2
                     , MCA Student, Department of Computer Science, Hazratbal Srinagar 190006


    Abstract:- Utterances used by a human while framing a response during the interaction with a
    software agent like Spoken Dialogue system(SDS) carries a lot of valuable information with regards to
    the internal mental state of the user. Opinion mining is an analysis of the mental state of a person, his
    opinion, appraisal or emotion towards an event, entities or their attributes. The users level of certainty
    about a topic could be determined by the analysis not only of the text used in the utterance but by
    studying the prosody information structure. Prosody reveals Information about the context by
    highlighting information structure and aspects of the speaker hearer relationship. Most often it is
    observed that the speaker‘s internal state is not depicted by the words he uses but by the tone of his
    utterance or facial expression of the user. In this paper we had analyzed a sample of student‘s after a
    lecture on operating system‘s subject and used the prosodic features of each dialogue with the student
    based on few questions. It was determined whether the students were certain, uncertain or neutral
    about their understanding of the lecture contents. This paper uses PRAAT a software tool for speech
    analysis which uses 15 acoustic features to determine the certainty of the responses of the user
    through classification by RAPIDMINER based on the prosody information which will actually aid the
    dialogue management component of the SDS in framing a better dialogue strategy.

    Keywords:- Uncertainty handling, Prosody information, Spoken language understanding, Machine
    learning.

                                       I.         INTRODUCTION
          Spoken language is an intuitive form of interaction between humans and computers. Spoken Language
Understanding has been a challenge in the design of the Spoken Dialogue System where the intention of the
speaker has to be identified from the words used in his utterances. Typically a spoken dialogue system
comprises a four main components an automatic speech recognition system (ASR), Spoken language
understanding component (SLU), Dialogue manager (DM) and an Speech synthesis system which converts the
text to speech (TTS). Spoken Language understanding deals with understanding the intent of the speaker from
the words used by him in his utterances. The accuracy of the speech recognition system is questionable and
researchers have provided various solutions to the problem of automatic speech recognition which lagged
behind human performance [2],[3] there have been some notable recent advances in discriminative training [4];
e.g., maximum mutual information (MMI) estimation [5], minimum classification error (MCE) training [6], [7],
and minimum phone error (MPE) training [8], [9]), in large-margin techniques (such as large margin estimation
[10], [11], large margin hidden Markov model (HMM) [12], large-margin MCE [13]–[15], and boosted MMI
[16]), as well as in novel acoustic models (such as conditional random fields (CRFs) [17]–[19], hidden CRFs
[20],[21] and segmental CRFs [22]),training densely connected, directed belief nets with many hidden layers
which learn a hierarchy of nonlinear feature detectors that can capture complex statistical patterns in data [23].
          There are many cases of experiences by the users when the computers either do not understand the
intended meaning of the user even after correctly recognizing the spoken utterances. One of the reason may be
that in a face to face human conversation, there are contextual, audio and visual cues [1] which aid the
knowledge requirements of the users for the efficient communication as the users other than hearing the
utterances are able to sense the mood and tone of the user by which they come to know whether the speaker is
certain or not. This is, absent in a dialogue between a computer and a human because in many potential
applications there is only audio input and no video input. If the Spoken Dialogue Systems are improved to use
the prosodic information from the spoken utterance they will definitely benefit from the level of certainty of the
user [24] such as spoken tutorial dialogue systems[25], language learning systems [26] and voice search
applications [27].
                                                        1
Opinion Mining Using The Prosodic Information In The Spoken...

          Our primary goal is to make use of prosodic information for aiding the dialogue manager in selecting
the dialogue strategy for effective interaction and influencing the final outcome. Technically Prosody is defined
as the rhythm, stress, and intonation of speech which reflect various features such as emotional state of the
speaker, the form of the utterance (statement, question, or command, the presence of irony or sarcasm,
emphasis, contrast, and focus or other elements of language) that may not be encoded by grammar or choice of
vocabulary. Prosodic information of an utterance can be used to determine how certain a speaker is and hence
the internal state of mind [28] which can be used for tasks from detecting frustration[29], to detecting flirtation
[30] and other intentions. The model proposed that uses prosodic information to classify utterances has
effectively colored the system responses in a student evaluation information system and performed better than a
trivial non-prosodic baseline model.
          In the context of human computer interaction, the study of prosodic information has been aimed at
extracting mood features in order to be able to dynamically adapt a dialog strategy by the automatic Spoken
Dialogue System.

                     II.         CORPUS AND CERTAINITY ANNOTATION
         It is very important to understand that not only what words are spoken by a speaker in his utterance but
how the words are spoken along with the certainty factor can actually guide the dialogue process between the
machine and the user. The spoken utterance may be perceived as uncertain, certain, neutral or mixed which
helps the dialogue system to make a guess about the mental state of the user about the utterance or about the
concept about which he is speaking about. In this paper we examine the impact of a lecture on operating system
subject on the understanding of the students as it is expressed within the context of a spoken dialogue.

                           AGENT :         What is an operating system.
                           STUDENT:        It is a set of software or hardware may be (UNCERTAIN)
                           AGENT :          Is it hardware or software.
                           STUDENT:        Software(CERTAIN)
                           AGENT :         What is the main function of Operating System.
                           STUDENT:        TO provide a interface between user and machine(CERTAIN)
                           AGENT :         What do you know about round robin scheduling.
                           STUDENT:        Uh-uhh (NEUTRAL)
                             Fig 1. An annotated excerpt from the student corpus.

         A corpus of 15 lecture related dialogs are selected and after listening each sentence of the student is
labeled by an annotator with either certain or uncertain or neutral. The dialog were also lexically annotated
based on the words used as certain, uncertain and neutral. The percentage of corpus with certainty, uncertainty
and neutral for the auditory and lexical conditions as annotated by listening to the audio of the dialog context
and annotated based on the lexical structure of the dialogues are shown in the Table I.

                       Table I : Percentage of Corpus with different levels of certainty.
                           CONDITIO                       LEVEL
                           N                Certain      Un-            Neutral
                                                         Certain
                           Auditory         22.3%        18.4%          59.3%
                            Lexical         12.1%        11.7%          76.2%

         It was observed that 40.7% non-neutral corpus could be decided as certain or uncertain based on the
audio and the dialog context compared to the 23.8% based on the lexical information. As such we used the
acoustic-prosody features for further information about the certainty or uncertainty.

                                     III.         METHODOLOGY
         For the basic model we compute values for 15 prosodic features as given in the Table I for each
utterance in the corpus of student lecture data set using PRAAT ( a program for speech analysis and synthesis)
[34] and Wavesurfer for extracting the f0 contour. Feature values are represented as z-scores normalized by
speaker. The temporal features like voice breaks, unvoiced frames, degree of voice breaks, Total duration are
not normalized. The set of features were selected in order to be comparable with Liscombe et al [ 31] who used
the same features along with turn related features for classifying uncertainty.



                                                        2
Opinion Mining Using The Prosodic Information In The Spoken...

                                                      TABLE II
                                          EXTRACTED AND SELECTED FEATURES.


                                N     NO. OF
                                      FEATURES                         FEATURES



                                                         Minimum, Maximum and Standard
                                                         deviation, relative position min f0,
                                          6
                                                         relative position max f0 statistics of
                                                         fundamental frequency (f0) Pitch
                                                         Minimum, Maximum, Mean
                                          4               and Standard Deviation (RMS), statistics
                                                         of Intensity
                                                         Ratio of voiced frames to total frames in
                                          1              the speech signal as an approximation of
                                                         speaking rate
                                          2              Total silence, Percent silence.
                                          2              Speaking duration, Total duration.


                                           IV.          RESULTS
          The features extracted are used as input variables to RAPID MINER machine learning software which
built C4.5 decision tree models that iteratively builds weak models and then combines them to form a better
model to predict the classification of unseen data. As an initial model we train a single decision tree using the
selected 15 features as listed in Table II. The model was evaluated over all the utterances of the corpus and it
classified within the classification classes, certain, uncertain and neutral and cross validated with an accuracy of
65% as compared to the non-prosodic model which had a an classification accuracy of 51.1%.




                             Fig. 2. Decision tree obtained after data discretisation.




                                         Fig. 2. A Plot of f0 and Output.
                                                         3
Opinion Mining Using The Prosodic Information In The Spoken...




                                   Fig. 2. Undiscretised Data Decision Tree.




                                       Fig. 4. The Performance Statistics

                                       V.          CONCLUSION
          In human computer interaction, the computer have to act human like so that other than the lexical
information, computer should be able to utilize the auditory and visual cues so that the users are responded in a
manner which is based on their emotions and the system looks to be more user friendly. In an automated
performance grading system students should be graded, selecting few based on his information and can help the
automated SGS to design a course layout which is more based on the preferences and prosodic information of
the user. When the system talks about certain concepts the prosodic features can indicate how much certain the
student is about the terms. Thus prosodic information provides information regarding the internal state of mind
of the user and would help the dialogue manager to dynamically select the strategy based on the certainty or
uncertainty.
          In our experiment we used a small set of prosodic features that have been examined in related work by
other researchers. Using and expanded set of features would improve the results and the accuracy with which
the certainty can be detected. In the future work we would be using the visual cues like facial expressions, body
language, emotions and other inputs by a human to maximize the ability to determine the internal mental state of
the user which can give the spoken dialogue system a mechanism to select dynamic dialogue strategy.

                                               REFERENCES
[1].     E. Krahmer and M. Swerts, ―How children and adults produce and perceive uncertainty in audiovisual
         speech,‖ Language and Speech, 48(1), 2005, pp 29–53
[2].     J. M. Baker, L. Deng, J. Glass, S. Khudanpur, C. Lee, N. Morgan, and D. O‘Shaugnessy, ―Research
         developments and directions in speech recognition and understanding, part 2,‖ IEEE Signal Processing
         Magazine, vol. 26, no. 4, pp. 78–85, 2009.
                                                       4
Opinion Mining Using The Prosodic Information In The Spoken...

[3].    J. M. Baker, L. Deng, J. Glass, S. Khudanpur, C. Lee, N. Morgan, and D. O‘Shaugnessy, ―Research
        developments and directions in speech recognition and understanding, part 1,‖ IEEE Signal Processing
        Magazine, vol. 26, no. 3, pp. 75–80, 2009.
[4].    X. He, L. Deng, and W. Chou, ―Discriminative learning in sequential pattern recognition — a unifying
        review for optimization-oriented speech recognition,‖ IEEE Signal Processing Magazine, vol. 25, no. 5,
        pp. 14– 36, 2008.
[5].    S. Kapadia, V. Valtchev, and S. J. Young, ―MMI training for continuous phoneme recognition on the
        TIMIT database,‖ In Proc. ICASSP, vol. 2, 1993, pp. 491–494.
[6].    B. H. Juang, W. Chou, and C. H. Lee, ―Minimum classification error rate methods for speech
        recognition,‖ IEEE Transactions on Speech and Audio Processing, vol. 5, no. 3, pp. 257–265, 1997.
[7].    E. McDermott, T. Hazen, J. L. Roux, A. Nakamura, and S. Katagiri, ―Discriminative training for large
        vocabulary speech recognition using minimum classification error,‖ IEEE Transactions on Speech and
        Audio Processing, vol. 15, no. 1, pp. 203–223, 2007.
[8].    D. Povey and P. Woodland, ―Minimum phone error and i-smoothing for improved discriminative
        training,‖ in Proc. ICASSP, 2002, pp. 105–108.
[9].    D. Povey, ―Discriminative training for large vocabulary speech recognition,‖ Ph.D. dissertation,
        Cambridge University Engineering Dept, 2003.
[10].    X. Li, H. Jiang, and C. Liu, ―Large margin HMMs for speech recognition,‖ in Proc. ICASSP, 2005, pp.
        513–516.
[11].   H. Jiang and X. Li, ―Incorporating training errors for large margin HMMs under semi-definite
        programming framework,‖ in Proc. ICASSP, vol. 4, 2007, pp. 629–632.
[12].   F. Sha and L. Saul, ―Large margin gaussian mixture modeling for phonetic classification and
        recognition,‖ in Proc. ICASSP, 2006, pp. 265–268.
[13].   D. Yu, L. Deng, X. He, and A. Acero, ―Use of incrementally regulated discriminative margins in MCE
        training for speech recognition,‖ in Proc. ICSLP, 2006, pp. 2418–2421.
[14].   D. Yu and L. Deng, ―Large-margin discriminative training of hidden Markov models for speech
        recognition,‖ in Proc. ICSC, 2007, pp. 429– 436.
[15].   D. Yu, L. Deng, X. He, and A. Acero, ―Large-margin minimum classification error training for large-
        scale speech recognition tasks,‖ in Proc. ICASSP, vol. 4, 2007, pp. 1137–1140.
[16].   D. Povey, D. Kanevsky, B. Kingsbury, B. Ramabhadran, G. Saon, and K. Visweswariah, ―Boosted
        MMI for model and feature space discriminative training,‖ in Proc. ICASSP, 2008, pp. 4057–4060.
[17].   Y. Hifny and S. Renals, ―Speech recognition using augmented conditional random fields,‖ IEEE
        Transactions on Audio, Speech & Language Processing, vol. 17, no. 2, pp. 354–365, 2009.
[18].   J. Morris and E. Fosler-Lussier, ―Combining phonetic attributes using conditional random fields,‖ in
        Proc. Interspeech, 2006, pp. 597–600.
[19].   G. Heigold, ―A log-linear discriminative modeling framework for speech recognition,‖ PhD Thesis,
        Aachen, Germany, 2010.
[20].   A. Gunawardana, M. Mahajan, A. Acero, and J. C. Platt, ―Hidden conditional random fields for phone
        classification,‖ in Proc. Interspeech, 2005, pp. 1117–1120.
[21].   D. Yu and L. Deng, ―Deep-structured hidden conditional random fields for phonetic recognition,‖ in
        Proc. Interspeech, 2010, pp. 2986–2989.
[22].    G. Zweig and P. Nguyen, ―A segmental conditional random field toolkit for speech recognition,‖ in
        Proc. Interspeech, 2010.
[23].   G. E. Hinton, S. Osindero, and Y. Teh, ―A fast learning algorithm for deep belief nets,‖ Neural
        Computation, vol. 18, pp. 1527–1554, 2006.
[24].   Heather Pon-Barry, Stuart M. Shieber, ―Recognizing Uncertainty in Speech‖, EURASIP Journal on
        Advances in Signal Processing, Volume 2011
[25].   K. Forbes-Riley and D. Litman, ―Adapting to student uncertainty improves tutoring dialogues,‖
        Frontiers in Artificial Intelligence and Applications, vol. 200, no. 1, pp. 33–40, 2009.
[26].   A. Alwan, Y. Bai, M. Black et al., ―A system for technology based assessment of language and literacy
        in young children: the role of multiple information sources,‖ in Proceedings of the 9th IEEE
        International Workshop on Multimedia Signal Processing (MMSP ‘07), pp. 26–30, Chania, Greece,
        October 2007.
[27].   T. Paek and Y.-C. Ju, ―Accommodating explicit user expressionsof uncertainty in voice search or
        something like that,‖ in Proceedings of the 9th Annual Conference of the International Speech
        Communication Association (INTERSPEECH ‘08), pp. 1165–1168, Brisbane, Australia, September
        2008.


                                                     5
Opinion Mining Using The Prosodic Information In The Spoken...

[28].    C. M. Lee and S. S. Narayanan, ―Toward detecting emotions in spoken dialogs,‖ IEEE Transactions
        on Speech and Audio Processing, vol. 13, no. 2, pp. 293–303, 2005.
[29].   J. Ang, R. Dhillon, A. Krupski, E. Shriberg, and A. Stolcke, ―Prosody-based automatic detection of
        annoyance and frustration in human computerdialog,‖ in Proceedings of the International Conference
        on Spoken Language Processing, pp. 2037–2040, Denver, Colo, USA, 2002.
[30].   R. Ranganath, D. Jurafsky, and D.McFarland, ― It‘s not you it‘s me: detecting flirting and its
        misperception in speed-dates,‖ in Proceedings of the Conference on Empirical Methods in Natural
        Language Processing (EMNLP ‘09), pp. 334–342, Singapore, 2009.
[31].   J. Liscombe, J. Hirschberg, and J. J. Venditti, ―Detecting certainness in spoken tutorial dialogues,‖ in
        Proceedings of the 9th European Conference on Speech Communication and Technology, pp. 1837–
        1840, Lisbon,Portugal, September 2005.
[32].   K. Forbes-Riley, D. Litman, and M. Rotaru, ―Responding to student uncertainty during computer
        tutoring: an experimental evaluation,‖ in Proceedings of the 9th International Conference on Intelligent
        Tutoring Systems (ITS ‘08), pp. 60–69, Montreal, Canada, June 2008.
[33].   J. C. Acosta and N. G. Ward, ―Responding to user emotional state by adding emotional coloring to
        utterances,‖ in Proceedings of the 10th Annual Conference of the International Speech Communication
        Association (INTERSPEECH ‘09), pp. 1587– 1590, Brighton, UK, September 2009.
[34].    P. Boersma, ―Praat, a system for doing phonetics by computer ―. Glot International, vol 5 no. 9/10,
        pp.341-345, 2001




                                                      6

More Related Content

What's hot

IRJET- Emotion recognition using Speech Signal: A Review
IRJET-  	  Emotion recognition using Speech Signal: A ReviewIRJET-  	  Emotion recognition using Speech Signal: A Review
IRJET- Emotion recognition using Speech Signal: A ReviewIRJET Journal
 
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTS
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTSEFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTS
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTSijnlc
 
Marathi Isolated Word Recognition System using MFCC and DTW Features
Marathi Isolated Word Recognition System using MFCC and DTW FeaturesMarathi Isolated Word Recognition System using MFCC and DTW Features
Marathi Isolated Word Recognition System using MFCC and DTW FeaturesIDES Editor
 
Report for Speech Emotion Recognition
Report for Speech Emotion RecognitionReport for Speech Emotion Recognition
Report for Speech Emotion RecognitionDongang (Sean) Wang
 
Isolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationIsolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationeSAT Publishing House
 
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...csandit
 
Speaker recognition in android
Speaker recognition in androidSpeaker recognition in android
Speaker recognition in androidAnshuli Mittal
 
A critical insight into multi-languages speech emotion databases
A critical insight into multi-languages speech emotion databasesA critical insight into multi-languages speech emotion databases
A critical insight into multi-languages speech emotion databasesjournalBEEI
 
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...mathsjournal
 
A Survey on Speaker Recognition System
A Survey on Speaker Recognition SystemA Survey on Speaker Recognition System
A Survey on Speaker Recognition SystemVani011
 
International journal of signal and image processing issues vol 2015 - no 1...
International journal of signal and image processing issues   vol 2015 - no 1...International journal of signal and image processing issues   vol 2015 - no 1...
International journal of signal and image processing issues vol 2015 - no 1...sophiabelthome
 
IRJET- Automatic Real-Time Speech Emotion Recognition using Support Vecto...
IRJET-  	  Automatic Real-Time Speech Emotion Recognition using Support Vecto...IRJET-  	  Automatic Real-Time Speech Emotion Recognition using Support Vecto...
IRJET- Automatic Real-Time Speech Emotion Recognition using Support Vecto...IRJET Journal
 
NLP BASED INTERVIEW ASSESSMENT SYSTEM
NLP BASED INTERVIEW ASSESSMENT SYSTEMNLP BASED INTERVIEW ASSESSMENT SYSTEM
NLP BASED INTERVIEW ASSESSMENT SYSTEMvivatechijri
 
5215ijcseit01
5215ijcseit015215ijcseit01
5215ijcseit01ijcsit
 

What's hot (15)

IRJET- Emotion recognition using Speech Signal: A Review
IRJET-  	  Emotion recognition using Speech Signal: A ReviewIRJET-  	  Emotion recognition using Speech Signal: A Review
IRJET- Emotion recognition using Speech Signal: A Review
 
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTS
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTSEFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTS
EFFECT OF MFCC BASED FEATURES FOR SPEECH SIGNAL ALIGNMENTS
 
Marathi Isolated Word Recognition System using MFCC and DTW Features
Marathi Isolated Word Recognition System using MFCC and DTW FeaturesMarathi Isolated Word Recognition System using MFCC and DTW Features
Marathi Isolated Word Recognition System using MFCC and DTW Features
 
Report for Speech Emotion Recognition
Report for Speech Emotion RecognitionReport for Speech Emotion Recognition
Report for Speech Emotion Recognition
 
Isolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantizationIsolated word recognition using lpc & vector quantization
Isolated word recognition using lpc & vector quantization
 
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...
COMBINED FEATURE EXTRACTION TECHNIQUES AND NAIVE BAYES CLASSIFIER FOR SPEECH ...
 
ECCAA
ECCAAECCAA
ECCAA
 
Speaker recognition in android
Speaker recognition in androidSpeaker recognition in android
Speaker recognition in android
 
A critical insight into multi-languages speech emotion databases
A critical insight into multi-languages speech emotion databasesA critical insight into multi-languages speech emotion databases
A critical insight into multi-languages speech emotion databases
 
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...
Speech Emotion Recognition by Using Combinations of Support Vector Machine (S...
 
A Survey on Speaker Recognition System
A Survey on Speaker Recognition SystemA Survey on Speaker Recognition System
A Survey on Speaker Recognition System
 
International journal of signal and image processing issues vol 2015 - no 1...
International journal of signal and image processing issues   vol 2015 - no 1...International journal of signal and image processing issues   vol 2015 - no 1...
International journal of signal and image processing issues vol 2015 - no 1...
 
IRJET- Automatic Real-Time Speech Emotion Recognition using Support Vecto...
IRJET-  	  Automatic Real-Time Speech Emotion Recognition using Support Vecto...IRJET-  	  Automatic Real-Time Speech Emotion Recognition using Support Vecto...
IRJET- Automatic Real-Time Speech Emotion Recognition using Support Vecto...
 
NLP BASED INTERVIEW ASSESSMENT SYSTEM
NLP BASED INTERVIEW ASSESSMENT SYSTEMNLP BASED INTERVIEW ASSESSMENT SYSTEM
NLP BASED INTERVIEW ASSESSMENT SYSTEM
 
5215ijcseit01
5215ijcseit015215ijcseit01
5215ijcseit01
 

Viewers also liked

Dia mundial decimo
Dia  mundial decimo Dia  mundial decimo
Dia mundial decimo tiffaeli
 
שבת בנחל הפולג 6.3.2010
שבת בנחל הפולג 6.3.2010שבת בנחל הפולג 6.3.2010
שבת בנחל הפולג 6.3.2010Hemi Shturman
 
Ppt noms de vaixells
Ppt noms de vaixellsPpt noms de vaixells
Ppt noms de vaixellsArnaulary
 
Киев-Класик, приглашение в Бельгию
Киев-Класик,  приглашение в БельгиюКиев-Класик,  приглашение в Бельгию
Киев-Класик, приглашение в БельгиюЛилия Горелая
 
Электронный Факультет Английского Языка
Электронный Факультет Английского ЯзыкаЭлектронный Факультет Английского Языка
Электронный Факультет Английского ЯзыкаNatalya Gataullina
 
татьянин день день студента
татьянин день   день студентататьянин день   день студента
татьянин день день студентаTatyanaPechennikova
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)IJERD Editor
 
新人が考える、UXと魔法の板っきれの話
新人が考える、UXと魔法の板っきれの話新人が考える、UXと魔法の板っきれの話
新人が考える、UXと魔法の板っきれの話Natsumi Kashiwa
 
Cover letter
Cover letter Cover letter
Cover letter mrym47
 
3 Destroy wrong thinking
3 Destroy wrong thinking3 Destroy wrong thinking
3 Destroy wrong thinkingLeon du Preez
 

Viewers also liked (17)

Dia mundial decimo
Dia  mundial decimo Dia  mundial decimo
Dia mundial decimo
 
Informat
InformatInformat
Informat
 
שבת בנחל הפולג 6.3.2010
שבת בנחל הפולג 6.3.2010שבת בנחל הפולג 6.3.2010
שבת בנחל הפולג 6.3.2010
 
Personal Brand Launch
Personal Brand LaunchPersonal Brand Launch
Personal Brand Launch
 
Ppt noms de vaixells
Ppt noms de vaixellsPpt noms de vaixells
Ppt noms de vaixells
 
Киев-Класик, приглашение в Бельгию
Киев-Класик,  приглашение в БельгиюКиев-Класик,  приглашение в Бельгию
Киев-Класик, приглашение в Бельгию
 
Электронный Факультет Английского Языка
Электронный Факультет Английского ЯзыкаЭлектронный Факультет Английского Языка
Электронный Факультет Английского Языка
 
Presentación1 mariiii
Presentación1 mariiiiPresentación1 mariiii
Presentación1 mariiii
 
татьянин день день студента
татьянин день   день студентататьянин день   день студента
татьянин день день студента
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)Welcome to International Journal of Engineering Research and Development (IJERD)
Welcome to International Journal of Engineering Research and Development (IJERD)
 
新人が考える、UXと魔法の板っきれの話
新人が考える、UXと魔法の板っきれの話新人が考える、UXと魔法の板っきれの話
新人が考える、UXと魔法の板っきれの話
 
Cover letter
Cover letter Cover letter
Cover letter
 
Brand audit
Brand auditBrand audit
Brand audit
 
3 Destroy wrong thinking
3 Destroy wrong thinking3 Destroy wrong thinking
3 Destroy wrong thinking
 
4 proreportcraigslistprofits
4 proreportcraigslistprofits4 proreportcraigslistprofits
4 proreportcraigslistprofits
 
Happy new year 2013
Happy new year 2013Happy new year 2013
Happy new year 2013
 

Similar to Welcome to International Journal of Engineering Research and Development (IJERD)

A Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningA Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningIRJET Journal
 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specificationijtsrd
 
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHM
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHMANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHM
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHMIRJET Journal
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifiereSAT Publishing House
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifiereSAT Journals
 
IRJET - Gesture based Communication Recognition System
IRJET -  	  Gesture based Communication Recognition SystemIRJET -  	  Gesture based Communication Recognition System
IRJET - Gesture based Communication Recognition SystemIRJET Journal
 
Signal Processing Tool for Emotion Recognition
Signal Processing Tool for Emotion RecognitionSignal Processing Tool for Emotion Recognition
Signal Processing Tool for Emotion Recognitionidescitation
 
Natural Language Processing (NLP).pdf
Natural Language Processing (NLP).pdfNatural Language Processing (NLP).pdf
Natural Language Processing (NLP).pdfMoar Digital 360
 
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...AIRCC Publishing Corporation
 
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...IRJET Journal
 
Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...ijctcm
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARijcseit
 
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...IRJET Journal
 
Speech emotion recognition using 2D-convolutional neural network
Speech emotion recognition using 2D-convolutional neural  networkSpeech emotion recognition using 2D-convolutional neural  network
Speech emotion recognition using 2D-convolutional neural networkIJECEIAES
 
SECOND LANGUAGE RESEARCH.pptx
SECOND LANGUAGE RESEARCH.pptxSECOND LANGUAGE RESEARCH.pptx
SECOND LANGUAGE RESEARCH.pptxssuser1ac0fa
 

Similar to Welcome to International Journal of Engineering Research and Development (IJERD) (20)

A Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep LearningA Review Paper on Speech Based Emotion Detection Using Deep Learning
A Review Paper on Speech Based Emotion Detection Using Deep Learning
 
A Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language SpecificationA Survey on Speech Recognition with Language Specification
A Survey on Speech Recognition with Language Specification
 
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHM
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHMANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHM
ANALYSING SPEECH EMOTION USING NEURAL NETWORK ALGORITHM
 
F334047
F334047F334047
F334047
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifier
 
Emotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifierEmotional telugu speech signals classification based on k nn classifier
Emotional telugu speech signals classification based on k nn classifier
 
IRJET - Gesture based Communication Recognition System
IRJET -  	  Gesture based Communication Recognition SystemIRJET -  	  Gesture based Communication Recognition System
IRJET - Gesture based Communication Recognition System
 
Signal Processing Tool for Emotion Recognition
Signal Processing Tool for Emotion RecognitionSignal Processing Tool for Emotion Recognition
Signal Processing Tool for Emotion Recognition
 
Natural Language Processing (NLP).pdf
Natural Language Processing (NLP).pdfNatural Language Processing (NLP).pdf
Natural Language Processing (NLP).pdf
 
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...
CURVELET BASED SPEECH RECOGNITION SYSTEM IN NOISY ENVIRONMENT: A STATISTICAL ...
 
Ijetcas14 390
Ijetcas14 390Ijetcas14 390
Ijetcas14 390
 
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...
DATABASES, FEATURES, CLASSIFIERS AND CHALLENGES IN AUTOMATIC SPEECH RECOGNITI...
 
Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...Automatic classification of bengali sentences based on sense definitions pres...
Automatic classification of bengali sentences based on sense definitions pres...
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMARSYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
SYLLABLE-BASED SPEECH RECOGNITION SYSTEM FOR MYANMAR
 
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...
STUDENT TEACHER INTERACTION ANALYSIS WITH EMOTION RECOGNITION FROM VIDEO AND ...
 
Speech emotion recognition using 2D-convolutional neural network
Speech emotion recognition using 2D-convolutional neural  networkSpeech emotion recognition using 2D-convolutional neural  network
Speech emotion recognition using 2D-convolutional neural network
 
SECOND LANGUAGE RESEARCH.pptx
SECOND LANGUAGE RESEARCH.pptxSECOND LANGUAGE RESEARCH.pptx
SECOND LANGUAGE RESEARCH.pptx
 

More from IJERD Editor

A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksA Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksIJERD Editor
 
MEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEMEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEIJERD Editor
 
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...IJERD Editor
 
Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’IJERD Editor
 
Reducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignReducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignIJERD Editor
 
Router 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationRouter 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationIJERD Editor
 
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...IJERD Editor
 
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRMitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRIJERD Editor
 
Study on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingStudy on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingIJERD Editor
 
Spyware triggering system by particular string value
Spyware triggering system by particular string valueSpyware triggering system by particular string value
Spyware triggering system by particular string valueIJERD Editor
 
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...IJERD Editor
 
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeSecure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeIJERD Editor
 
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...IJERD Editor
 
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraGesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraIJERD Editor
 
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...IJERD Editor
 
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...IJERD Editor
 
Moon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingMoon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingIJERD Editor
 
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...IJERD Editor
 
Importance of Measurements in Smart Grid
Importance of Measurements in Smart GridImportance of Measurements in Smart Grid
Importance of Measurements in Smart GridIJERD Editor
 
Study of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderStudy of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderIJERD Editor
 

More from IJERD Editor (20)

A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service AttacksA Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
A Novel Method for Prevention of Bandwidth Distributed Denial of Service Attacks
 
MEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACEMEMS MICROPHONE INTERFACE
MEMS MICROPHONE INTERFACE
 
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...Influence of tensile behaviour of slab on the structural Behaviour of shear c...
Influence of tensile behaviour of slab on the structural Behaviour of shear c...
 
Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’Gold prospecting using Remote Sensing ‘A case study of Sudan’
Gold prospecting using Remote Sensing ‘A case study of Sudan’
 
Reducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding DesignReducing Corrosion Rate by Welding Design
Reducing Corrosion Rate by Welding Design
 
Router 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and VerificationRouter 1X3 – RTL Design and Verification
Router 1X3 – RTL Design and Verification
 
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
Active Power Exchange in Distributed Power-Flow Controller (DPFC) At Third Ha...
 
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVRMitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
Mitigation of Voltage Sag/Swell with Fuzzy Control Reduced Rating DVR
 
Study on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive ManufacturingStudy on the Fused Deposition Modelling In Additive Manufacturing
Study on the Fused Deposition Modelling In Additive Manufacturing
 
Spyware triggering system by particular string value
Spyware triggering system by particular string valueSpyware triggering system by particular string value
Spyware triggering system by particular string value
 
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
A Blind Steganalysis on JPEG Gray Level Image Based on Statistical Features a...
 
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid SchemeSecure Image Transmission for Cloud Storage System Using Hybrid Scheme
Secure Image Transmission for Cloud Storage System Using Hybrid Scheme
 
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
Application of Buckley-Leverett Equation in Modeling the Radius of Invasion i...
 
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web CameraGesture Gaming on the World Wide Web Using an Ordinary Web Camera
Gesture Gaming on the World Wide Web Using an Ordinary Web Camera
 
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
Hardware Analysis of Resonant Frequency Converter Using Isolated Circuits And...
 
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
Simulated Analysis of Resonant Frequency Converter Using Different Tank Circu...
 
Moon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF DxingMoon-bounce: A Boon for VHF Dxing
Moon-bounce: A Boon for VHF Dxing
 
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
“MS-Extractor: An Innovative Approach to Extract Microsatellites on „Y‟ Chrom...
 
Importance of Measurements in Smart Grid
Importance of Measurements in Smart GridImportance of Measurements in Smart Grid
Importance of Measurements in Smart Grid
 
Study of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powderStudy of Macro level Properties of SCC using GGBS and Lime stone powder
Study of Macro level Properties of SCC using GGBS and Lime stone powder
 

Welcome to International Journal of Engineering Research and Development (IJERD)

  • 1. International Journal of Engineering Research and Development e-ISSN: 2278-067X, p-ISSN : 2278-800X, www.ijerd.com Volume 5, Issue 3 (December 2012), PP. 01-06 Opinion Mining Using The Prosodic Information In The Spoken Utterance’s For Efficient Dialogue Management In A Spoken Dialogue System Manzoor Ahmad1, Romana Wani2 1 University of Kashmir, Scientist ‗D‘, Department of Computer Science, Hazratbal Srinagar 190006 2 , MCA Student, Department of Computer Science, Hazratbal Srinagar 190006 Abstract:- Utterances used by a human while framing a response during the interaction with a software agent like Spoken Dialogue system(SDS) carries a lot of valuable information with regards to the internal mental state of the user. Opinion mining is an analysis of the mental state of a person, his opinion, appraisal or emotion towards an event, entities or their attributes. The users level of certainty about a topic could be determined by the analysis not only of the text used in the utterance but by studying the prosody information structure. Prosody reveals Information about the context by highlighting information structure and aspects of the speaker hearer relationship. Most often it is observed that the speaker‘s internal state is not depicted by the words he uses but by the tone of his utterance or facial expression of the user. In this paper we had analyzed a sample of student‘s after a lecture on operating system‘s subject and used the prosodic features of each dialogue with the student based on few questions. It was determined whether the students were certain, uncertain or neutral about their understanding of the lecture contents. This paper uses PRAAT a software tool for speech analysis which uses 15 acoustic features to determine the certainty of the responses of the user through classification by RAPIDMINER based on the prosody information which will actually aid the dialogue management component of the SDS in framing a better dialogue strategy. Keywords:- Uncertainty handling, Prosody information, Spoken language understanding, Machine learning. I. INTRODUCTION Spoken language is an intuitive form of interaction between humans and computers. Spoken Language Understanding has been a challenge in the design of the Spoken Dialogue System where the intention of the speaker has to be identified from the words used in his utterances. Typically a spoken dialogue system comprises a four main components an automatic speech recognition system (ASR), Spoken language understanding component (SLU), Dialogue manager (DM) and an Speech synthesis system which converts the text to speech (TTS). Spoken Language understanding deals with understanding the intent of the speaker from the words used by him in his utterances. The accuracy of the speech recognition system is questionable and researchers have provided various solutions to the problem of automatic speech recognition which lagged behind human performance [2],[3] there have been some notable recent advances in discriminative training [4]; e.g., maximum mutual information (MMI) estimation [5], minimum classification error (MCE) training [6], [7], and minimum phone error (MPE) training [8], [9]), in large-margin techniques (such as large margin estimation [10], [11], large margin hidden Markov model (HMM) [12], large-margin MCE [13]–[15], and boosted MMI [16]), as well as in novel acoustic models (such as conditional random fields (CRFs) [17]–[19], hidden CRFs [20],[21] and segmental CRFs [22]),training densely connected, directed belief nets with many hidden layers which learn a hierarchy of nonlinear feature detectors that can capture complex statistical patterns in data [23]. There are many cases of experiences by the users when the computers either do not understand the intended meaning of the user even after correctly recognizing the spoken utterances. One of the reason may be that in a face to face human conversation, there are contextual, audio and visual cues [1] which aid the knowledge requirements of the users for the efficient communication as the users other than hearing the utterances are able to sense the mood and tone of the user by which they come to know whether the speaker is certain or not. This is, absent in a dialogue between a computer and a human because in many potential applications there is only audio input and no video input. If the Spoken Dialogue Systems are improved to use the prosodic information from the spoken utterance they will definitely benefit from the level of certainty of the user [24] such as spoken tutorial dialogue systems[25], language learning systems [26] and voice search applications [27]. 1
  • 2. Opinion Mining Using The Prosodic Information In The Spoken... Our primary goal is to make use of prosodic information for aiding the dialogue manager in selecting the dialogue strategy for effective interaction and influencing the final outcome. Technically Prosody is defined as the rhythm, stress, and intonation of speech which reflect various features such as emotional state of the speaker, the form of the utterance (statement, question, or command, the presence of irony or sarcasm, emphasis, contrast, and focus or other elements of language) that may not be encoded by grammar or choice of vocabulary. Prosodic information of an utterance can be used to determine how certain a speaker is and hence the internal state of mind [28] which can be used for tasks from detecting frustration[29], to detecting flirtation [30] and other intentions. The model proposed that uses prosodic information to classify utterances has effectively colored the system responses in a student evaluation information system and performed better than a trivial non-prosodic baseline model. In the context of human computer interaction, the study of prosodic information has been aimed at extracting mood features in order to be able to dynamically adapt a dialog strategy by the automatic Spoken Dialogue System. II. CORPUS AND CERTAINITY ANNOTATION It is very important to understand that not only what words are spoken by a speaker in his utterance but how the words are spoken along with the certainty factor can actually guide the dialogue process between the machine and the user. The spoken utterance may be perceived as uncertain, certain, neutral or mixed which helps the dialogue system to make a guess about the mental state of the user about the utterance or about the concept about which he is speaking about. In this paper we examine the impact of a lecture on operating system subject on the understanding of the students as it is expressed within the context of a spoken dialogue. AGENT : What is an operating system. STUDENT: It is a set of software or hardware may be (UNCERTAIN) AGENT : Is it hardware or software. STUDENT: Software(CERTAIN) AGENT : What is the main function of Operating System. STUDENT: TO provide a interface between user and machine(CERTAIN) AGENT : What do you know about round robin scheduling. STUDENT: Uh-uhh (NEUTRAL) Fig 1. An annotated excerpt from the student corpus. A corpus of 15 lecture related dialogs are selected and after listening each sentence of the student is labeled by an annotator with either certain or uncertain or neutral. The dialog were also lexically annotated based on the words used as certain, uncertain and neutral. The percentage of corpus with certainty, uncertainty and neutral for the auditory and lexical conditions as annotated by listening to the audio of the dialog context and annotated based on the lexical structure of the dialogues are shown in the Table I. Table I : Percentage of Corpus with different levels of certainty. CONDITIO LEVEL N Certain Un- Neutral Certain Auditory 22.3% 18.4% 59.3% Lexical 12.1% 11.7% 76.2% It was observed that 40.7% non-neutral corpus could be decided as certain or uncertain based on the audio and the dialog context compared to the 23.8% based on the lexical information. As such we used the acoustic-prosody features for further information about the certainty or uncertainty. III. METHODOLOGY For the basic model we compute values for 15 prosodic features as given in the Table I for each utterance in the corpus of student lecture data set using PRAAT ( a program for speech analysis and synthesis) [34] and Wavesurfer for extracting the f0 contour. Feature values are represented as z-scores normalized by speaker. The temporal features like voice breaks, unvoiced frames, degree of voice breaks, Total duration are not normalized. The set of features were selected in order to be comparable with Liscombe et al [ 31] who used the same features along with turn related features for classifying uncertainty. 2
  • 3. Opinion Mining Using The Prosodic Information In The Spoken... TABLE II EXTRACTED AND SELECTED FEATURES. N NO. OF FEATURES FEATURES Minimum, Maximum and Standard deviation, relative position min f0, 6 relative position max f0 statistics of fundamental frequency (f0) Pitch Minimum, Maximum, Mean 4 and Standard Deviation (RMS), statistics of Intensity Ratio of voiced frames to total frames in 1 the speech signal as an approximation of speaking rate 2 Total silence, Percent silence. 2 Speaking duration, Total duration. IV. RESULTS The features extracted are used as input variables to RAPID MINER machine learning software which built C4.5 decision tree models that iteratively builds weak models and then combines them to form a better model to predict the classification of unseen data. As an initial model we train a single decision tree using the selected 15 features as listed in Table II. The model was evaluated over all the utterances of the corpus and it classified within the classification classes, certain, uncertain and neutral and cross validated with an accuracy of 65% as compared to the non-prosodic model which had a an classification accuracy of 51.1%. Fig. 2. Decision tree obtained after data discretisation. Fig. 2. A Plot of f0 and Output. 3
  • 4. Opinion Mining Using The Prosodic Information In The Spoken... Fig. 2. Undiscretised Data Decision Tree. Fig. 4. The Performance Statistics V. CONCLUSION In human computer interaction, the computer have to act human like so that other than the lexical information, computer should be able to utilize the auditory and visual cues so that the users are responded in a manner which is based on their emotions and the system looks to be more user friendly. In an automated performance grading system students should be graded, selecting few based on his information and can help the automated SGS to design a course layout which is more based on the preferences and prosodic information of the user. When the system talks about certain concepts the prosodic features can indicate how much certain the student is about the terms. Thus prosodic information provides information regarding the internal state of mind of the user and would help the dialogue manager to dynamically select the strategy based on the certainty or uncertainty. In our experiment we used a small set of prosodic features that have been examined in related work by other researchers. Using and expanded set of features would improve the results and the accuracy with which the certainty can be detected. In the future work we would be using the visual cues like facial expressions, body language, emotions and other inputs by a human to maximize the ability to determine the internal mental state of the user which can give the spoken dialogue system a mechanism to select dynamic dialogue strategy. REFERENCES [1]. E. Krahmer and M. Swerts, ―How children and adults produce and perceive uncertainty in audiovisual speech,‖ Language and Speech, 48(1), 2005, pp 29–53 [2]. J. M. Baker, L. Deng, J. Glass, S. Khudanpur, C. Lee, N. Morgan, and D. O‘Shaugnessy, ―Research developments and directions in speech recognition and understanding, part 2,‖ IEEE Signal Processing Magazine, vol. 26, no. 4, pp. 78–85, 2009. 4
  • 5. Opinion Mining Using The Prosodic Information In The Spoken... [3]. J. M. Baker, L. Deng, J. Glass, S. Khudanpur, C. Lee, N. Morgan, and D. O‘Shaugnessy, ―Research developments and directions in speech recognition and understanding, part 1,‖ IEEE Signal Processing Magazine, vol. 26, no. 3, pp. 75–80, 2009. [4]. X. He, L. Deng, and W. Chou, ―Discriminative learning in sequential pattern recognition — a unifying review for optimization-oriented speech recognition,‖ IEEE Signal Processing Magazine, vol. 25, no. 5, pp. 14– 36, 2008. [5]. S. Kapadia, V. Valtchev, and S. J. Young, ―MMI training for continuous phoneme recognition on the TIMIT database,‖ In Proc. ICASSP, vol. 2, 1993, pp. 491–494. [6]. B. H. Juang, W. Chou, and C. H. Lee, ―Minimum classification error rate methods for speech recognition,‖ IEEE Transactions on Speech and Audio Processing, vol. 5, no. 3, pp. 257–265, 1997. [7]. E. McDermott, T. Hazen, J. L. Roux, A. Nakamura, and S. Katagiri, ―Discriminative training for large vocabulary speech recognition using minimum classification error,‖ IEEE Transactions on Speech and Audio Processing, vol. 15, no. 1, pp. 203–223, 2007. [8]. D. Povey and P. Woodland, ―Minimum phone error and i-smoothing for improved discriminative training,‖ in Proc. ICASSP, 2002, pp. 105–108. [9]. D. Povey, ―Discriminative training for large vocabulary speech recognition,‖ Ph.D. dissertation, Cambridge University Engineering Dept, 2003. [10]. X. Li, H. Jiang, and C. Liu, ―Large margin HMMs for speech recognition,‖ in Proc. ICASSP, 2005, pp. 513–516. [11]. H. Jiang and X. Li, ―Incorporating training errors for large margin HMMs under semi-definite programming framework,‖ in Proc. ICASSP, vol. 4, 2007, pp. 629–632. [12]. F. Sha and L. Saul, ―Large margin gaussian mixture modeling for phonetic classification and recognition,‖ in Proc. ICASSP, 2006, pp. 265–268. [13]. D. Yu, L. Deng, X. He, and A. Acero, ―Use of incrementally regulated discriminative margins in MCE training for speech recognition,‖ in Proc. ICSLP, 2006, pp. 2418–2421. [14]. D. Yu and L. Deng, ―Large-margin discriminative training of hidden Markov models for speech recognition,‖ in Proc. ICSC, 2007, pp. 429– 436. [15]. D. Yu, L. Deng, X. He, and A. Acero, ―Large-margin minimum classification error training for large- scale speech recognition tasks,‖ in Proc. ICASSP, vol. 4, 2007, pp. 1137–1140. [16]. D. Povey, D. Kanevsky, B. Kingsbury, B. Ramabhadran, G. Saon, and K. Visweswariah, ―Boosted MMI for model and feature space discriminative training,‖ in Proc. ICASSP, 2008, pp. 4057–4060. [17]. Y. Hifny and S. Renals, ―Speech recognition using augmented conditional random fields,‖ IEEE Transactions on Audio, Speech & Language Processing, vol. 17, no. 2, pp. 354–365, 2009. [18]. J. Morris and E. Fosler-Lussier, ―Combining phonetic attributes using conditional random fields,‖ in Proc. Interspeech, 2006, pp. 597–600. [19]. G. Heigold, ―A log-linear discriminative modeling framework for speech recognition,‖ PhD Thesis, Aachen, Germany, 2010. [20]. A. Gunawardana, M. Mahajan, A. Acero, and J. C. Platt, ―Hidden conditional random fields for phone classification,‖ in Proc. Interspeech, 2005, pp. 1117–1120. [21]. D. Yu and L. Deng, ―Deep-structured hidden conditional random fields for phonetic recognition,‖ in Proc. Interspeech, 2010, pp. 2986–2989. [22]. G. Zweig and P. Nguyen, ―A segmental conditional random field toolkit for speech recognition,‖ in Proc. Interspeech, 2010. [23]. G. E. Hinton, S. Osindero, and Y. Teh, ―A fast learning algorithm for deep belief nets,‖ Neural Computation, vol. 18, pp. 1527–1554, 2006. [24]. Heather Pon-Barry, Stuart M. Shieber, ―Recognizing Uncertainty in Speech‖, EURASIP Journal on Advances in Signal Processing, Volume 2011 [25]. K. Forbes-Riley and D. Litman, ―Adapting to student uncertainty improves tutoring dialogues,‖ Frontiers in Artificial Intelligence and Applications, vol. 200, no. 1, pp. 33–40, 2009. [26]. A. Alwan, Y. Bai, M. Black et al., ―A system for technology based assessment of language and literacy in young children: the role of multiple information sources,‖ in Proceedings of the 9th IEEE International Workshop on Multimedia Signal Processing (MMSP ‘07), pp. 26–30, Chania, Greece, October 2007. [27]. T. Paek and Y.-C. Ju, ―Accommodating explicit user expressionsof uncertainty in voice search or something like that,‖ in Proceedings of the 9th Annual Conference of the International Speech Communication Association (INTERSPEECH ‘08), pp. 1165–1168, Brisbane, Australia, September 2008. 5
  • 6. Opinion Mining Using The Prosodic Information In The Spoken... [28]. C. M. Lee and S. S. Narayanan, ―Toward detecting emotions in spoken dialogs,‖ IEEE Transactions on Speech and Audio Processing, vol. 13, no. 2, pp. 293–303, 2005. [29]. J. Ang, R. Dhillon, A. Krupski, E. Shriberg, and A. Stolcke, ―Prosody-based automatic detection of annoyance and frustration in human computerdialog,‖ in Proceedings of the International Conference on Spoken Language Processing, pp. 2037–2040, Denver, Colo, USA, 2002. [30]. R. Ranganath, D. Jurafsky, and D.McFarland, ― It‘s not you it‘s me: detecting flirting and its misperception in speed-dates,‖ in Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP ‘09), pp. 334–342, Singapore, 2009. [31]. J. Liscombe, J. Hirschberg, and J. J. Venditti, ―Detecting certainness in spoken tutorial dialogues,‖ in Proceedings of the 9th European Conference on Speech Communication and Technology, pp. 1837– 1840, Lisbon,Portugal, September 2005. [32]. K. Forbes-Riley, D. Litman, and M. Rotaru, ―Responding to student uncertainty during computer tutoring: an experimental evaluation,‖ in Proceedings of the 9th International Conference on Intelligent Tutoring Systems (ITS ‘08), pp. 60–69, Montreal, Canada, June 2008. [33]. J. C. Acosta and N. G. Ward, ―Responding to user emotional state by adding emotional coloring to utterances,‖ in Proceedings of the 10th Annual Conference of the International Speech Communication Association (INTERSPEECH ‘09), pp. 1587– 1590, Brighton, UK, September 2009. [34]. P. Boersma, ―Praat, a system for doing phonetics by computer ―. Glot International, vol 5 no. 9/10, pp.341-345, 2001 6