SlideShare a Scribd company logo
1 of 42
Dialog System
A comprehensive understanding
Mr. T
Perception
Dialog System
Natural Language
Understanding
• Domain identification
• User intent detection
• Slot Filling
Dialogue
Management
• Dialogue state tracking
• Dialogue policy optimization
Natural
Language
Generation
Backend
Knowledge
Providers
Speech
Recognition
Speech
Synthesis
Utterance
Response
Trigger
Semantic Frame
Ask_weather(date=weekends)
Text Input
“I will go out at weekends, what is the weather?”
System Action
Request_locationText Response
“Where will you go?/”Where you want to ask for the weather this weekends?”
Trigger Word
A solution for a trigger word system
Convoluted Neural Network Recurrent Neural Network
Wave Sound Frequency Domain
Trigger
Word
Unknown
Output
Low level features
high level features
Patterns extraction Classifier
What’s else…
Hey Bot/Ok Bot!
Speech Recognition
Speech Recognition
Acoustic Dictionary
(Pronunciation model)
Pre-processing
Acoustic Features
Speech wave
Decoder Acoustic Model
WORD PRON (ipa)
v ə ˨˩ˀ
w e
vợ
quê
“Vợ tôi ở quê rất đẹp”
NGRAMs SCORE
Vợ
quê
2.5
0.7
Language Model
v
w
Speech Synthesis
Ideas of TTS
PG & E will file schedules on April, 20
https://web.stanford.edu/~jurafsky/slp3/ed3book.pdf
Look up and Concatenate
No!
Pipeline for Text To Speech
Text Input
• Sentence Tokenization
• Non-standard words
• Trained by Machine Learning (SVM, DT, LR)
Text Normalization
• Disambiguation words
• Dictionary Look Up
Phonetic Analysis
• Names Spelling
• Grapheme 2 Phoneme
𝑃 = 𝑎𝑟𝑔𝑚𝑎𝑥 𝑃 (𝑃|𝑆)
Prosodic Analysis
• Prosody Structure
• Prosody Prominence
• Tune
Voice Output
https://web.stanford.edu/~jurafsky/slp3/ed3book.pdf
The state-of-the-art
Where GAN comes into play.
Natural Language Understanding
King
Queen
Man
Woman
Boy
Girl
(0.12, 0.23, 0.43)
(0.14, 0.57, 0.88)
(0.44, 0.90, 0.11)
(0.19, 0.23, 0.53)
(0.12, 0.65, 0.42)
(0.34, 0.44, 0.68)
EmbeddingWords
Word Embedding
(1, 0, 0, 0, 0, 0, 0)
(0, 1, 0, 0, 0, 0, 0)
(0, 0, 1, 0, 0, 0, 0)
)
(0, 0, 0, 0, 1, 0, 0)
(0, 0, 0, 0, 0, 1, 0)
One-hot Encoding Vector
10 0 1 0
0 0 0 2
4 0 7 0
0 5 0 12
Frequency Based Vector
Documents
1 2 3 4
1
2
3
4
Terms
Docs Vector
word Vector
word Vector
man
woman
queen
king Predication Based Vector
One-hot Encoding Vector
1 2 3 4 5 6 7 8
Co gai 1 0 0 0 0 0 0 0
hot girl 0 1 0 0 0 0 0 0
xinh dep 0 0 1 0 0 0 0 0
truoc day 0 0 0 1 0 0 0 0
la 0 0 0 0 1 0 0 0
mot 0 0 0 0 0 1 0 0
chang trai 0 0 0 0 0 0 1 0
dam my 0 0 0 0 0 0 0 1
each word
Gets a 1x 8
vector
representation
What’s wrong…
Corpus:
Co gai, hot girl,
xinh đep, truoc
day, la, mot,
chang trai, dam
my
Custom Encoding Vector
Corpus:
Co gai, hot girl,
xinh đep, truoc
day, la, mot,
chang trai, dam
my
nguoi ban chat
Thoi
gian
So
dem
Nu
tinh
Co gai 1 0 0 0 1
hot girl 0.7 1 0 0 0.7
xinh dep 0.6 1 0 0 0.5
truoc day 0 0 1 1 0
la 0 0 0 0 0
mot 0 0 0 1 0
chang trai 1 0 0 0 0
dam my 0.7 1 0 0 0
each word
Gets a 1x5
vector
representation
Custom Encoding Vector
each word
Gets a 1x5
vector
representation
And better
relationship
Corpus:
Co gai, hot girl,
xinh đep, truoc
day, la, mot,
chang trai, dam
my
nguoi ban chat
Thoi
gian
So
dem
Nu
tinh
Co gai 1 0 0 0 1
hot girl 0.7 1 0 0 0.7
xinh dep 0.6 1 0 0 0.5
truoc day 0 0 1 1 0
la 0 0 0 0 0
mot 0 0 0 1 0
chang trai 1 0 0 0 0
dam my 0.7 1 0 0 0
Count Vector
Let us understand this using a simple example.
• D1: He is a lazy boy. She is also lazy.
• D2: Neeraj is a lazy person.
Dictionary = [‘He’, ‘She’, ‘lazy’, ‘boy’, ‘Neeraj’, ‘person’]
D=2 (# docs), N=6 (# words in the dictionary)
He She lazy boy Neeraj person
D1 1 1 2 1 0 0
D2 0 0 1 0 1 1
Count Vector matrix M = DXN, vector (“lazy”) = [2, 1]
TF-IDF vectorization
TF = (Number of times term t appears in a document)/(Number of terms
in the document)
So, TF(This,Document1) = 1/8
TF(This, Document2)=1/5
DF = log(N/n), where, N is the number of documents and n is the number
of documents a term t has appeared in.
where N is the number of documents and n is the number of documents
a term t has appeared in.
So, IDF(This) = log(2/2) = 0.
Let us compute IDF for the word ‘Messi’.
IDF(Messi) = log(2/1) = 0.301.
Now, let us compare the TF-IDF for a common word ‘This’
and a word ‘Messi’ which seems to be of relevance to Document 1.
TF-IDF(This,Document1) = (1/8) * (0) = 0
TF-IDF(This, Document2) = (1/5) * (0) = 0
TF-IDF(Messi, Document1) = (4/8)*0.301 = 0.15
TF-IDF penalizes the word ‘This’
but assigns greater weight to
‘Messi’.
Co-Occurrence Matrix with a fixed context window
The big idea – Similar words tend to occur together and will have
similar context for example –
“Apple is a fruit. Mango is a fruit.”
Apple and mango tend to have a similar context i.e fruit.
Not preferred in practical
Prediction based Vector
• Continuous Bag of words & Skip-Grams model
https://www.analyticsvidhya.com/blog/2017/06/word-embeddings-count-word2veec/
Input Weight matrix = a word vector
P(word|context) P(context|word)
Intent and Entities
• Intent = topic/domain
• Entities = keywords
“Go home to have the dinner”
Action ObjectLocation
Intent = “Home_activity”
Dialogue Management
VS
Statefulness is the key
• Follow-up
• Pending action
Natural Language Generation
• Fix Response + slot filling + random from a pool
User: Do you know “I’m really quite something”?
Bot: “I’m really quite something” composed by Son Tung-MTP
• Using Neural Network and Language Model
Not recommended
Future of End2End
Data Driven
• Seq2Seq
• Reinforcement
https://aclweb.org/anthology/C18-3006
Tips
• Script Writer
• Personality
• Control the dialogue
• API saves time
• Label Intent and Entities
• Design the flow
• Expandable
• Lots of testing
Applications
Cloud Services
Database
Data warehouse
Web service
Rasa: Open source
conversational AI
PRACTICAL TIME
Use case 1: Health-care
Speech
Google
Virtual
Assistant
Text
(Rest API)
Analyzed
Text
Request
Text
NLP
Intent
Entities
Dialog
Management
Data
Data
Logical Functions
Emotion
Detection
Health-Care
System
Recommendation
System
EmployeesHR staffs
Training
Monitoring
Supporting
Communicate
Company’s
Knowledge
Resource
Chatbot
Use case 2: HR-chatbot
THANK YOU!

More Related Content

Similar to Dialog system understanding

Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
Infrastructures et recommandations pour les Humanités Numériques - Big Data e...Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
Patrice Bellot - Aix-Marseille Université / CNRS (LIS, INS2I)
 

Similar to Dialog system understanding (20)

Pycon ke word vectors
Pycon ke   word vectorsPycon ke   word vectors
Pycon ke word vectors
 
Engineering Intelligent NLP Applications Using Deep Learning – Part 1
Engineering Intelligent NLP Applications Using Deep Learning – Part 1Engineering Intelligent NLP Applications Using Deep Learning – Part 1
Engineering Intelligent NLP Applications Using Deep Learning – Part 1
 
Data Science Your Vacation
Data Science Your VacationData Science Your Vacation
Data Science Your Vacation
 
Document similarity
Document similarityDocument similarity
Document similarity
 
Subword tokenizers
Subword tokenizersSubword tokenizers
Subword tokenizers
 
Data Science Your Vacation
Data Science Your VacationData Science Your Vacation
Data Science Your Vacation
 
Using topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic searchUsing topic modelling frameworks for NLP and semantic search
Using topic modelling frameworks for NLP and semantic search
 
Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)Neural Text Embeddings for Information Retrieval (WSDM 2017)
Neural Text Embeddings for Information Retrieval (WSDM 2017)
 
LSA algorithm
LSA algorithmLSA algorithm
LSA algorithm
 
Search pitb
Search pitbSearch pitb
Search pitb
 
MACHINE-DRIVEN TEXT ANALYSIS
MACHINE-DRIVEN TEXT ANALYSISMACHINE-DRIVEN TEXT ANALYSIS
MACHINE-DRIVEN TEXT ANALYSIS
 
A Panorama of Natural Language Processing
A Panorama of Natural Language ProcessingA Panorama of Natural Language Processing
A Panorama of Natural Language Processing
 
Webinar: Simpler Semantic Search with Solr
Webinar: Simpler Semantic Search with SolrWebinar: Simpler Semantic Search with Solr
Webinar: Simpler Semantic Search with Solr
 
Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
Infrastructures et recommandations pour les Humanités Numériques - Big Data e...Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
Infrastructures et recommandations pour les Humanités Numériques - Big Data e...
 
Vectorization In NLP.pptx
Vectorization In NLP.pptxVectorization In NLP.pptx
Vectorization In NLP.pptx
 
Natural Language Processing: L02 words
Natural Language Processing: L02 wordsNatural Language Processing: L02 words
Natural Language Processing: L02 words
 
Some Information Retrieval Models and Our Experiments for TREC KBA
Some Information Retrieval Models and Our Experiments for TREC KBASome Information Retrieval Models and Our Experiments for TREC KBA
Some Information Retrieval Models and Our Experiments for TREC KBA
 
Deep learning for nlp
Deep learning for nlpDeep learning for nlp
Deep learning for nlp
 
Sequence to sequence (encoder-decoder) learning
Sequence to sequence (encoder-decoder) learningSequence to sequence (encoder-decoder) learning
Sequence to sequence (encoder-decoder) learning
 
Anthiil Inside workshop on NLP
Anthiil Inside workshop on NLPAnthiil Inside workshop on NLP
Anthiil Inside workshop on NLP
 

Recently uploaded

Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
panagenda
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Victor Rentea
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Victor Rentea
 

Recently uploaded (20)

ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
Why Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire businessWhy Teams call analytics are critical to your entire business
Why Teams call analytics are critical to your entire business
 
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdfRising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
Rising Above_ Dubai Floods and the Fortitude of Dubai International Airport.pdf
 
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWEREMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
EMPOWERMENT TECHNOLOGY GRADE 11 QUARTER 2 REVIEWER
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost SavingRepurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
Repurposing LNG terminals for Hydrogen Ammonia: Feasibility and Cost Saving
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf[BuildWithAI] Introduction to Gemini.pdf
[BuildWithAI] Introduction to Gemini.pdf
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
Boost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdfBoost Fertility New Invention Ups Success Rates.pdf
Boost Fertility New Invention Ups Success Rates.pdf
 
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot ModelMcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
Mcleodganj Call Girls 🥰 8617370543 Service Offer VIP Hot Model
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
WSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering DevelopersWSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering Developers
 
Six Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal OntologySix Myths about Ontologies: The Basics of Formal Ontology
Six Myths about Ontologies: The Basics of Formal Ontology
 
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemkeProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
ProductAnonymous-April2024-WinProductDiscovery-MelissaKlemke
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 

Dialog system understanding

  • 1. Dialog System A comprehensive understanding Mr. T
  • 3. Dialog System Natural Language Understanding • Domain identification • User intent detection • Slot Filling Dialogue Management • Dialogue state tracking • Dialogue policy optimization Natural Language Generation Backend Knowledge Providers Speech Recognition Speech Synthesis Utterance Response Trigger Semantic Frame Ask_weather(date=weekends) Text Input “I will go out at weekends, what is the weather?” System Action Request_locationText Response “Where will you go?/”Where you want to ask for the weather this weekends?”
  • 4.
  • 6. A solution for a trigger word system Convoluted Neural Network Recurrent Neural Network Wave Sound Frequency Domain Trigger Word Unknown Output Low level features high level features Patterns extraction Classifier What’s else… Hey Bot/Ok Bot!
  • 8. Speech Recognition Acoustic Dictionary (Pronunciation model) Pre-processing Acoustic Features Speech wave Decoder Acoustic Model WORD PRON (ipa) v ə ˨˩ˀ w e vợ quê “Vợ tôi ở quê rất đẹp” NGRAMs SCORE Vợ quê 2.5 0.7 Language Model v w
  • 10.
  • 11. Ideas of TTS PG & E will file schedules on April, 20 https://web.stanford.edu/~jurafsky/slp3/ed3book.pdf
  • 12. Look up and Concatenate
  • 13. No!
  • 14. Pipeline for Text To Speech Text Input • Sentence Tokenization • Non-standard words • Trained by Machine Learning (SVM, DT, LR) Text Normalization • Disambiguation words • Dictionary Look Up Phonetic Analysis • Names Spelling • Grapheme 2 Phoneme 𝑃 = 𝑎𝑟𝑔𝑚𝑎𝑥 𝑃 (𝑃|𝑆) Prosodic Analysis • Prosody Structure • Prosody Prominence • Tune Voice Output https://web.stanford.edu/~jurafsky/slp3/ed3book.pdf
  • 15. The state-of-the-art Where GAN comes into play.
  • 17. King Queen Man Woman Boy Girl (0.12, 0.23, 0.43) (0.14, 0.57, 0.88) (0.44, 0.90, 0.11) (0.19, 0.23, 0.53) (0.12, 0.65, 0.42) (0.34, 0.44, 0.68) EmbeddingWords Word Embedding (1, 0, 0, 0, 0, 0, 0) (0, 1, 0, 0, 0, 0, 0) (0, 0, 1, 0, 0, 0, 0) ) (0, 0, 0, 0, 1, 0, 0) (0, 0, 0, 0, 0, 1, 0) One-hot Encoding Vector 10 0 1 0 0 0 0 2 4 0 7 0 0 5 0 12 Frequency Based Vector Documents 1 2 3 4 1 2 3 4 Terms Docs Vector word Vector word Vector man woman queen king Predication Based Vector
  • 18. One-hot Encoding Vector 1 2 3 4 5 6 7 8 Co gai 1 0 0 0 0 0 0 0 hot girl 0 1 0 0 0 0 0 0 xinh dep 0 0 1 0 0 0 0 0 truoc day 0 0 0 1 0 0 0 0 la 0 0 0 0 1 0 0 0 mot 0 0 0 0 0 1 0 0 chang trai 0 0 0 0 0 0 1 0 dam my 0 0 0 0 0 0 0 1 each word Gets a 1x 8 vector representation What’s wrong… Corpus: Co gai, hot girl, xinh đep, truoc day, la, mot, chang trai, dam my
  • 19. Custom Encoding Vector Corpus: Co gai, hot girl, xinh đep, truoc day, la, mot, chang trai, dam my nguoi ban chat Thoi gian So dem Nu tinh Co gai 1 0 0 0 1 hot girl 0.7 1 0 0 0.7 xinh dep 0.6 1 0 0 0.5 truoc day 0 0 1 1 0 la 0 0 0 0 0 mot 0 0 0 1 0 chang trai 1 0 0 0 0 dam my 0.7 1 0 0 0 each word Gets a 1x5 vector representation
  • 20. Custom Encoding Vector each word Gets a 1x5 vector representation And better relationship Corpus: Co gai, hot girl, xinh đep, truoc day, la, mot, chang trai, dam my nguoi ban chat Thoi gian So dem Nu tinh Co gai 1 0 0 0 1 hot girl 0.7 1 0 0 0.7 xinh dep 0.6 1 0 0 0.5 truoc day 0 0 1 1 0 la 0 0 0 0 0 mot 0 0 0 1 0 chang trai 1 0 0 0 0 dam my 0.7 1 0 0 0
  • 21. Count Vector Let us understand this using a simple example. • D1: He is a lazy boy. She is also lazy. • D2: Neeraj is a lazy person. Dictionary = [‘He’, ‘She’, ‘lazy’, ‘boy’, ‘Neeraj’, ‘person’] D=2 (# docs), N=6 (# words in the dictionary) He She lazy boy Neeraj person D1 1 1 2 1 0 0 D2 0 0 1 0 1 1 Count Vector matrix M = DXN, vector (“lazy”) = [2, 1]
  • 22. TF-IDF vectorization TF = (Number of times term t appears in a document)/(Number of terms in the document) So, TF(This,Document1) = 1/8 TF(This, Document2)=1/5 DF = log(N/n), where, N is the number of documents and n is the number of documents a term t has appeared in. where N is the number of documents and n is the number of documents a term t has appeared in. So, IDF(This) = log(2/2) = 0. Let us compute IDF for the word ‘Messi’. IDF(Messi) = log(2/1) = 0.301. Now, let us compare the TF-IDF for a common word ‘This’ and a word ‘Messi’ which seems to be of relevance to Document 1. TF-IDF(This,Document1) = (1/8) * (0) = 0 TF-IDF(This, Document2) = (1/5) * (0) = 0 TF-IDF(Messi, Document1) = (4/8)*0.301 = 0.15 TF-IDF penalizes the word ‘This’ but assigns greater weight to ‘Messi’.
  • 23. Co-Occurrence Matrix with a fixed context window The big idea – Similar words tend to occur together and will have similar context for example – “Apple is a fruit. Mango is a fruit.” Apple and mango tend to have a similar context i.e fruit. Not preferred in practical
  • 24. Prediction based Vector • Continuous Bag of words & Skip-Grams model https://www.analyticsvidhya.com/blog/2017/06/word-embeddings-count-word2veec/ Input Weight matrix = a word vector P(word|context) P(context|word)
  • 25. Intent and Entities • Intent = topic/domain • Entities = keywords “Go home to have the dinner” Action ObjectLocation Intent = “Home_activity”
  • 27. Statefulness is the key • Follow-up • Pending action
  • 28. Natural Language Generation • Fix Response + slot filling + random from a pool User: Do you know “I’m really quite something”? Bot: “I’m really quite something” composed by Son Tung-MTP • Using Neural Network and Language Model Not recommended
  • 30. Data Driven • Seq2Seq • Reinforcement https://aclweb.org/anthology/C18-3006
  • 31. Tips • Script Writer • Personality • Control the dialogue • API saves time • Label Intent and Entities • Design the flow • Expandable • Lots of testing
  • 33.
  • 34.
  • 35.
  • 36.
  • 37.
  • 40. Use case 1: Health-care Speech Google Virtual Assistant Text (Rest API) Analyzed Text Request Text NLP Intent Entities Dialog Management Data Data Logical Functions Emotion Detection Health-Care System Recommendation System

Editor's Notes

  1. Data cleaning: transform special characters, number, date, times, into words. Could use Hidden Markov Model or Neural Network to convert because of some ambiguous cases such as spelling years, numbers. 1984. Especially in English, some words are written the same but spelling different. Segmentation, Fragmentation: from text to words, Phone dictionary: contain a reference of word and phone