SlideShare a Scribd company logo
1 of 44
Download to read offline
Incrementality and Anticipation in a
Scaleable Network Model of Linguistic
Competence and Performance
Marshall R. Mayberry, III
Matthew W. Crocker
Saarland University
Goal of Computational Psycholinguistics
• Incremental models of language understanding
• Competence
– Transparent semantic representation
– Wide coverage
– Accuracy
• Performance
– Incremental interpretation and ambiguity resolution
– On-line behavior (e.g., RTs, ERPs, and eye-tracking)
– Adaptation to context
– Robustness
Experience-based Psycholinguistic Models
• Probabilistic (Jurafsky, 1996; Crocker and Brants, 2000)
– Transparent
– Principled
– Scaleable
• Connectionist (Elman, 1990; Rohde, 2002)
– Integrate diverse information sources
– Gradient
– Adaptive
– Robust
Connectionist Language Modelling
• General architecture with simple memory
• Parsimonious account of language processing
• Sentence mapped into a case-role representation
• Incremental development of interpretation with revision
BUT …
• Fixed output with prespecified role assemblies
• Multiple case-role frames? Binding?
• Embedded clauses? Multiple Phrases?
• Cognitively appealing but difficult to scale up.
INSOMNet
Incremental Nonmonotonic Self-Organization of
Meaning Network (Mayberry, 2003)
• Based on Simple Recurrent Network (Elman, 1990)
• Generates explicit compositional representations
• Incremental sentence processing
• Scaled up to Redwoods Treebank/VerbMobil (English)
NEGRA (German)
John
John saw
Object anticipated
John saw Mary
John saw Mary leave
Change in subcategorization frame for saw causes a
different map node to be used.
Scaling Up Without Dumbing Down
• LinGO Redwoods Treebank (Oepen et al, 2002)
– 5000 sentences from VerbMobil (Wahlster, 2000) -
dialogues of face-to-face business travel arrangements
– Rich semantic annotation based on Minimal Recursion
Semantics (Copestake et al, 2001)
– 71.6% recovery of highly-activated dependents
• NEGRA (Skut et al, 1997)
– 20000 sentences of German newspaper text
– Syntactic annotation based on dependency grammar
– 62% accuracy on sentences of less than 15 words
INSOMNet Summary
 
Model of comprehension
 
Wide coverage
 
Explicit semantic representations constructed
incrementally
1. Preferred interpretation
2. Nonmonotonic revision/reanalysis
3. Anticipation
4. Demonstrate observed on-line behavior
5. Adaptation to diverse constraints
6. Modelling Comprehension in Visual Worlds
Visual Worlds
• Attention in visual scene reveals interpretation
– Anticipation (Altmann & Kamide, 1999)
– Rapid integration of diverse linguistic constraints
• Case and Argument structure (Kamide et al, 2003)
• Plausibility (Altmann & Kamide, 1999)
• Prosody (Weber et al, 2003)
• Scene itself can influence interpretation
– Preferred interpretation (Tanenhaus et al, 1995)
– Establish relationships
• Referential Context (Tanenhaus et al, 1995)
• Depicted Actions (Knoeferle et al, 2003)
German
• Flexible constituent order
– Second element must be verb or auxiliary
– Subject or object can occur before or after the verb
• Case-marking often marks subject or object
Die Prinzessin wäscht gleich der Pirat.
(“The princessOBJ washes soon the pirateSUBJ ”)
• Facilitates the study of incremental interpretation
of thematic roles
Linguistic Constraints on Anticipation
• Case and Plausibility (Kamide et al, 2003)
Der Hase frisst gleich den Kohl. / Den Hasen frisst gleich der Fuchs.
(“The hareSUBJ eats soon the cabbageOBJ”) / (“The hareOBJ eats soon the foxSUBJ”)
• Argument Structure (Scheepers et al, submitted)
Der Hase interessiert gleich den Fuchs. / Den Hasen interessiert gleich der Kohl.
(“The hareSUBJ interests soon the foxOBJ”) / (“The hareOBJ interests soon the cabbageSUBJ”)
Influence of Visual Context
Knoeferle, Crocker, Scheepers, and Pickering (to appear)
• SVO: Die Prinzessin malt offensichtlich den Fechter.
(“The princessSUBJ paints soon the fencerOBJ”)
• OVS: Die Prinzessin wäscht offensichtlich der Pirat.
(“The princessOBJ washes soon the pirateSUBJ”)
Empirical Results
• Attention to scene shows
– Scene information rapidly integrated with
utterance to anticipate upcoming arguments.
(Kamide et al, 2003)
• Influence from scene shows
– Depicted actions allow early role
disambiguation even in absence of selectional
restrictions. (Knoeferle et al, to appear)
Modelling Incremental Interpretation
and Anticipation
• Three experiments with complementary properties
– Plausibility
– Depicted actions
– Ambiguity
• Training data consists of all combinations of referents such
that actual experimental sentences are held out.
• Scene information is provided as context for half of
training data. Testing with and without scene.
– Demonstrate anticipatory behavior of empirical studies.
– Time course of ambiguity resolution should show
correspondence to gaze fixations.
Simulations
• Two simulations trained on data from three experiments
• Small-scale
– 16 verbs
– 24 nouns
– 48 non-experimental nouns
– All sentences have matching scene
– 50% trained with and without scene
• Large-scale
– 80 verbs
– 120 nouns
– 240 non-experimental nouns
– Fillers with extra sentence constructions and words
– Non-matching scenes
Network Input/Output
(Linguistic Constraints)
der
Gray nodes denote partially activated arguments
Scene permits more accurate anticipation
der Hase
der Hase interessiert
Scene allows network to recover missing argument
der Hase interessiert gleich
der Hase interessiert gleich den
der Hase interessiert gleich den Fuchs
Network Input/Output
Non-linguistic Constraints
die
die Prinzessin
die Prinzessin wäscht
Scene allows early disambiguation
die Prinzessin wäscht gleich
die Prinzessin wäscht gleich der
die Prinzessin wäscht gleich der Pirat
Small-scale Simulation
• Close to 100% accuracy with and without
scene information
– End of sentence
– Proper selection of plausible arguments and
linking in first experiment
– Early disambiguation using scene in second
• Demonstrates model’s potential to adapt to
and use scene information when available.
Large-Scale Simulation
• Scaling up to complete experiments
– 15% of sentences match empirical behavior.
– Typically agent and patient roles are confused.
• Only a small portion of architectural
parameters have been explored so far.
– How best to encode scene?
– Sizes of scene and output maps?
– Relative strengths of links?
– How much noise to add?
Conclusions
• Incremental Nonmonotonic Sentence Processing
• Semantic comprehension with soft constraints
– Cognitive model of sentence processing
– Parallel, nonmonotonic interpretations
• Scales up to realistic language
• Shows promise for modelling fine-grained
adaptive behavior to context
Linguistic Constraints on Anticipation
• Case and Plausibility
(Kamide et al, 2003)
– Der Hase frisst gleich den
Kohl.
– Den Hasen frisst gleich
der Fuchs.
• Argument Structure
(Scheepers et al, submitted)
– Der Hase interessiert
gleich den Fuchs.
– Den Hasen interessiert
gleich der Kohl.
Sequence Processor
• SRN
– Processes sentence incrementally
– Generates explicit semantic representation
• SARDNet helps retain long-distance dependencies
Semantic Frame Encoder/Decoder
• SRN output is map of encoded semantic frames
• Dedicated links for specific components
– Verb frames have WORD, TENSE, AGT, and PAT links
– Noun frames have WORD and GENDER links
Frame Selector
• Frame Modulator Map
– One-to-one correspondence with Frame Nodes
– Gives graded frame selection
• Target Indicator Map
– One-to-one correspondence with Frame Nodes
– Binary selection
Semantic Self-Organization
• Frame Modulator Map
– Provides graded Frame Node selection
– Trained through back-propagation
– Target is the Target Indicator Map
• Target Indicator Map
– Only used during training
– Self-organized using compressed frames
– Selects Frame Nodes to encode semantic frames
INSOMNet
Putting It All Together
• Sequence Processor reads in sentence incrementally
• Frame En/Decoder represents semantic frames
• Frame Selector models graded frame selection
Human Sentence Processing
• Sentence interpretation is constructed incrementally
• Many factors involved (e.g., experience, plausibility)
• Ambiguity is pervasive
– How is preferred interpretation determined?
Conceptual/Semantic Representation

More Related Content

What's hot

Constructive Hybrid Logics
Constructive Hybrid LogicsConstructive Hybrid Logics
Constructive Hybrid LogicsValeria de Paiva
 
Langacker's cognitive grammar
Langacker's cognitive grammarLangacker's cognitive grammar
Langacker's cognitive grammarJOy Verzosa
 
Cognitive grammar
Cognitive grammarCognitive grammar
Cognitive grammarsafwan aziz
 
Cognitive Grammar Lesson
Cognitive Grammar LessonCognitive Grammar Lesson
Cognitive Grammar LessonMelinda50
 
Constructive Description Logics 2006
Constructive Description Logics 2006Constructive Description Logics 2006
Constructive Description Logics 2006Valeria de Paiva
 
Compositional Distributional Models of Meaning
Compositional Distributional Models of MeaningCompositional Distributional Models of Meaning
Compositional Distributional Models of MeaningDimitrios Kartsaklis
 
Cognitive Linguistics: The Case Of Find
Cognitive Linguistics: The Case Of FindCognitive Linguistics: The Case Of Find
Cognitive Linguistics: The Case Of FindJESSIE GRACE RUBRICO
 
Sentiment Analysis in Twitter with Lightweight Discourse Analysis
Sentiment Analysis in Twitter with Lightweight Discourse AnalysisSentiment Analysis in Twitter with Lightweight Discourse Analysis
Sentiment Analysis in Twitter with Lightweight Discourse AnalysisSubhabrata Mukherjee
 
Minimalist program
Minimalist programMinimalist program
Minimalist programRabbiaAzam
 
Supporting language learners with the
Supporting language learners with theSupporting language learners with the
Supporting language learners with theijaia
 
Text, genre and discourse shifts in translation
Text, genre and discourse shifts in translationText, genre and discourse shifts in translation
Text, genre and discourse shifts in translationAisyah Revolusioner
 
Authorship attribution
Authorship attributionAuthorship attribution
Authorship attributionReza Ramezani
 
The Evolution of Morphological Agreement
The Evolution of Morphological AgreementThe Evolution of Morphological Agreement
The Evolution of Morphological AgreementRichard Littauer
 
Semantic discourse analysis
Semantic discourse analysisSemantic discourse analysis
Semantic discourse analysisblessedkkr
 
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...Association for Computational Linguistics
 

What's hot (20)

Constructive Hybrid Logics
Constructive Hybrid LogicsConstructive Hybrid Logics
Constructive Hybrid Logics
 
Langacker's cognitive grammar
Langacker's cognitive grammarLangacker's cognitive grammar
Langacker's cognitive grammar
 
Cognitive grammar
Cognitive grammarCognitive grammar
Cognitive grammar
 
Cognitive Grammar Lesson
Cognitive Grammar LessonCognitive Grammar Lesson
Cognitive Grammar Lesson
 
Constructive Description Logics 2006
Constructive Description Logics 2006Constructive Description Logics 2006
Constructive Description Logics 2006
 
Minimalist program
Minimalist programMinimalist program
Minimalist program
 
Compositional Distributional Models of Meaning
Compositional Distributional Models of MeaningCompositional Distributional Models of Meaning
Compositional Distributional Models of Meaning
 
The Cognitive Perspective in Teaching Chinese measure words
The Cognitive Perspective in Teaching Chinese measure words The Cognitive Perspective in Teaching Chinese measure words
The Cognitive Perspective in Teaching Chinese measure words
 
Cognitive Linguistics: The Case Of Find
Cognitive Linguistics: The Case Of FindCognitive Linguistics: The Case Of Find
Cognitive Linguistics: The Case Of Find
 
Sentiment Analysis in Twitter with Lightweight Discourse Analysis
Sentiment Analysis in Twitter with Lightweight Discourse AnalysisSentiment Analysis in Twitter with Lightweight Discourse Analysis
Sentiment Analysis in Twitter with Lightweight Discourse Analysis
 
Lfg and gpsg
Lfg and gpsgLfg and gpsg
Lfg and gpsg
 
Minimalist program
Minimalist programMinimalist program
Minimalist program
 
Supporting language learners with the
Supporting language learners with theSupporting language learners with the
Supporting language learners with the
 
Nlp (1)
Nlp (1)Nlp (1)
Nlp (1)
 
Text, genre and discourse shifts in translation
Text, genre and discourse shifts in translationText, genre and discourse shifts in translation
Text, genre and discourse shifts in translation
 
Authorship attribution
Authorship attributionAuthorship attribution
Authorship attribution
 
The Evolution of Morphological Agreement
The Evolution of Morphological AgreementThe Evolution of Morphological Agreement
The Evolution of Morphological Agreement
 
Semantic discourse analysis
Semantic discourse analysisSemantic discourse analysis
Semantic discourse analysis
 
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...
Chris Dyer - 2017 - CoNLL Invited Talk: Should Neural Network Architecture Re...
 
AI Lesson 09
AI Lesson 09AI Lesson 09
AI Lesson 09
 

Similar to Incrementality

A Neural Probabilistic Language Model.pptx
A Neural Probabilistic Language Model.pptxA Neural Probabilistic Language Model.pptx
A Neural Probabilistic Language Model.pptxRama Irsheidat
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language ProcessingToine Bogers
 
dialogue act modeling for automatic tagging and recognition
 dialogue act modeling for automatic tagging and recognition dialogue act modeling for automatic tagging and recognition
dialogue act modeling for automatic tagging and recognitionVipul Munot
 
Colloquium talk on modal sense classification using a convolutional neural ne...
Colloquium talk on modal sense classification using a convolutional neural ne...Colloquium talk on modal sense classification using a convolutional neural ne...
Colloquium talk on modal sense classification using a convolutional neural ne...Ana Marasović
 
Natural Language Processing (NLP)
Natural Language Processing (NLP)Natural Language Processing (NLP)
Natural Language Processing (NLP)Yuriy Guts
 
Turkish language modeling using BERT
Turkish language modeling using BERTTurkish language modeling using BERT
Turkish language modeling using BERTAbdurrahimDerric
 
State-of-the-Art Text Classification using Deep Contextual Word Representations
State-of-the-Art Text Classification using Deep Contextual Word RepresentationsState-of-the-Art Text Classification using Deep Contextual Word Representations
State-of-the-Art Text Classification using Deep Contextual Word RepresentationsAusaf Ahmed
 
Gathering Lexical Linked Data and Knowledge Patterns from FrameNet
Gathering Lexical Linked Data and Knowledge Patterns from FrameNetGathering Lexical Linked Data and Knowledge Patterns from FrameNet
Gathering Lexical Linked Data and Knowledge Patterns from FrameNetAndrea Nuzzolese
 
NLP introduced and in 47 slides Lecture 1.ppt
NLP introduced and in 47 slides Lecture 1.pptNLP introduced and in 47 slides Lecture 1.ppt
NLP introduced and in 47 slides Lecture 1.pptOlusolaTop
 
Introduction to Natural Language Processing
Introduction to Natural Language ProcessingIntroduction to Natural Language Processing
Introduction to Natural Language ProcessingPranav Gupta
 
WORD RECOGNITION MASLP
WORD RECOGNITION MASLPWORD RECOGNITION MASLP
WORD RECOGNITION MASLPHimaniBansal15
 
Engineering Intelligent NLP Applications Using Deep Learning – Part 2
Engineering Intelligent NLP Applications Using Deep Learning – Part 2 Engineering Intelligent NLP Applications Using Deep Learning – Part 2
Engineering Intelligent NLP Applications Using Deep Learning – Part 2 Saurabh Kaushik
 
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)Zachary S. Brown
 
Challenges in transfer learning in nlp
Challenges in transfer learning in nlpChallenges in transfer learning in nlp
Challenges in transfer learning in nlpLaraOlmosCamarena
 

Similar to Incrementality (20)

A Neural Probabilistic Language Model.pptx
A Neural Probabilistic Language Model.pptxA Neural Probabilistic Language Model.pptx
A Neural Probabilistic Language Model.pptx
 
Natural Language Processing
Natural Language ProcessingNatural Language Processing
Natural Language Processing
 
dialogue act modeling for automatic tagging and recognition
 dialogue act modeling for automatic tagging and recognition dialogue act modeling for automatic tagging and recognition
dialogue act modeling for automatic tagging and recognition
 
Colloquium talk on modal sense classification using a convolutional neural ne...
Colloquium talk on modal sense classification using a convolutional neural ne...Colloquium talk on modal sense classification using a convolutional neural ne...
Colloquium talk on modal sense classification using a convolutional neural ne...
 
Natural Language Processing (NLP)
Natural Language Processing (NLP)Natural Language Processing (NLP)
Natural Language Processing (NLP)
 
Turkish language modeling using BERT
Turkish language modeling using BERTTurkish language modeling using BERT
Turkish language modeling using BERT
 
What is word2vec?
What is word2vec?What is word2vec?
What is word2vec?
 
State-of-the-Art Text Classification using Deep Contextual Word Representations
State-of-the-Art Text Classification using Deep Contextual Word RepresentationsState-of-the-Art Text Classification using Deep Contextual Word Representations
State-of-the-Art Text Classification using Deep Contextual Word Representations
 
Deep learning for nlp
Deep learning for nlpDeep learning for nlp
Deep learning for nlp
 
Sentence Processing by Muhammad Saleem.pptx
Sentence Processing by Muhammad Saleem.pptxSentence Processing by Muhammad Saleem.pptx
Sentence Processing by Muhammad Saleem.pptx
 
Gathering Lexical Linked Data and Knowledge Patterns from FrameNet
Gathering Lexical Linked Data and Knowledge Patterns from FrameNetGathering Lexical Linked Data and Knowledge Patterns from FrameNet
Gathering Lexical Linked Data and Knowledge Patterns from FrameNet
 
haenelt.ppt
haenelt.ppthaenelt.ppt
haenelt.ppt
 
NLP introduced and in 47 slides Lecture 1.ppt
NLP introduced and in 47 slides Lecture 1.pptNLP introduced and in 47 slides Lecture 1.ppt
NLP introduced and in 47 slides Lecture 1.ppt
 
Semeval Deep Learning In Semantic Similarity
Semeval Deep Learning In Semantic SimilaritySemeval Deep Learning In Semantic Similarity
Semeval Deep Learning In Semantic Similarity
 
Introduction to Natural Language Processing
Introduction to Natural Language ProcessingIntroduction to Natural Language Processing
Introduction to Natural Language Processing
 
WORD RECOGNITION MASLP
WORD RECOGNITION MASLPWORD RECOGNITION MASLP
WORD RECOGNITION MASLP
 
Engineering Intelligent NLP Applications Using Deep Learning – Part 2
Engineering Intelligent NLP Applications Using Deep Learning – Part 2 Engineering Intelligent NLP Applications Using Deep Learning – Part 2
Engineering Intelligent NLP Applications Using Deep Learning – Part 2
 
NLP_KASHK: Introduction
NLP_KASHK: Introduction NLP_KASHK: Introduction
NLP_KASHK: Introduction
 
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
Deep Learning and Modern Natural Language Processing (AnacondaCon2019)
 
Challenges in transfer learning in nlp
Challenges in transfer learning in nlpChallenges in transfer learning in nlp
Challenges in transfer learning in nlp
 

Incrementality

  • 1. Incrementality and Anticipation in a Scaleable Network Model of Linguistic Competence and Performance Marshall R. Mayberry, III Matthew W. Crocker Saarland University
  • 2. Goal of Computational Psycholinguistics • Incremental models of language understanding • Competence – Transparent semantic representation – Wide coverage – Accuracy • Performance – Incremental interpretation and ambiguity resolution – On-line behavior (e.g., RTs, ERPs, and eye-tracking) – Adaptation to context – Robustness
  • 3. Experience-based Psycholinguistic Models • Probabilistic (Jurafsky, 1996; Crocker and Brants, 2000) – Transparent – Principled – Scaleable • Connectionist (Elman, 1990; Rohde, 2002) – Integrate diverse information sources – Gradient – Adaptive – Robust
  • 4. Connectionist Language Modelling • General architecture with simple memory • Parsimonious account of language processing • Sentence mapped into a case-role representation • Incremental development of interpretation with revision
  • 5. BUT … • Fixed output with prespecified role assemblies • Multiple case-role frames? Binding? • Embedded clauses? Multiple Phrases? • Cognitively appealing but difficult to scale up.
  • 6. INSOMNet Incremental Nonmonotonic Self-Organization of Meaning Network (Mayberry, 2003) • Based on Simple Recurrent Network (Elman, 1990) • Generates explicit compositional representations • Incremental sentence processing • Scaled up to Redwoods Treebank/VerbMobil (English) NEGRA (German)
  • 10. John saw Mary leave Change in subcategorization frame for saw causes a different map node to be used.
  • 11. Scaling Up Without Dumbing Down • LinGO Redwoods Treebank (Oepen et al, 2002) – 5000 sentences from VerbMobil (Wahlster, 2000) - dialogues of face-to-face business travel arrangements – Rich semantic annotation based on Minimal Recursion Semantics (Copestake et al, 2001) – 71.6% recovery of highly-activated dependents • NEGRA (Skut et al, 1997) – 20000 sentences of German newspaper text – Syntactic annotation based on dependency grammar – 62% accuracy on sentences of less than 15 words
  • 12. INSOMNet Summary   Model of comprehension   Wide coverage   Explicit semantic representations constructed incrementally 1. Preferred interpretation 2. Nonmonotonic revision/reanalysis 3. Anticipation 4. Demonstrate observed on-line behavior 5. Adaptation to diverse constraints 6. Modelling Comprehension in Visual Worlds
  • 13. Visual Worlds • Attention in visual scene reveals interpretation – Anticipation (Altmann & Kamide, 1999) – Rapid integration of diverse linguistic constraints • Case and Argument structure (Kamide et al, 2003) • Plausibility (Altmann & Kamide, 1999) • Prosody (Weber et al, 2003) • Scene itself can influence interpretation – Preferred interpretation (Tanenhaus et al, 1995) – Establish relationships • Referential Context (Tanenhaus et al, 1995) • Depicted Actions (Knoeferle et al, 2003)
  • 14. German • Flexible constituent order – Second element must be verb or auxiliary – Subject or object can occur before or after the verb • Case-marking often marks subject or object Die Prinzessin wäscht gleich der Pirat. (“The princessOBJ washes soon the pirateSUBJ ”) • Facilitates the study of incremental interpretation of thematic roles
  • 15. Linguistic Constraints on Anticipation • Case and Plausibility (Kamide et al, 2003) Der Hase frisst gleich den Kohl. / Den Hasen frisst gleich der Fuchs. (“The hareSUBJ eats soon the cabbageOBJ”) / (“The hareOBJ eats soon the foxSUBJ”) • Argument Structure (Scheepers et al, submitted) Der Hase interessiert gleich den Fuchs. / Den Hasen interessiert gleich der Kohl. (“The hareSUBJ interests soon the foxOBJ”) / (“The hareOBJ interests soon the cabbageSUBJ”)
  • 16. Influence of Visual Context Knoeferle, Crocker, Scheepers, and Pickering (to appear) • SVO: Die Prinzessin malt offensichtlich den Fechter. (“The princessSUBJ paints soon the fencerOBJ”) • OVS: Die Prinzessin wäscht offensichtlich der Pirat. (“The princessOBJ washes soon the pirateSUBJ”)
  • 17. Empirical Results • Attention to scene shows – Scene information rapidly integrated with utterance to anticipate upcoming arguments. (Kamide et al, 2003) • Influence from scene shows – Depicted actions allow early role disambiguation even in absence of selectional restrictions. (Knoeferle et al, to appear)
  • 18. Modelling Incremental Interpretation and Anticipation • Three experiments with complementary properties – Plausibility – Depicted actions – Ambiguity • Training data consists of all combinations of referents such that actual experimental sentences are held out. • Scene information is provided as context for half of training data. Testing with and without scene. – Demonstrate anticipatory behavior of empirical studies. – Time course of ambiguity resolution should show correspondence to gaze fixations.
  • 19. Simulations • Two simulations trained on data from three experiments • Small-scale – 16 verbs – 24 nouns – 48 non-experimental nouns – All sentences have matching scene – 50% trained with and without scene • Large-scale – 80 verbs – 120 nouns – 240 non-experimental nouns – Fillers with extra sentence constructions and words – Non-matching scenes
  • 21. der Gray nodes denote partially activated arguments Scene permits more accurate anticipation
  • 23. der Hase interessiert Scene allows network to recover missing argument
  • 25. der Hase interessiert gleich den
  • 26. der Hase interessiert gleich den Fuchs
  • 28. die
  • 30. die Prinzessin wäscht Scene allows early disambiguation
  • 33. die Prinzessin wäscht gleich der Pirat
  • 34. Small-scale Simulation • Close to 100% accuracy with and without scene information – End of sentence – Proper selection of plausible arguments and linking in first experiment – Early disambiguation using scene in second • Demonstrates model’s potential to adapt to and use scene information when available.
  • 35. Large-Scale Simulation • Scaling up to complete experiments – 15% of sentences match empirical behavior. – Typically agent and patient roles are confused. • Only a small portion of architectural parameters have been explored so far. – How best to encode scene? – Sizes of scene and output maps? – Relative strengths of links? – How much noise to add?
  • 36. Conclusions • Incremental Nonmonotonic Sentence Processing • Semantic comprehension with soft constraints – Cognitive model of sentence processing – Parallel, nonmonotonic interpretations • Scales up to realistic language • Shows promise for modelling fine-grained adaptive behavior to context
  • 37. Linguistic Constraints on Anticipation • Case and Plausibility (Kamide et al, 2003) – Der Hase frisst gleich den Kohl. – Den Hasen frisst gleich der Fuchs. • Argument Structure (Scheepers et al, submitted) – Der Hase interessiert gleich den Fuchs. – Den Hasen interessiert gleich der Kohl.
  • 38. Sequence Processor • SRN – Processes sentence incrementally – Generates explicit semantic representation • SARDNet helps retain long-distance dependencies
  • 39. Semantic Frame Encoder/Decoder • SRN output is map of encoded semantic frames • Dedicated links for specific components – Verb frames have WORD, TENSE, AGT, and PAT links – Noun frames have WORD and GENDER links
  • 40. Frame Selector • Frame Modulator Map – One-to-one correspondence with Frame Nodes – Gives graded frame selection • Target Indicator Map – One-to-one correspondence with Frame Nodes – Binary selection
  • 41. Semantic Self-Organization • Frame Modulator Map – Provides graded Frame Node selection – Trained through back-propagation – Target is the Target Indicator Map • Target Indicator Map – Only used during training – Self-organized using compressed frames – Selects Frame Nodes to encode semantic frames
  • 42. INSOMNet Putting It All Together • Sequence Processor reads in sentence incrementally • Frame En/Decoder represents semantic frames • Frame Selector models graded frame selection
  • 43. Human Sentence Processing • Sentence interpretation is constructed incrementally • Many factors involved (e.g., experience, plausibility) • Ambiguity is pervasive – How is preferred interpretation determined?