SlideShare a Scribd company logo
1 of 90
Download to read offline
Squashing
Computational Linguistics
Noah A. Smith
Paul G. Allen School of Computer Science & Engineering
University of Washington
Seattle, USA
@nlpnoah
Research supported in part by: NSF, DARPA DEFT, DARPA CWC, Facebook, Google, Samsung, University of Washington.
data
Applications of NLP in 2017
• Conversation, IE, MT, QA, summarization, text categorization
Applications of NLP in 2017
• Conversation, IE, MT, QA, summarization, text categorization
• Machine-in-the-loop tools for (human) authors
Elizabeth Clark
Collaborate with an NLP
model through an
“exquisite corpse”
storytelling game
Chenhao Tan
Revise your
message with help
from NLP
tremoloop.com
Applications of NLP in 2017
• Conversation, IE, MT, QA, summarization, text categorization
• Machine-in-the-loop tools for (human) authors
• Analysis tools for measuring social phenomena
Lucy Lin
Sensationalism in
science news
bit.ly/sensational-news
… bookmark this survey!
Dallas Card
Track ideas, propositions,
frames in discourse over
time
data
?
Squash
Squash Networks
• Parameterized differentiable functions composed out of simpler
parameterized differentiable functions, some nonlinear
Squash Networks
• Parameterized differentiable functions composed out of simpler
parameterized differentiable functions, some nonlinear
From Jack (2010),Dynamic
System Modeling and
Control, goo.gl/pGvJPS
*Yes, rectified linear units (relus) are only half-squash; hat-tip Martha White.
Squash Networks
• Parameterized differentiable functions composed out of simpler
parameterized differentiable functions, some nonlinear
• Estimate parameters using Leibniz (1676)
From existentialcomics.com
Who wants an all-squash diet?
wow
very Curcurbita
much festive
many dropout
Linguistic Structure Prediction
input (text)
output (structure)
Linguistic Structure Prediction
input (text)
output (structure)
sequences,
trees,
graphs, …
“gold” output
Linguistic Structure Prediction
input (text)
output (structure)
sequences,
trees,
graphs, …
“gold” output
Linguistic Structure Prediction
input representation
input (text)
output (structure)
sequences,
trees,
graphs, …
“gold” output
Linguistic Structure Prediction
input representation
input (text)
output (structure)
clusters,lexicons,
embeddings, …
sequences,
trees,
graphs, …
“gold” output
Linguistic Structure Prediction
input representation
input (text)
output (structure)
training objective
clusters,lexicons,
embeddings, …
sequences,
trees,
graphs, …
“gold” output
Linguistic Structure Prediction
input representation
input (text)
output (structure)
training objective
clusters,lexicons,
embeddings, …
sequences,
trees,
graphs, …
probabilistic,
cost-aware,…
“gold” output
Linguistic Structure Prediction
input representation
input (text)
part representations
output (structure)
training objective
clusters,lexicons,
embeddings, …
sequences,
trees,
graphs, …
probabilistic,
cost-aware,…
“gold” output
Linguistic Structure Prediction
input representation
input (text)
part representations
output (structure)
training objective
clusters,lexicons,
embeddings, …
segments/spans, arcs,
graph fragments, …
sequences,
trees,
graphs, …
probabilistic,
cost-aware,…
“gold” output
Linguistic Structure Prediction
input representation
input (text)
part representations
output (structure)
training objective
“gold” output
Linguistic Structure Prediction
input representation
input (text)
part representations
output (structure)
training objective
error definitions
& weights
regularization
annotation
conventions
& theory
constraints &
independence
assumptions
dataselection
“task”
Inductive Bias
• What does your learning
algorithm assume?
• How will it choose among good
predictive functions?
See also:
No Free Lunch Theorem
(Mitchell, 1980;Wolpert, 1996)
data
bias
Three New Models
• Parsing sentences into
predicate-argument structures
• Fillmore frames
• Semantic dependency graphs
• Language models that
dynamically track entities
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
Original story on Slate.com: http://goo.gl/Hp89tD
Frame-Semantic Analysis
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
FrameNet: https://framenet.icsi.berkeley.edu
Frame-Semantic Analysis
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
FrameNet: https://framenet.icsi.berkeley.edu
Frame-Semantic Analysis
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
cognizer: Democrats
topic: why … Clinton
FrameNet: https://framenet.icsi.berkeley.edu
Frame-Semantic Analysis
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
cognizer: Democrats
topic: why … Clinton
explanation: why
degree: so much
content: of Clinton
experiencer: ?
helper: Stephanopoulos … Carville
goal: to put over
time: in 1992
benefited_party: ?
FrameNet: https://framenet.icsi.berkeley.edu
landmark event: Democrats … Clinton
trajector event: they… 1992
entity: so … Clinton
degree: so
mass: resentment of Clinton
time: When … Clinton
required situation: they … to look … 1992
time: When … Clinton
cognizer agent: they
ground: much … 1992
sought entity: ?
topic: about … 1992
trajector event: the Big Lie … over
landmark period: 1992
FrameNet: https://framenet.icsi.berkeley.edu
brood, consider,
contemplate, deliberate, …
appraise,
assess,
evaluate, …
commit to
memory, learn,
memorize, …
agonize, fret, fuss,
lose sleep, …
translate bracket,
categorize,
class, classify
Frame-Semantic Analysis
When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much
further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992.
cognizer: Democrats
topic: why … Clinton
explanation: why
degree: so much
content: of Clinton
experiencer: ?
helper: Stephanopoulos … Carville
goal: to put over
time: in 1992
benefited_party: ?
FrameNet: https://framenet.icsi.berkeley.edu
landmark event: Democrats … Clinton
trajector event: they… 1992
entity: so … Clinton
degree: so
mass: resentment of Clinton
time: When … Clinton
required situation: they … to look … 1992
time: When … Clinton
cognizer agent: they
ground: much … 1992
sought entity: ?
topic: about … 1992
trajector event: the Big Lie … over
landmark period: 1992
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … words + frame
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
biLSTM
(contextualized
word vectors)
words + frame
parts: segments
up to length d
scored by
another biLSTM,
with labels
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
…
biLSTM
(contextualized
word vectors)
words + frame
parts: segments
up to length d
scored by
another biLSTM,
with labels
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
…
biLSTM
(contextualized
word vectors)
words + frame
output: covering sequence of nonoverlapping segments
Segmental RNN
(Lingpeng Kong, Chris Dyer, N.A.S., ICLR 2016)
biLSTM
(contextualized
word vectors)
input sequence
parts: segments
up to length d
scored by
another biLSTM,
with labels
training objective:
log loss
output: covering sequence of nonoverlapping segments, recovered in O(Ldn); see Sarawagi & Cohen, 2004
…
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
∅
∅
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
Inference via dynamic programming in O(Ldn)
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
Open-SESAME
(Swabha Swayamdipta, Sam Thomson,
Chris Dyer, N.A.S., arXiv:1706.09528)
biLSTM
(contextualized
word vectors)
words + frame
parts: segments
with role labels,
scored by
another biLSTM
training objective:
recall-oriented
softmax margin
(Gimpel et al.,
2010)
output: labeled
argument spans
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer: Democrats
topic: why there is so much resentment of Clinton
…
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
syntax features?
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
yes no yes yesno
Penn Treebank (Marcus et al., 1993)
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer topic ∅
wonderCogitation
∅
wonderCogitation
wonderCogitation
wonderCogitation
∅
main task
scaffold task
yes no yes yesno
Multitask Representation Learning
(Caruana, 1997)
“gold” output
input representation
input (text)
part representations
output (structure)
training objective
“gold” output
input representation
input (text)
part representations
output (structure)
training objective
main task: find and
label semantic
arguments
scaffold task: predict
syntactic constituents
shared
training datasets need not overlap
output structures need not be consistent
ours
65
66
67
68
69
70
71
72
SEMAFOR 1.0
(Das et al.,
2014)
SEMAFOR 2.0
(Kshirsagar et
al.,2015)
Open-SESAME … with
syntactic
scaffold
… with syntax
features
Framat (Roth,
2016)
FitzGerald et
al. (2015)
single model ensemble
F1 on frame-semantic parsing (frames & arguments), FrameNet 1.5 test set.
open-source systems
ours
65
66
67
68
69
70
71
72
SEMAFOR 1.0
(Das et al.,
2014)
SEMAFOR 2.0
(Kshirsagar et
al.,2015)
Open-SESAME … with
syntactic
scaffold
… with syntax
features
Framat (Roth,
2016)
FitzGerald et
al. (2015)
single model ensemble
F1 on frame-semantic parsing (frames & arguments), FrameNet 1.5 test set.
open-source systems
biLSTM
(contextualized
word vectors)
words + frame
training objective:
recall-oriented
softmax margin
(Gimpel et al.,
2010)
output: labeled
argument spans
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer: Democrats
topic: why there is so much resentment of Clinton
…
segments
get scores
Bias?
parts: segments
with role labels,
scored by
another biLSTM
biLSTM
(contextualized
word vectors)
words + frame
training objective:
recall-oriented
softmax margin
(Gimpel et al.,
2010)
output: labeled
argument spans
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer: Democrats
topic: why there is so much resentment of Clinton
…
segments
get scores
syntactic scaffold
Bias?
parts: segments
with role labels,
scored by
another biLSTM
biLSTM
(contextualized
word vectors)
words + frame
training objective:
recall-oriented
softmax margin
(Gimpel et al.,
2010)
output: labeled
argument spans
When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
cognizer: Democrats
topic: why there is so much resentment of Clinton
…
segments
get scores
recall-oriented
cost
syntactic scaffold
Bias?
parts: segments
with role labels,
scored by
another biLSTM
Semantic Dependency Graphs
(DELPH-IN minimal recursion semantics-derived representation; “DM”)
Oepen et al. (SemEval 2014; 2015),see also http://sdp.delph-in.net
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Democrats wonder
arg1
tanh(C[hwonder; hDemocrats] + b) · arg1
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
When Democrats Democrats wonder why is
is resentment
so much much resentment
resentment of of Clinton
arg2 arg1
arg1
comp_so
arg1
arg1 arg1arg1
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
When Democrats
wonder why
Democrats wonder why is
is resentment
so much much resentment
resentment of of Clinton
arg2 arg1
arg1
comp_so
arg1
arg1
arg1 arg1arg1
wonder there
arg1
wonder is
arg1
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Inference via AD3
(alternating directions dual decomposition; Martins et al., 2014)
When Democrats
wonder why
Democrats wonder why is
is resentment
so much much resentment
resentment of of Clinton
arg2 arg1
arg1
comp_so
arg1
arg1
arg1 arg1arg1
wonder there
arg1
wonder is
arg1
Neurboparser
(Hao Peng, Sam Thomson, N.A.S., ACL 2017)
biLSTM
(contextualized
word vectors)
words
parts: labeled
bilexical
dependencies
training objective:
structured hinge
loss
output: labeled
semantic
dependency
graph with
constraints
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
…
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Three Formalisms, Three Separate Parsers
Democrats wonder
arg1
formalism 3 (PSD)
formalism 2 (PAS)
formalism 1 (DM)
Democrats wonder
arg1
Democrats wonder
act
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Shared Input Representations
(Daumé, 2007)
Democrats wonder
arg1
shared across all
formalism 3 (PSD)
formalism 2 (PAS)
formalism 1 (DM)
Democrats wonder
arg1
Democrats wonder
act
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Cross-Task Parts
Democrats wonder
arg1
Democrats wonder
arg1
Democrats wonder
act
shared across all
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
Both: Shared Input Representations
& Cross-Task Parts
Democrats wonder
arg1
shared across all
formalism 3 (PSD)
formalism 2 (PAS)
formalism 1 (DM)
Democrats wonder
arg1
Democrats wonder
act
Multitask Learning: Many Possibilities
• Shared input representations, parts? Which parts?
• Joint decoding?
• Overlapping training data?
• Scaffold tasks?
“gold” output “gold” output“gold” output
formalism 1 (DM) formalism 2 (PAS) formalism 3 (PSD)
ours
80
81
82
83
84
85
86
87
88
Du et al.,2015 Almeida &
Martins, 2015
(no syntax)
Almeida &
Martins, 2015
(syntax)
Neurboparser … with shared
input
representations
and cross-task
parts
WSJ (in-domain) Brown (out-of-domain)
F1 averaged on three semantic dependency parsing formalisms, SemEval 2015 test set.
50
60
70
80
90
100
DM PAS PSD
WSJ (in-domain) Brown (out-of-domain)
Neurboparser F1 on three semantic dependency parsing formalisms, SemEval 2015 test set.
good enough?
biLSTM
(contextualized
word vectors)
words
training objective:
structured hinge
loss
output: labeled
semantic
dependency
graph with
constraints
When Democrats wonderwhy there is so much resentment of Clinton, they don’t need …
…
Bias? cross-formalism
sharing
parts: labeled
bilexical
dependencies
Text
Text ≠ Sentences
larger context
Generative Language Models
history next word
p(W | history)
Generative Language Models
history next word
p(W | history)
When Democrats wonderwhy there is so much resentment of
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of
entity 1
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of Clinton,
1. new entity with new
vector
2. mention word will be
“Clinton”
entity 1
entity 2
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of Clinton,
entity 1
entity 2
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of Clinton, they
1. coreferent of entity 1
(previously known as
“Democrats”)
2. mention word will be
“they”
3. embedding of entity 1
will be updated
entity 1
entity 2
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of Clinton, they
entity 1
entity 2
Entity Language Model
(Yangfeng Ji, Chenhao Tan, Sebastian Martschat,
Yejin Choi, N.A.S., EMNLP 2017)
When Democrats wonderwhy there is so much resentment of Clinton, they don’t
1. not part of an entity
mention
2. “don’t”
entity 1
entity 2
128
132
136
140
5-gram LM RNN LM entity
language
model
Perplexity on CoNLL 2012 test set. CoNLL 2012 coreference evaluation.
50
55
60
65
70
75
Martschat and
Strube (2015)
reranked with entity
LM
CoNLL MUC F1 B3 CEAF
25
35
45
55
65
75
always
new
shallow
features
Modi et
al. (2017)
entity LM human
Accuracy: InScript.
history next word
p(W | history)
entities
Bias?
Bias in the Future?
• Linguistic scaffold tasks.
Bias in the Future?
• Linguistic scaffold tasks.
• Language is by and about people.
Bias in the Future?
• Linguistic scaffold tasks.
• Language is by and about people.
• NLP is needed when texts are costly to read.
Bias in the Future?
• Linguistic scaffold tasks.
• Language is by and about people.
• NLP is needed when texts are costly to read.
• Polyglot learning.
data
bias
Thank you!

More Related Content

Similar to Noah A Smith - 2017 - Invited Keynote: Squashing Computational Linguistics

Tim Estes - Generating dynamic social networks from large scale unstructured ...
Tim Estes - Generating dynamic social networks from large scale unstructured ...Tim Estes - Generating dynamic social networks from large scale unstructured ...
Tim Estes - Generating dynamic social networks from large scale unstructured ...Digital Reasoning
 
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...Andre Freitas
 
Data Science Workshop - day 1
Data Science Workshop - day 1Data Science Workshop - day 1
Data Science Workshop - day 1Aseel Addawood
 
Building AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsBuilding AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsAndre Freitas
 
Ppt Pak kasim
Ppt Pak kasimPpt Pak kasim
Ppt Pak kasimdarmigita
 
What data scientists really do, according to 50 data scientists
What data scientists really do, according to 50 data scientistsWhat data scientists really do, according to 50 data scientists
What data scientists really do, according to 50 data scientistsHugo Bowne-Anderson
 
Why Watson Won: A cognitive perspective
Why Watson Won: A cognitive perspectiveWhy Watson Won: A cognitive perspective
Why Watson Won: A cognitive perspectiveJames Hendler
 
Fairness, Transparency, and Privacy in AI @ LinkedIn
Fairness, Transparency, and Privacy in AI @ LinkedInFairness, Transparency, and Privacy in AI @ LinkedIn
Fairness, Transparency, and Privacy in AI @ LinkedInKrishnaram Kenthapadi
 
Machine learning technology for publishing industry / Buchmesse 2018
Machine learning technology for publishing industry / Buchmesse 2018Machine learning technology for publishing industry / Buchmesse 2018
Machine learning technology for publishing industry / Buchmesse 2018Maxim Kublitski
 
Silk data - machine learning
Silk data - machine learning Silk data - machine learning
Silk data - machine learning SaltoDigitale
 
BSSML17 - Topic Models
BSSML17 - Topic ModelsBSSML17 - Topic Models
BSSML17 - Topic ModelsBigML, Inc
 
Data Science Workshop - day 2
Data Science Workshop - day 2Data Science Workshop - day 2
Data Science Workshop - day 2Aseel Addawood
 
Frontiers of Computational Journalism week 2 - Text Analysis
Frontiers of Computational Journalism week 2 - Text AnalysisFrontiers of Computational Journalism week 2 - Text Analysis
Frontiers of Computational Journalism week 2 - Text AnalysisJonathan Stray
 
Session 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectSession 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectbodaceacat
 
Session 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectSession 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectSara-Jayne Terp
 
Modern text mining – understanding a million comments in 60 minutes
Modern text mining – understanding a million comments in 60 minutesModern text mining – understanding a million comments in 60 minutes
Modern text mining – understanding a million comments in 60 minutesZOLLHOF - Tech Incubator
 

Similar to Noah A Smith - 2017 - Invited Keynote: Squashing Computational Linguistics (20)

Tim Estes - Generating dynamic social networks from large scale unstructured ...
Tim Estes - Generating dynamic social networks from large scale unstructured ...Tim Estes - Generating dynamic social networks from large scale unstructured ...
Tim Estes - Generating dynamic social networks from large scale unstructured ...
 
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...
Talking to your Data: Natural Language Interfaces for a schema-less world (Ke...
 
Data Science Workshop - day 1
Data Science Workshop - day 1Data Science Workshop - day 1
Data Science Workshop - day 1
 
Building AI Applications using Knowledge Graphs
Building AI Applications using Knowledge GraphsBuilding AI Applications using Knowledge Graphs
Building AI Applications using Knowledge Graphs
 
Ppt Pak kasim
Ppt Pak kasimPpt Pak kasim
Ppt Pak kasim
 
BD-ACA Week8a
BD-ACA Week8aBD-ACA Week8a
BD-ACA Week8a
 
What data scientists really do, according to 50 data scientists
What data scientists really do, according to 50 data scientistsWhat data scientists really do, according to 50 data scientists
What data scientists really do, according to 50 data scientists
 
Why Watson Won: A cognitive perspective
Why Watson Won: A cognitive perspectiveWhy Watson Won: A cognitive perspective
Why Watson Won: A cognitive perspective
 
Fairness, Transparency, and Privacy in AI @ LinkedIn
Fairness, Transparency, and Privacy in AI @ LinkedInFairness, Transparency, and Privacy in AI @ LinkedIn
Fairness, Transparency, and Privacy in AI @ LinkedIn
 
Machine learning technology for publishing industry / Buchmesse 2018
Machine learning technology for publishing industry / Buchmesse 2018Machine learning technology for publishing industry / Buchmesse 2018
Machine learning technology for publishing industry / Buchmesse 2018
 
Silk data - machine learning
Silk data - machine learning Silk data - machine learning
Silk data - machine learning
 
BSSML17 - Topic Models
BSSML17 - Topic ModelsBSSML17 - Topic Models
BSSML17 - Topic Models
 
Data Science Workshop - day 2
Data Science Workshop - day 2Data Science Workshop - day 2
Data Science Workshop - day 2
 
BDACA - Lecture4
BDACA - Lecture4BDACA - Lecture4
BDACA - Lecture4
 
Frontiers of Computational Journalism week 2 - Text Analysis
Frontiers of Computational Journalism week 2 - Text AnalysisFrontiers of Computational Journalism week 2 - Text Analysis
Frontiers of Computational Journalism week 2 - Text Analysis
 
Session 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectSession 01 designing and scoping a data science project
Session 01 designing and scoping a data science project
 
Session 01 designing and scoping a data science project
Session 01 designing and scoping a data science projectSession 01 designing and scoping a data science project
Session 01 designing and scoping a data science project
 
Modern text mining – understanding a million comments in 60 minutes
Modern text mining – understanding a million comments in 60 minutesModern text mining – understanding a million comments in 60 minutes
Modern text mining – understanding a million comments in 60 minutes
 
Gate
GateGate
Gate
 
Open University - TU100 Day school 1
Open University - TU100 Day school 1Open University - TU100 Day school 1
Open University - TU100 Day school 1
 

More from Association for Computational Linguistics

Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...
Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...
Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...Association for Computational Linguistics
 
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...Association for Computational Linguistics
 
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsDaniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsAssociation for Computational Linguistics
 
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsDaniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsAssociation for Computational Linguistics
 
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...Association for Computational Linguistics
 
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Association for Computational Linguistics
 
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Association for Computational Linguistics
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...Association for Computational Linguistics
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Association for Computational Linguistics
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Association for Computational Linguistics
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopAssociation for Computational Linguistics
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Association for Computational Linguistics
 

More from Association for Computational Linguistics (20)

Muis - 2016 - Weak Semi-Markov CRFs for NP Chunking in Informal Text
Muis - 2016 - Weak Semi-Markov CRFs for NP Chunking in Informal TextMuis - 2016 - Weak Semi-Markov CRFs for NP Chunking in Informal Text
Muis - 2016 - Weak Semi-Markov CRFs for NP Chunking in Informal Text
 
Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...
Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...
Castro - 2018 - A High Coverage Method for Automatic False Friends Detection ...
 
Castro - 2018 - A Crowd-Annotated Spanish Corpus for Humour Analysis
Castro - 2018 - A Crowd-Annotated Spanish Corpus for Humour AnalysisCastro - 2018 - A Crowd-Annotated Spanish Corpus for Humour Analysis
Castro - 2018 - A Crowd-Annotated Spanish Corpus for Humour Analysis
 
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...
Muthu Kumar Chandrasekaran - 2018 - Countering Position Bias in Instructor In...
 
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsDaniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
 
Elior Sulem - 2018 - Semantic Structural Evaluation for Text Simplification
Elior Sulem - 2018 - Semantic Structural Evaluation for Text SimplificationElior Sulem - 2018 - Semantic Structural Evaluation for Text Simplification
Elior Sulem - 2018 - Semantic Structural Evaluation for Text Simplification
 
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future DirectionsDaniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
Daniel Gildea - 2018 - The ACL Anthology: Current State and Future Directions
 
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...
Wenqiang Lei - 2018 - Sequicity: Simplifying Task-oriented Dialogue Systems w...
 
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
Matthew Marge - 2017 - Exploring Variation of Natural Human Commands to a Rob...
 
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
Venkatesh Duppada - 2017 - SeerNet at EmoInt-2017: Tweet Emotion Intensity Es...
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
John Richardson - 2015 - KyotoEBMT System Description for the 2nd Workshop on...
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
 
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
Zhongyuan Zhu - 2015 - Evaluating Neural Machine Translation in English-Japan...
 
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
Hyoung-Gyu Lee - 2015 - NAVER Machine Translation System for WAT 2015
 
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 WorkshopSatoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
Satoshi Sonoh - 2015 - Toshiba MT System Description for the WAT2015 Workshop
 
Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015Chenchen Ding - 2015 - NICT at WAT 2015
Chenchen Ding - 2015 - NICT at WAT 2015
 
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
Graham Neubig - 2015 - Neural Reranking Improves Subjective Quality of Machin...
 

Recently uploaded

SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentationcamerronhm
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsMebane Rash
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17Celine George
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsTechSoup
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxAreebaZafar22
 
Third Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxThird Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxAmita Gupta
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxRamakrishna Reddy Bijjam
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxheathfieldcps1
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfPoh-Sun Goh
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.christianmathematics
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Association for Project Management
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxnegromaestrong
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxAmanpreet Kaur
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibitjbellavia9
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...Nguyen Thanh Tu Collection
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxVishalSingh1417
 

Recently uploaded (20)

SOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning PresentationSOC 101 Demonstration of Learning Presentation
SOC 101 Demonstration of Learning Presentation
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Third Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptxThird Battle of Panipat detailed notes.pptx
Third Battle of Panipat detailed notes.pptx
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Micro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdfMicro-Scholarship, What it is, How can it help me.pdf
Micro-Scholarship, What it is, How can it help me.pdf
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
Sociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning ExhibitSociology 101 Demonstration of Learning Exhibit
Sociology 101 Demonstration of Learning Exhibit
 
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
TỔNG ÔN TẬP THI VÀO LỚP 10 MÔN TIẾNG ANH NĂM HỌC 2023 - 2024 CÓ ĐÁP ÁN (NGỮ Â...
 
Asian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptxAsian American Pacific Islander Month DDSD 2024.pptx
Asian American Pacific Islander Month DDSD 2024.pptx
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 

Noah A Smith - 2017 - Invited Keynote: Squashing Computational Linguistics

  • 1. Squashing Computational Linguistics Noah A. Smith Paul G. Allen School of Computer Science & Engineering University of Washington Seattle, USA @nlpnoah Research supported in part by: NSF, DARPA DEFT, DARPA CWC, Facebook, Google, Samsung, University of Washington.
  • 3. Applications of NLP in 2017 • Conversation, IE, MT, QA, summarization, text categorization
  • 4. Applications of NLP in 2017 • Conversation, IE, MT, QA, summarization, text categorization • Machine-in-the-loop tools for (human) authors Elizabeth Clark Collaborate with an NLP model through an “exquisite corpse” storytelling game Chenhao Tan Revise your message with help from NLP tremoloop.com
  • 5. Applications of NLP in 2017 • Conversation, IE, MT, QA, summarization, text categorization • Machine-in-the-loop tools for (human) authors • Analysis tools for measuring social phenomena Lucy Lin Sensationalism in science news bit.ly/sensational-news … bookmark this survey! Dallas Card Track ideas, propositions, frames in discourse over time
  • 8. Squash Networks • Parameterized differentiable functions composed out of simpler parameterized differentiable functions, some nonlinear
  • 9. Squash Networks • Parameterized differentiable functions composed out of simpler parameterized differentiable functions, some nonlinear From Jack (2010),Dynamic System Modeling and Control, goo.gl/pGvJPS *Yes, rectified linear units (relus) are only half-squash; hat-tip Martha White.
  • 10. Squash Networks • Parameterized differentiable functions composed out of simpler parameterized differentiable functions, some nonlinear • Estimate parameters using Leibniz (1676) From existentialcomics.com
  • 11. Who wants an all-squash diet? wow very Curcurbita much festive many dropout
  • 12.
  • 13. Linguistic Structure Prediction input (text) output (structure)
  • 14. Linguistic Structure Prediction input (text) output (structure) sequences, trees, graphs, …
  • 15. “gold” output Linguistic Structure Prediction input (text) output (structure) sequences, trees, graphs, …
  • 16. “gold” output Linguistic Structure Prediction input representation input (text) output (structure) sequences, trees, graphs, …
  • 17. “gold” output Linguistic Structure Prediction input representation input (text) output (structure) clusters,lexicons, embeddings, … sequences, trees, graphs, …
  • 18. “gold” output Linguistic Structure Prediction input representation input (text) output (structure) training objective clusters,lexicons, embeddings, … sequences, trees, graphs, …
  • 19. “gold” output Linguistic Structure Prediction input representation input (text) output (structure) training objective clusters,lexicons, embeddings, … sequences, trees, graphs, … probabilistic, cost-aware,…
  • 20. “gold” output Linguistic Structure Prediction input representation input (text) part representations output (structure) training objective clusters,lexicons, embeddings, … sequences, trees, graphs, … probabilistic, cost-aware,…
  • 21. “gold” output Linguistic Structure Prediction input representation input (text) part representations output (structure) training objective clusters,lexicons, embeddings, … segments/spans, arcs, graph fragments, … sequences, trees, graphs, … probabilistic, cost-aware,…
  • 22. “gold” output Linguistic Structure Prediction input representation input (text) part representations output (structure) training objective
  • 23. “gold” output Linguistic Structure Prediction input representation input (text) part representations output (structure) training objective error definitions & weights regularization annotation conventions & theory constraints & independence assumptions dataselection “task”
  • 24. Inductive Bias • What does your learning algorithm assume? • How will it choose among good predictive functions? See also: No Free Lunch Theorem (Mitchell, 1980;Wolpert, 1996)
  • 26. Three New Models • Parsing sentences into predicate-argument structures • Fillmore frames • Semantic dependency graphs • Language models that dynamically track entities
  • 27. When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. Original story on Slate.com: http://goo.gl/Hp89tD
  • 28. Frame-Semantic Analysis When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. FrameNet: https://framenet.icsi.berkeley.edu
  • 29. Frame-Semantic Analysis When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. FrameNet: https://framenet.icsi.berkeley.edu
  • 30. Frame-Semantic Analysis When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. cognizer: Democrats topic: why … Clinton FrameNet: https://framenet.icsi.berkeley.edu
  • 31. Frame-Semantic Analysis When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. cognizer: Democrats topic: why … Clinton explanation: why degree: so much content: of Clinton experiencer: ? helper: Stephanopoulos … Carville goal: to put over time: in 1992 benefited_party: ? FrameNet: https://framenet.icsi.berkeley.edu landmark event: Democrats … Clinton trajector event: they… 1992 entity: so … Clinton degree: so mass: resentment of Clinton time: When … Clinton required situation: they … to look … 1992 time: When … Clinton cognizer agent: they ground: much … 1992 sought entity: ? topic: about … 1992 trajector event: the Big Lie … over landmark period: 1992
  • 32. FrameNet: https://framenet.icsi.berkeley.edu brood, consider, contemplate, deliberate, … appraise, assess, evaluate, … commit to memory, learn, memorize, … agonize, fret, fuss, lose sleep, … translate bracket, categorize, class, classify
  • 33. Frame-Semantic Analysis When Democrats wonder why there is so much resentment of Clinton, they don’t need to look much further than the Big Lie about philandering that Stephanopoulos, Carville helped to put over in 1992. cognizer: Democrats topic: why … Clinton explanation: why degree: so much content: of Clinton experiencer: ? helper: Stephanopoulos … Carville goal: to put over time: in 1992 benefited_party: ? FrameNet: https://framenet.icsi.berkeley.edu landmark event: Democrats … Clinton trajector event: they… 1992 entity: so … Clinton degree: so mass: resentment of Clinton time: When … Clinton required situation: they … to look … 1992 time: When … Clinton cognizer agent: they ground: much … 1992 sought entity: ? topic: about … 1992 trajector event: the Big Lie … over landmark period: 1992
  • 34. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … words + frame
  • 35. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … biLSTM (contextualized word vectors) words + frame
  • 36. parts: segments up to length d scored by another biLSTM, with labels When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … … biLSTM (contextualized word vectors) words + frame
  • 37. parts: segments up to length d scored by another biLSTM, with labels When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … … biLSTM (contextualized word vectors) words + frame output: covering sequence of nonoverlapping segments
  • 38. Segmental RNN (Lingpeng Kong, Chris Dyer, N.A.S., ICLR 2016) biLSTM (contextualized word vectors) input sequence parts: segments up to length d scored by another biLSTM, with labels training objective: log loss output: covering sequence of nonoverlapping segments, recovered in O(Ldn); see Sarawagi & Cohen, 2004 …
  • 39. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
  • 40. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need …
  • 41. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ ∅ ∅
  • 42. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 43. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 44. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 45. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 46. Inference via dynamic programming in O(Ldn) When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 47. Open-SESAME (Swabha Swayamdipta, Sam Thomson, Chris Dyer, N.A.S., arXiv:1706.09528) biLSTM (contextualized word vectors) words + frame parts: segments with role labels, scored by another biLSTM training objective: recall-oriented softmax margin (Gimpel et al., 2010) output: labeled argument spans When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer: Democrats topic: why there is so much resentment of Clinton …
  • 48. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅
  • 49. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅ syntax features?
  • 50. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … yes no yes yesno Penn Treebank (Marcus et al., 1993)
  • 51. When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer topic ∅ wonderCogitation ∅ wonderCogitation wonderCogitation wonderCogitation ∅ main task scaffold task yes no yes yesno
  • 52. Multitask Representation Learning (Caruana, 1997) “gold” output input representation input (text) part representations output (structure) training objective “gold” output input representation input (text) part representations output (structure) training objective main task: find and label semantic arguments scaffold task: predict syntactic constituents shared training datasets need not overlap output structures need not be consistent
  • 53. ours 65 66 67 68 69 70 71 72 SEMAFOR 1.0 (Das et al., 2014) SEMAFOR 2.0 (Kshirsagar et al.,2015) Open-SESAME … with syntactic scaffold … with syntax features Framat (Roth, 2016) FitzGerald et al. (2015) single model ensemble F1 on frame-semantic parsing (frames & arguments), FrameNet 1.5 test set. open-source systems
  • 54. ours 65 66 67 68 69 70 71 72 SEMAFOR 1.0 (Das et al., 2014) SEMAFOR 2.0 (Kshirsagar et al.,2015) Open-SESAME … with syntactic scaffold … with syntax features Framat (Roth, 2016) FitzGerald et al. (2015) single model ensemble F1 on frame-semantic parsing (frames & arguments), FrameNet 1.5 test set. open-source systems
  • 55. biLSTM (contextualized word vectors) words + frame training objective: recall-oriented softmax margin (Gimpel et al., 2010) output: labeled argument spans When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer: Democrats topic: why there is so much resentment of Clinton … segments get scores Bias? parts: segments with role labels, scored by another biLSTM
  • 56. biLSTM (contextualized word vectors) words + frame training objective: recall-oriented softmax margin (Gimpel et al., 2010) output: labeled argument spans When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer: Democrats topic: why there is so much resentment of Clinton … segments get scores syntactic scaffold Bias? parts: segments with role labels, scored by another biLSTM
  • 57. biLSTM (contextualized word vectors) words + frame training objective: recall-oriented softmax margin (Gimpel et al., 2010) output: labeled argument spans When Democrats wonderCogitation why there is so much resentment of Clinton, they don’t need … cognizer: Democrats topic: why there is so much resentment of Clinton … segments get scores recall-oriented cost syntactic scaffold Bias? parts: segments with role labels, scored by another biLSTM
  • 58. Semantic Dependency Graphs (DELPH-IN minimal recursion semantics-derived representation; “DM”) Oepen et al. (SemEval 2014; 2015),see also http://sdp.delph-in.net
  • 59. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Democrats wonder arg1 tanh(C[hwonder; hDemocrats] + b) · arg1
  • 60. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … When Democrats Democrats wonder why is is resentment so much much resentment resentment of of Clinton arg2 arg1 arg1 comp_so arg1 arg1 arg1arg1
  • 61. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … When Democrats wonder why Democrats wonder why is is resentment so much much resentment resentment of of Clinton arg2 arg1 arg1 comp_so arg1 arg1 arg1 arg1arg1 wonder there arg1 wonder is arg1
  • 62. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Inference via AD3 (alternating directions dual decomposition; Martins et al., 2014) When Democrats wonder why Democrats wonder why is is resentment so much much resentment resentment of of Clinton arg2 arg1 arg1 comp_so arg1 arg1 arg1 arg1arg1 wonder there arg1 wonder is arg1
  • 63. Neurboparser (Hao Peng, Sam Thomson, N.A.S., ACL 2017) biLSTM (contextualized word vectors) words parts: labeled bilexical dependencies training objective: structured hinge loss output: labeled semantic dependency graph with constraints When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … …
  • 64. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Three Formalisms, Three Separate Parsers Democrats wonder arg1 formalism 3 (PSD) formalism 2 (PAS) formalism 1 (DM) Democrats wonder arg1 Democrats wonder act
  • 65. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Shared Input Representations (Daumé, 2007) Democrats wonder arg1 shared across all formalism 3 (PSD) formalism 2 (PAS) formalism 1 (DM) Democrats wonder arg1 Democrats wonder act
  • 66. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Cross-Task Parts Democrats wonder arg1 Democrats wonder arg1 Democrats wonder act shared across all
  • 67. When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … Both: Shared Input Representations & Cross-Task Parts Democrats wonder arg1 shared across all formalism 3 (PSD) formalism 2 (PAS) formalism 1 (DM) Democrats wonder arg1 Democrats wonder act
  • 68. Multitask Learning: Many Possibilities • Shared input representations, parts? Which parts? • Joint decoding? • Overlapping training data? • Scaffold tasks? “gold” output “gold” output“gold” output formalism 1 (DM) formalism 2 (PAS) formalism 3 (PSD)
  • 69. ours 80 81 82 83 84 85 86 87 88 Du et al.,2015 Almeida & Martins, 2015 (no syntax) Almeida & Martins, 2015 (syntax) Neurboparser … with shared input representations and cross-task parts WSJ (in-domain) Brown (out-of-domain) F1 averaged on three semantic dependency parsing formalisms, SemEval 2015 test set.
  • 70. 50 60 70 80 90 100 DM PAS PSD WSJ (in-domain) Brown (out-of-domain) Neurboparser F1 on three semantic dependency parsing formalisms, SemEval 2015 test set. good enough?
  • 71. biLSTM (contextualized word vectors) words training objective: structured hinge loss output: labeled semantic dependency graph with constraints When Democrats wonderwhy there is so much resentment of Clinton, they don’t need … … Bias? cross-formalism sharing parts: labeled bilexical dependencies
  • 72. Text
  • 74. Generative Language Models history next word p(W | history)
  • 75. Generative Language Models history next word p(W | history)
  • 76. When Democrats wonderwhy there is so much resentment of Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017)
  • 77. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of entity 1
  • 78. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of Clinton, 1. new entity with new vector 2. mention word will be “Clinton” entity 1 entity 2
  • 79. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of Clinton, entity 1 entity 2
  • 80. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of Clinton, they 1. coreferent of entity 1 (previously known as “Democrats”) 2. mention word will be “they” 3. embedding of entity 1 will be updated entity 1 entity 2
  • 81. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of Clinton, they entity 1 entity 2
  • 82. Entity Language Model (Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, N.A.S., EMNLP 2017) When Democrats wonderwhy there is so much resentment of Clinton, they don’t 1. not part of an entity mention 2. “don’t” entity 1 entity 2
  • 83. 128 132 136 140 5-gram LM RNN LM entity language model Perplexity on CoNLL 2012 test set. CoNLL 2012 coreference evaluation. 50 55 60 65 70 75 Martschat and Strube (2015) reranked with entity LM CoNLL MUC F1 B3 CEAF 25 35 45 55 65 75 always new shallow features Modi et al. (2017) entity LM human Accuracy: InScript.
  • 84. history next word p(W | history) entities Bias?
  • 85. Bias in the Future? • Linguistic scaffold tasks.
  • 86. Bias in the Future? • Linguistic scaffold tasks. • Language is by and about people.
  • 87. Bias in the Future? • Linguistic scaffold tasks. • Language is by and about people. • NLP is needed when texts are costly to read.
  • 88. Bias in the Future? • Linguistic scaffold tasks. • Language is by and about people. • NLP is needed when texts are costly to read. • Polyglot learning.