This document summarizes Yejin Choi's research on modeling commonsense knowledge in frame semantics. The research aims to model both physical knowledge implied by language (e.g. verbs) and inherent physical properties of objects. It does this by developing a model called VerbPhysics that represents commonsense knowledge across five physical dimensions for verbs and object relationships. The model is trained on text using loopy belief propagation with embeddings to jointly learn verb frames and object-object relations without external knowledge or embodiment. The goal is to reverse engineer commonsense physical knowledge from language alone.
The document discusses using a hybrid approach combining predictability and the Dempster-Shafer method of modeling randomness and fuzziness. It provides motivation for considering external factors beyond what is in training data. Basic definitions of randomness and fuzziness are given. The Dempster-Shafer theory allows combining evidence from independent sources and handling contradictions. Several datasets are tested using a modified Dempster-Shafer approach to potentially increase predictive performance over solely using historic data.
1. The document discusses universal quantification and quantifiers. Universal quantification refers to statements that are true for all variables, while quantifiers are words like "some" or "all" that refer to quantities.
2. It explains that a universally quantified statement is of the form "For all x, P(x) is true" and is defined to be true if P(x) is true for every x, and false if P(x) is false for at least one x.
3. When the universe of discourse can be listed as x1, x2, etc., a universal statement is the same as the conjunction P(x1) and P(x2) and etc., because this
Distributional semantics models can provide probabilistic information about word meanings based on contextual clues. An agent can use distributional evidence to update its probabilistic information state about unknown words. In experiments, distributional similarity evidence increased the probability that properties of a known word (like "crocodile") also apply to an unknown word ("alligator"). Higher similarities led to more confident inferences. Combining multiple pieces of evidence further increased probabilities, allowing an agent to infer an unknown word refers to an animal based on its similarities to both "crocodile" and "trout".
The document discusses various natural language processing (NLP) tasks including named entity recognition, entity linking, question answering, sentiment analysis, dependency parsing, and semantic role labeling. It provides examples and explanations of how each task can be approached, common challenges, and relevant datasets and resources.
This document provides an overview of some basic probability concepts and probability distributions relevant to biostatistics. It defines key probability terms like classical probability, relative frequency probability, subjective probability, and the three axioms of probability. It also explains how to calculate probabilities of single and joint events, conditional probability, and introduces important probability distributions like the binomial, Poisson, and normal distributions. Worked examples are provided to illustrate probability calculations.
Slides to present our paper
Sommerauer, Pia, and Antske Fokkens. "Firearms and Tigers are Dangerous, Kitchen Knives and Zebras are Not: Testing whether Word Embeddings Can Tell." In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pp. 276-286. 2018.
This document introduces the Word-Sensibility model for predicting and responding to language. It discusses:
1) Representing words as "responsive units" that map to topics like space, time, mental states, and locomotion. These units respond dynamically based on their relationships.
2) Organizing words into "word-topics" with four dimensions called "quadranyms" representing sources and targets of responses. Linking units forms "scripts" and stacking units makes hierarchical "layers."
3) Using an example of understanding the sentence "I will know as soon as I walk through the door" by mapping words to responsive topics and having the topics calibrate responses.
4) Representing words
Redevelop 2019 - Debugging our biases and intuition in software developmentDave Hulbert
We program algorithms with shortcuts known as heuristics. These allow us to get a good enough answer to a problem with less CPU and memory usage. Brains attempt to take shortcuts too, using intuition and biases to figure things out with less thinking or knowledge. Heuristics are valuable but they're not perfect. We can evaluate best and worst cases for code but how do we do the same with our own decision making process?
In this talk, we'll go through the ups and downs of heuristics and biases that exist in a developer's world. We'll look at ways to reduce any resulting fallacies, whilst still taking advantage of the performance improvement.
The document discusses using a hybrid approach combining predictability and the Dempster-Shafer method of modeling randomness and fuzziness. It provides motivation for considering external factors beyond what is in training data. Basic definitions of randomness and fuzziness are given. The Dempster-Shafer theory allows combining evidence from independent sources and handling contradictions. Several datasets are tested using a modified Dempster-Shafer approach to potentially increase predictive performance over solely using historic data.
1. The document discusses universal quantification and quantifiers. Universal quantification refers to statements that are true for all variables, while quantifiers are words like "some" or "all" that refer to quantities.
2. It explains that a universally quantified statement is of the form "For all x, P(x) is true" and is defined to be true if P(x) is true for every x, and false if P(x) is false for at least one x.
3. When the universe of discourse can be listed as x1, x2, etc., a universal statement is the same as the conjunction P(x1) and P(x2) and etc., because this
Distributional semantics models can provide probabilistic information about word meanings based on contextual clues. An agent can use distributional evidence to update its probabilistic information state about unknown words. In experiments, distributional similarity evidence increased the probability that properties of a known word (like "crocodile") also apply to an unknown word ("alligator"). Higher similarities led to more confident inferences. Combining multiple pieces of evidence further increased probabilities, allowing an agent to infer an unknown word refers to an animal based on its similarities to both "crocodile" and "trout".
The document discusses various natural language processing (NLP) tasks including named entity recognition, entity linking, question answering, sentiment analysis, dependency parsing, and semantic role labeling. It provides examples and explanations of how each task can be approached, common challenges, and relevant datasets and resources.
This document provides an overview of some basic probability concepts and probability distributions relevant to biostatistics. It defines key probability terms like classical probability, relative frequency probability, subjective probability, and the three axioms of probability. It also explains how to calculate probabilities of single and joint events, conditional probability, and introduces important probability distributions like the binomial, Poisson, and normal distributions. Worked examples are provided to illustrate probability calculations.
Slides to present our paper
Sommerauer, Pia, and Antske Fokkens. "Firearms and Tigers are Dangerous, Kitchen Knives and Zebras are Not: Testing whether Word Embeddings Can Tell." In Proceedings of the 2018 EMNLP Workshop BlackboxNLP: Analyzing and Interpreting Neural Networks for NLP, pp. 276-286. 2018.
This document introduces the Word-Sensibility model for predicting and responding to language. It discusses:
1) Representing words as "responsive units" that map to topics like space, time, mental states, and locomotion. These units respond dynamically based on their relationships.
2) Organizing words into "word-topics" with four dimensions called "quadranyms" representing sources and targets of responses. Linking units forms "scripts" and stacking units makes hierarchical "layers."
3) Using an example of understanding the sentence "I will know as soon as I walk through the door" by mapping words to responsive topics and having the topics calibrate responses.
4) Representing words
Redevelop 2019 - Debugging our biases and intuition in software developmentDave Hulbert
We program algorithms with shortcuts known as heuristics. These allow us to get a good enough answer to a problem with less CPU and memory usage. Brains attempt to take shortcuts too, using intuition and biases to figure things out with less thinking or knowledge. Heuristics are valuable but they're not perfect. We can evaluate best and worst cases for code but how do we do the same with our own decision making process?
In this talk, we'll go through the ups and downs of heuristics and biases that exist in a developer's world. We'll look at ways to reduce any resulting fallacies, whilst still taking advantage of the performance improvement.
FIID 2015 Guadalajara, Mexico
Graphic organizers for vocabulary learning
Several types and purposes of graphic organizers have been proposed over the years. In this presentation, several explore types and purposes of graphic organizers preserving the themes of using these in a scalable manner across the grade levels and between various disciplines through a literacy lens (e.g., sciences, mathematics, physical education, literary study, etc.). The primary focus of this session is to demonstrate how vocabulary, especially, can be more effectively taught using graphic organizers. Participants will learn to create graphic organizers to address specific learning purposes rather than relying solely on generic organizers. Five different organizers will be introduced, and participants will be able to experience the use of the organizer. Digital versions, paper versions prepared by the teacher, and manipulative versions created by the students will be explored.
This document provides an introduction to action research and how it can be applied. It discusses that action research involves collaboration between researchers and clients to diagnose and solve problems. The action research cycle involves planning a change, acting on it, observing the results, and reflecting to plan further changes.
Two case studies are described: one where researchers worked with students to help them better evaluate information, and another where researchers collaborated with teachers across Europe to design an online course based on the teachers' desired learning goals. Finally, the document provides advice on when and how to appropriately use action research and links to additional resources.
Research methodology hypothesis constructionSami Nighaoui
This document provides guidance on constructing a research hypothesis. It discusses identifying dependent and independent variables related to the topic of illegal immigrants' rights in the US. It suggests drafting an initial research question and theoretical statement explaining the relationship between variables. The document emphasizes that a hypothesis formally defines the causal relationship and must be falsifiable. It provides an example null hypothesis that level of education does not impact views on immigrant rights, and that rejecting this would support the alternative hypothesis.
These slides can share you what is particle swam optimization which is one of the technique that comes under the soft computing and artificial intelligence which is an emerging technology now and trending technology that every corporate company is working over this and hear I have given an example of that on particle swam optimization.
A friendly introduction to using action research in library and information environments. Presented at LILAC 2018 by Sam Aston, Geoff Walton and Emma Coonan.
The document provides an introduction to first-order logic and its advantages over propositional logic. It discusses elements of first-order logic like objects, relations, functions, equality, variables, quantifiers, and converting English statements to formal logic. Examples are provided to illustrate modeling relationships between objects and reasoning about multiple objects at once using quantifiers. The limitations of propositional logic in modeling statements involving objects and relations are demonstrated.
This document discusses logic and propositional logic. It covers the following topics:
- The history and applications of logic.
- Different types of statements and their grammar.
- Propositional logic including symbols, connectives, truth tables, and semantics.
- Quantifiers, universal and existential quantification, and properties of quantifiers.
- Normal forms such as disjunctive normal form and conjunctive normal form.
- Inference rules and the principle of mathematical induction, illustrated with examples.
This document introduces the natural sciences as an area of knowledge. It defines natural sciences as knowledge of observable objects and processes in nature, such as biology and physics, as distinguished from abstract sciences like mathematics. It describes the scientific method as involving observation, hypothesis formation, prediction, and experimental testing to confirm or falsify hypotheses. Personal knowledge and imagination play a role in scientific discovery alongside shared experimentation and observation. Questions are raised about whether scientific knowledge can be considered absolutely true or reliable given its tentative nature and reliance on indirect observation tools.
This document discusses Bayesian inference and Bayes' theorem. It provides examples of how Bayesian inference can be used to calculate conditional probabilities based on new evidence or data. Specifically, it discusses using Bayesian inference to determine the probability a die is "good" based on rolling results, and the probability a woman has breast cancer based on positive or negative mammogram results.
This paper contributes a noun phrase-annotated SMS corpus and proposes a weak semi-Markov CRF model for noun phrase chunking in informal text. The weak semi-CRF model improves training speed over linear-CRF and semi-CRF models while maintaining similar accuracy. Experiments on the SMS corpus show the weak semi-CRF achieves F1 scores comparable to other models but trains faster, especially with larger training data sizes.
This document presents a new method for automatically detecting false friends between Spanish and Portuguese using word embeddings. The method builds word vector spaces for each language using word2vec, finds a linear transformation between the spaces, and measures vector distances to classify word pairs as cognates or false friends. In experiments on a dataset of 710 word pairs, the method achieved state-of-the-art accuracy of 77.28% and high coverage of 97.91%, outperforming previous work. Future work will explore using different word embeddings and fine-grained classifications of partial false friends.
This document describes a Spanish language corpus for humor analysis that was created through crowd-sourcing annotations. Over 27,000 tweets were collected from humorous accounts and annotated through a web interface. The corpus contains over 100,000 annotations of the tweets' humor and funniness. Inter-annotator agreement was higher for this corpus than a previous Spanish humor corpus. The dataset will help analyze subjectivity in humor and was used in a shared task on humor classification and funniness prediction.
This document discusses position bias in instructor interventions in MOOC discussion forums. It finds that instructors are more likely to intervene in threads that appear higher on the discussion forum user interface due to their recent activity. To address this, it proposes a debiased classifier that weights examples based on their propensity for intervention. It finds this approach identifies intervention opportunities that were overlooked due to position bias. The debiased classifier outperforms a standard classifier on several metrics, demonstrating it can better predict unbiased intervention needs.
The document summarizes the history and current state of the ACL Anthology, a repository of publications from ACL-sponsored conferences. It discusses how the Anthology was established in 2001 and is now maintained by volunteers, containing over 45,000 papers. The presentation calls for community involvement to help future-proof the Anthology through efforts like migrating its infrastructure and improving documentation. It also proposes hosting the Anthology on the main ACL website and recruiting a new editor.
The document presents SAMSA, a new automatic evaluation measure for structural text simplification. SAMSA uses semantic parsing to measure the preservation of semantic structures and relations between an original text and its simplified version. It correlates significantly better with human judgments of meaning preservation and structural simplicity than prior reference-based metrics. SAMSA is the first evaluation method designed specifically for structural simplification operations like sentence splitting.
(1) Sequicity is a framework that simplifies task-oriented dialogue systems using single sequence-to-sequence architectures.
(2) It formalizes dialogues as sequences of belief spans and responses and decodes them in two stages: generating a belief span followed by a response.
(3) An experiment on two datasets found that a two-stage CopyNet instantiation of Sequicity outperformed several baselines in effectiveness, efficiency and handling out-of-vocabulary requests.
The document summarizes a study that explored how people's strategies for giving commands to a robot change over time during a collaborative navigation task. Ten participants each directed a robot for one hour via dialogue. Initially, participants predominantly used metric units like distances in their commands, but over time their commands increasingly referred to environmental landmarks. The study collected audio, text, and robot data to analyze parameters in commands. Future work aims to automate dialogue response generation based on this data.
The document describes a system for estimating emotion intensity in tweets. It takes a lexicon-based and word vector-based approach to create sentence embeddings for tweets. Various regression models are trained and an ensemble is used to predict emotion intensity scores between 0-1 for anger, sadness, joy and fear. The system achieved third place in predicting emotion intensity and second place for intensities over 0.5. Future work involves using contextual sentence embeddings to improve predictions.
This document describes Toshiba's machine translation system submitted to the WAT2015 workshop. It discusses using statistical post-editing (SPE) to improve rule-based machine translation (RBMT) output, as well as combining SPE and SMT systems using reranking with recurrent neural network language models. Experimental results show that the combined system achieved the best BLEU and RIBES scores compared to the individual SPE and SMT systems on several language pairs, including Japanese-English and Chinese-Japanese. However, human evaluation correlations were not entirely clear.
FIID 2015 Guadalajara, Mexico
Graphic organizers for vocabulary learning
Several types and purposes of graphic organizers have been proposed over the years. In this presentation, several explore types and purposes of graphic organizers preserving the themes of using these in a scalable manner across the grade levels and between various disciplines through a literacy lens (e.g., sciences, mathematics, physical education, literary study, etc.). The primary focus of this session is to demonstrate how vocabulary, especially, can be more effectively taught using graphic organizers. Participants will learn to create graphic organizers to address specific learning purposes rather than relying solely on generic organizers. Five different organizers will be introduced, and participants will be able to experience the use of the organizer. Digital versions, paper versions prepared by the teacher, and manipulative versions created by the students will be explored.
This document provides an introduction to action research and how it can be applied. It discusses that action research involves collaboration between researchers and clients to diagnose and solve problems. The action research cycle involves planning a change, acting on it, observing the results, and reflecting to plan further changes.
Two case studies are described: one where researchers worked with students to help them better evaluate information, and another where researchers collaborated with teachers across Europe to design an online course based on the teachers' desired learning goals. Finally, the document provides advice on when and how to appropriately use action research and links to additional resources.
Research methodology hypothesis constructionSami Nighaoui
This document provides guidance on constructing a research hypothesis. It discusses identifying dependent and independent variables related to the topic of illegal immigrants' rights in the US. It suggests drafting an initial research question and theoretical statement explaining the relationship between variables. The document emphasizes that a hypothesis formally defines the causal relationship and must be falsifiable. It provides an example null hypothesis that level of education does not impact views on immigrant rights, and that rejecting this would support the alternative hypothesis.
These slides can share you what is particle swam optimization which is one of the technique that comes under the soft computing and artificial intelligence which is an emerging technology now and trending technology that every corporate company is working over this and hear I have given an example of that on particle swam optimization.
A friendly introduction to using action research in library and information environments. Presented at LILAC 2018 by Sam Aston, Geoff Walton and Emma Coonan.
The document provides an introduction to first-order logic and its advantages over propositional logic. It discusses elements of first-order logic like objects, relations, functions, equality, variables, quantifiers, and converting English statements to formal logic. Examples are provided to illustrate modeling relationships between objects and reasoning about multiple objects at once using quantifiers. The limitations of propositional logic in modeling statements involving objects and relations are demonstrated.
This document discusses logic and propositional logic. It covers the following topics:
- The history and applications of logic.
- Different types of statements and their grammar.
- Propositional logic including symbols, connectives, truth tables, and semantics.
- Quantifiers, universal and existential quantification, and properties of quantifiers.
- Normal forms such as disjunctive normal form and conjunctive normal form.
- Inference rules and the principle of mathematical induction, illustrated with examples.
This document introduces the natural sciences as an area of knowledge. It defines natural sciences as knowledge of observable objects and processes in nature, such as biology and physics, as distinguished from abstract sciences like mathematics. It describes the scientific method as involving observation, hypothesis formation, prediction, and experimental testing to confirm or falsify hypotheses. Personal knowledge and imagination play a role in scientific discovery alongside shared experimentation and observation. Questions are raised about whether scientific knowledge can be considered absolutely true or reliable given its tentative nature and reliance on indirect observation tools.
This document discusses Bayesian inference and Bayes' theorem. It provides examples of how Bayesian inference can be used to calculate conditional probabilities based on new evidence or data. Specifically, it discusses using Bayesian inference to determine the probability a die is "good" based on rolling results, and the probability a woman has breast cancer based on positive or negative mammogram results.
Similar to Yejin Choi - 2017 - From Naive Physics to Connotation: Modeling Commonsense in Frame Semantics (10)
This paper contributes a noun phrase-annotated SMS corpus and proposes a weak semi-Markov CRF model for noun phrase chunking in informal text. The weak semi-CRF model improves training speed over linear-CRF and semi-CRF models while maintaining similar accuracy. Experiments on the SMS corpus show the weak semi-CRF achieves F1 scores comparable to other models but trains faster, especially with larger training data sizes.
This document presents a new method for automatically detecting false friends between Spanish and Portuguese using word embeddings. The method builds word vector spaces for each language using word2vec, finds a linear transformation between the spaces, and measures vector distances to classify word pairs as cognates or false friends. In experiments on a dataset of 710 word pairs, the method achieved state-of-the-art accuracy of 77.28% and high coverage of 97.91%, outperforming previous work. Future work will explore using different word embeddings and fine-grained classifications of partial false friends.
This document describes a Spanish language corpus for humor analysis that was created through crowd-sourcing annotations. Over 27,000 tweets were collected from humorous accounts and annotated through a web interface. The corpus contains over 100,000 annotations of the tweets' humor and funniness. Inter-annotator agreement was higher for this corpus than a previous Spanish humor corpus. The dataset will help analyze subjectivity in humor and was used in a shared task on humor classification and funniness prediction.
This document discusses position bias in instructor interventions in MOOC discussion forums. It finds that instructors are more likely to intervene in threads that appear higher on the discussion forum user interface due to their recent activity. To address this, it proposes a debiased classifier that weights examples based on their propensity for intervention. It finds this approach identifies intervention opportunities that were overlooked due to position bias. The debiased classifier outperforms a standard classifier on several metrics, demonstrating it can better predict unbiased intervention needs.
The document summarizes the history and current state of the ACL Anthology, a repository of publications from ACL-sponsored conferences. It discusses how the Anthology was established in 2001 and is now maintained by volunteers, containing over 45,000 papers. The presentation calls for community involvement to help future-proof the Anthology through efforts like migrating its infrastructure and improving documentation. It also proposes hosting the Anthology on the main ACL website and recruiting a new editor.
The document presents SAMSA, a new automatic evaluation measure for structural text simplification. SAMSA uses semantic parsing to measure the preservation of semantic structures and relations between an original text and its simplified version. It correlates significantly better with human judgments of meaning preservation and structural simplicity than prior reference-based metrics. SAMSA is the first evaluation method designed specifically for structural simplification operations like sentence splitting.
(1) Sequicity is a framework that simplifies task-oriented dialogue systems using single sequence-to-sequence architectures.
(2) It formalizes dialogues as sequences of belief spans and responses and decodes them in two stages: generating a belief span followed by a response.
(3) An experiment on two datasets found that a two-stage CopyNet instantiation of Sequicity outperformed several baselines in effectiveness, efficiency and handling out-of-vocabulary requests.
The document summarizes a study that explored how people's strategies for giving commands to a robot change over time during a collaborative navigation task. Ten participants each directed a robot for one hour via dialogue. Initially, participants predominantly used metric units like distances in their commands, but over time their commands increasingly referred to environmental landmarks. The study collected audio, text, and robot data to analyze parameters in commands. Future work aims to automate dialogue response generation based on this data.
The document describes a system for estimating emotion intensity in tweets. It takes a lexicon-based and word vector-based approach to create sentence embeddings for tweets. Various regression models are trained and an ensemble is used to predict emotion intensity scores between 0-1 for anger, sadness, joy and fear. The system achieved third place in predicting emotion intensity and second place for intensities over 0.5. Future work involves using contextual sentence embeddings to improve predictions.
This document describes Toshiba's machine translation system submitted to the WAT2015 workshop. It discusses using statistical post-editing (SPE) to improve rule-based machine translation (RBMT) output, as well as combining SPE and SMT systems using reranking with recurrent neural network language models. Experimental results show that the combined system achieved the best BLEU and RIBES scores compared to the individual SPE and SMT systems on several language pairs, including Japanese-English and Chinese-Japanese. However, human evaluation correlations were not entirely clear.
The document describes improvements made to the KyotoEBMT machine translation system. It discusses using forest parsing of input sentences to handle parsing errors and syntactic divergences. It also describes using the Nile alignment tool along with constituent parsing to improve word alignments from the training corpus. New features were added and the reranking was improved by incorporating a neural machine translation-based bilingual language model.
El documento describe el sistema de traducción basado en ejemplos KyotoEBMT. El sistema utiliza análisis de dependencia tanto del idioma origen como del idioma destino y puede manejar ambigüedades en las hipótesis de traducción mediante el uso de reglas de rejilla. Los resultados oficiales del WAT2015 muestran mejoras en las métricas BLEU y RIBES con la reranqueación de traducciones, aunque la reranqueación empeora la evaluación humana para la dirección de traducción japonés-chino. El sistema Ky
This document evaluates several neural machine translation models for English to Japanese translation. It finds that simple neural models outperform statistical machine translation baselines. Soft attention models with LSTM units performed best. However, training these models on pre-reordered data hurt performance. The neural models tended to produce grammatically correct but incomplete translations by omitting information. Replacing unknown words helped some models but more sophisticated solutions are needed for models trained on natural order data.
This document evaluates various neural machine translation models for English to Japanese translation. It compares different network architectures, recurrent units, and training data configurations. Results show that soft-attention models outperformed multi-layer encoder-decoder models, and training on pre-reordered data hurt performance. Neural machine translation models tended to generate grammatically correct but incomplete translations.
This document describes NAVER's machine translation systems for the WAT 2015 evaluation. For English-to-Japanese translation, the best system combined tree-to-string syntax-based machine translation with neural machine translation re-ranking, achieving a BLEU score of 34.60. For Korean-to-Japanese translation, the top system used phrase-based machine translation and neural machine translation re-ranking, obtaining a BLEU score of 71.38. The document also analyzes the effectiveness of character-level tokenization and other techniques for neural machine translation.
Toshiba presented their machine translation system for the WAT2015 workshop. Their system uses statistical post-editing (SPE) to correct rule-based machine translation (RBMT) output. It also combines SPE and phrase-based statistical machine translation (SMT) results by reranking the merged n-best lists using a recurrent neural network language model. Evaluation showed the combined system achieved the best results on most language pairs compared to SPE and SMT individually. Analysis of system selections by the combination found it primarily chose translations from SPE.
The document summarizes research conducted by NICT at the WAT 2015 workshop. They tested simple translation techniques like reverse pre-reordering for Japanese-to-English and character-based translation for Korean-to-Japanese. The techniques were found to work effectively and the researchers encourage wider use of these techniques if confirmed through human evaluation at the workshop.
Neural reranking of machine translation output improves both automatic metrics and subjective human evaluations of translation quality. The document analyzes reranking results from a statistical machine translation system using an attentional neural machine translation model. Reranking corrected errors related to reordering, insertion, deletion, substitution and conjugation. Specifically, it improved phrasal reordering, auxiliary verb insertion/deletion, and coordinate structures. The gains were mainly in grammatical aspects rather than lexical selection. While reranking is shown to be effective, questions remain about comparing it to pure neural machine translation and neural language models.
More from Association for Computational Linguistics (20)
This presentation was provided by Rebecca Benner, Ph.D., of the American Society of Anesthesiologists, for the second session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session Two: 'Expanding Pathways to Publishing Careers,' was held June 13, 2024.
A Visual Guide to 1 Samuel | A Tale of Two HeartsSteve Thomason
These slides walk through the story of 1 Samuel. Samuel is the last judge of Israel. The people reject God and want a king. Saul is anointed as the first king, but he is not a good king. David, the shepherd boy is anointed and Saul is envious of him. David shows honor while Saul continues to self destruct.
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...TechSoup
Whether you're new to SEO or looking to refine your existing strategies, this webinar will provide you with actionable insights and practical tips to elevate your nonprofit's online presence.
Beyond Degrees - Empowering the Workforce in the Context of Skills-First.pptxEduSkills OECD
Iván Bornacelly, Policy Analyst at the OECD Centre for Skills, OECD, presents at the webinar 'Tackling job market gaps with a skills-first approach' on 12 June 2024
Temple of Asclepius in Thrace. Excavation resultsKrassimira Luka
The temple and the sanctuary around were dedicated to Asklepios Zmidrenus. This name has been known since 1875 when an inscription dedicated to him was discovered in Rome. The inscription is dated in 227 AD and was left by soldiers originating from the city of Philippopolis (modern Plovdiv).
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.pptHenry Hollis
The History of NZ 1870-1900.
Making of a Nation.
From the NZ Wars to Liberals,
Richard Seddon, George Grey,
Social Laboratory, New Zealand,
Confiscations, Kotahitanga, Kingitanga, Parliament, Suffrage, Repudiation, Economic Change, Agriculture, Gold Mining, Timber, Flax, Sheep, Dairying,
3. Intelligent Communication
è Reading between the lines
Understanding
what is said
+
what is not said
“CHEESEBURGER STABBING”
• Someone stabbed a cheeseburger?
• A cheeseburger stabbed someone?
• A cheeseburger stabbed another cheeseburger?
• Someone stabbed someone else with a cheeseburger?
• Someone stabbed someone else over a cheeseburger?
4. Intelligent Communication
è Reading between the lines
Understanding
what is said
+
what is not said
“CHEESEBURGER STABBING”
• Someone stabbed a cheeseburger?
• A cheeseburger stabbed someone?
• A cheeseburger stabbed another cheeseburger?
• Someone stabbed someone else with a cheeseburger?
• Someone stabbed someone else over a cheeseburger?
Knowledge about the world
to fill in the gap!
5. Blueberry Muffins
Ingredients
1 cup milk
1 egg
1/3 cup vegetable oil
2 cups all-purpose flour
2 teaspoons baking powder
1/2 cup white sugar
1/2 cup fresh blueberries
Procedure
1. Preheat oven to 400 degrees F. Line a 12-cup muffin tin with paper liners.
2. In a large bowl, stir together milk, egg, and oil. Add flour, baking powder, sugar, and
blueberries; gently mix the batter with only a few strokes. Spoon batter into cups.
3. Bake for 30 minutes. Serve hot.
http://allrecipes.com/Recipe/Blueberry-Muffins-I/
Intelligent Communication
Bake what? where?
Knowledge about the world
to fill in the gap!
6. Encyclopedic knowledge
– Who is the president of which
country and born in what year…
Commonsense knowledge
– It’s not possible to stab someone
using a cheeseburger
– Stabbing a cheeseburger is not
newsworthy…
– Stabbing someone is generally
immoral
– How to make a cheeseburger
Types of Knowledge
Information Extraction
Naïve Physics
Social Norms
Procedural (How-to)
Knowledge
9. Winograd Schema Challenge
• The trophy would not fit in the brown
suitcase because it was too big (small).
What was too big (small)?
Answer 0: the trophy
Answer 1: the suitcase
11. Support from Psychology Studies
• A familiar-size stroop effect. (Konkle, T., and Oliva, A. 2012.)
• Knowledge on size represented in a perceptual or analog format
(Moyer, 1973; Paivio, 1975; Rubinsten & Henik, 2002).
• Objects have a canonical visual size (Konkle & Oliva, 2011; Linsen,
Leyssen, Sammartino, & Palmer, 2011).
congruent incongruent
17. • Language – absolute
estimation
• “car is * x * m”
• “person is * m tall”
• Vision – relative
estimation
area(box1)
area(box2 )
×
depth(box1)2
depth(box2 )2
size(Oi )
size(Oj )
=
Are Elephants Bigger than Butterflies?
(Bagherinezhad et al. @ AAAI 2016)
18. Overcoming Reporting Bias
• People don’t state the obvious
– Elephants are bigger than butterflies
• Thoughts last year
– Look beyond language, such as vision!
– (AAAI 2016, ACL 2016, ICCV 2015)
– Can’t learn diverse physical knowledge (weight,
strength, rigidness, strength…)
• Thoughts this year
– Back to language, but with a different game plan!
22. x threw y
x is bigger than y
x weighs more than y
as a result, y will be moving faster than x
23. Let’s solve two related problems
Physical properties implied by predicates
“I picked up the <unk>.”
“The <unk> shattered when it hit
the ground
Physical properties of objects
“I dropped a cherry into
the <unk>.”
25. Learning Knowledge from Language
Using IE Patterns
(Gordon et al., 2010, Gordon and Schubert, 2012, Narisawa
et al. 2013, Tandon et al. 2014)
Narrative Schemas, Script knowledge
(Chambers and Jurafsky. 2009, Pichotta and Mooney 2014,
2016)
Inferring Knowledge through Reasoning
Inverting Grice’s maxims
Mohammad S. Sorrower, et al., 2011
Natural logic, natural reasoning, and entailment
MacCartney and Manning, 2007, Angeli and Manning 2013,
2014, Bowman et al., 2015
28. x threw y
⇒ x is larger than y
⇒ x is heavier than y
⇒ x is slower than y
R(agent, theme)
Representation: VerbPhysics Frames
action
theme
agent
tool
goal
…
“He threw the ball”
29. “He threw the ball”
x threw y
⇒ x is larger than y
⇒ x is heavier than y
⇒ x is slower than y
“We walked into the house”
x walked into y
⇒ x is smaller than y
⇒ x is lighter than y
⇒ x is faster than y
R(agent, theme)
Representation: VerbPhysics Frames
action
theme
agent
tool
goal
…
R(agent, goal)
30. “He threw the ball”
x threw y
⇒ x is larger than y
⇒ x is heavier than y
⇒ x is slower than y
“We walked into the house”
x walked into y
⇒ x is smaller than y
⇒ x is lighter than y
⇒ x is faster than y
“I squashed the bug with my boot”
squashed x with y
⇒ x is smaller than y
⇒ x is lighter than y
⇒ x is weaker than y
⇒ x is less rigid than y
⇒ x is slower than y
R(agent, theme) R(agent, goal) R(theme, goal)
Representation: VerbPhysics Frames
action
theme
agent
tool
goal
…
32. …
…
…
random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
attribute similarity
selectional preference
33. group of RVs
Each variable represents an entailment knowledge
“If X throws Y, then X > Y in terms of size”
Or
“If X throws Y, then Z in size” where Z = {>, <, =}
(Using Levin action verbs (Levin, 1993))
34. random variable (RV)
group of RVs
Each variable represents an entailment knowledge
“If X throws Y, then X > Y in terms of size”
Or
“If X throws Y, then Z in size” where Z = {>, <, =}
(Using Levin action verbs (Levin, 1993))
~ 8 frames per predicate
x threw y
PERSON threw x into y
PERSON threw x in y
PERSON threw x on y
PERSON threw x
PERSON threw out x
…
x threw out y
“threw”
35. random variable (RV)
group of RVs
factor connects RV
frame similarity
Each variable represents an entailment knowledge
“If X throws Y, then X > Y in terms of size”
Or
“If X throws Y, then Z in size” where Z = {>, <, =}
(Using Levin action verbs (Levin, 1993))
~ 8 frames per predicate
x threw y
PERSON threw x into y
PERSON threw x in y
PERSON threw x on y
PERSON threw x
PERSON threw out x
…
x threw out y
“threw”
37. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
38. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
Each variable represents relative object knowledge
“p is generally bigger than q in terms of size”
Or
“p Z q in size” where Z = {>, <, =}
39. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
Each variable represents relative object knowledge
“p is generally bigger than q in terms of size”
Or
“p Z q in size” where Z = {>, <, =}
40. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
Each variable represents relative object knowledge
“p is generally bigger than q in terms of size”
Or
“p Z q in size” where Z = {>, <, =}
Each variable represents an entailment knowledge
“If X throws Y, then X > Y in terms of size”
Or
“If X throws Y, then Z in size” where Z = {>, <, =}
41. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
selectional preference
Each variable represents relative object knowledge
“p is generally bigger than q in terms of size”
Or
“p Z q in size” where Z = {>, <, =}
Each variable represents an entailment knowledge
“If X throws Y, then X > Y in terms of size”
Or
“If X throws Y, then Z in size” where Z = {>, <, =}
42. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
selectional preference
43. random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
attribute similarity
selectional preference
44. …
random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
attribute similarity
selectional preference
45. …
…
…
random variable (RV)
group of RVs
factor connects RV
factors connect
subset of RVs
verb similarity
frame similarity
object similarity
attribute similarity
selectional preference
54. How Language Describes the World
“All I know is what I read in the papers.”
– Will Rogers
55. How Language Describes the World
(per pro-life)
Supreme
court
Texas
Law
“endanger”
Women
abortion
“protect
lives and
health”
“refused to hear”
“blocking regulations”
“protect”
“violate”
clinic
“close,
require”
56. An Alternative World View
(per pro-choice)
Supreme
court
Texas
Law
“strike down”
Women
abortion
“impose”
“seeks,
obtains”
Clinics
“violate,
cause”
“make choices,
has a constitutional right”
“shutdown,
force,
threaten”
57. Prior Literature
• A ton of work on sentiment analysis
- Especially on product reviews and movie reviews
- (Pang and Lee 2009 for survey, Socher et al 2013)
• Recent work on implied sentiment
– (Greene and Resnik 2009 , Feng et al 2013; Choi and Wiebe 2014;
Mohammad and Turney 2010; Deng and Wiebe 2014)
Supreme
court
Texas
Law
“strike down”
Women
“impose”
58. “criticize” frame
per left-leaning sources:
Obama criticize someone?
or
Obama is critized by someone?
(Based on the Stream Corpus (2014) w.r.t. 30 news sources with known liberal / conservative leanings)
59. Connotative Meaning of Words
• Semantic prosody (Sinclair 1991, Louw 1993)
“______ caused ______”
• Wittgenstein’s view:
– the meaning of a word is its use in the language
• Connotation arises from the (typical) context
in which words are used
• Can we encode connotation in frame
semantics?
headache, cancer, …
appreciation, award?
Texas
Law “cause”
61. “violate” frame
agent theme
writer
reader
=
=
=
• Perspective (x->y) : directed sentiment
• Effect (x) : whether the event is good for (or bad for) x
• State (x) : the private state of x as a result of the event
• Value (x) : the intrinsic value of x
Connotation judgments are not always categorical, some are better to be modeled as
distributions. (analogously to de Marneffe et al. 2012 on veridicality assessment)
62. “criticize” frame
per left-leaning sources:
Obama criticize someone?
or
Obama is critized by someone?
(Based on the Stream Corpus (2014) w.r.t. 30 news sources with known liberal / conservative leanings)
63. Left/right leaning sources
Verb Role Left-leaning sources Right-leaning sources
accuse subject [-] Putin, Progressives, Limbaugh,
Gingrich
activist, U.S., protestor, Chavez
object [+] Official, rival, administration,
leader
Romney, Iran, Gingrich, regime,
opposition
attack subject [-] McCain, Trump, Limbaugh Obama, campaign, Biden, Israel
object [+] Gingrich, Obama, policy citizen, Zimmerman
criticize subject [-] Ugandans, rival, Romney, Tyson Britain, passage, Obama,
Maddow
object [+] sensation, Obama, Allen,
Cameron, Congress
threat, Pelosi, Romey, GOP,
Republicans
suffer subject
[+]
woman, victim, officer, father Christie, Rangers, people, man
object [-] attack, blow, burn, seizure,
casualty
injury, fool, loss, concussion
(Based on the Stream Corpus (2014) w.r.t. 30 news sources with known liberal / conservative leanings)
64. Connotation Frames
1. Analysis on large-scale news corpora
2. Crowdsourcing experiments
15 annotators answered 16 questions for each of the top 1000 verbs from the NY times corpus.
Average Krippendorff’s alpha is 0.25 and 2-way soft agreement is 95%.
3. Computational models
Writer: The judge upheld the verdict.
Questions: How the judge feels about the verdict:
Negative PositiveNeutral
Negative or
Neutral
Positive or
Neutral
66. Base Model:
Embedding to Connotation Frames
word embedding for verb v
(Levy and Goldberg 2014)
…
Log-linear
parameterization
• Frame lexicon induction
• Separate frame prediction for each predicate
67. Connotation Dynamics
agent theme
writer
1. Perspective Triads:
~ social balance theory (Cartwright 1956)
is positive towards xj, and xj is positive towards xk, then we expect xi is al
mics hold for the negative case.
Pw!xi = ¬ (Pw!xj Pxi!xj )
ate has a positive effect on one argument, then we expect that the interactio
e. Similar dynamics hold for the negative case.
Exi = Pxj !xi
xi is presupposed as valuable, then we expect that the writer also views x
or the negative case.
Vxi = Pw!xi
icate has a positive effect on an argument xi, then we expect that xi will gain
mics hold for the negative case.
Sxi = Exi
ped Relations: we propose models that parameterize these dynamics using lo
68. Connotation Dynamics
agent theme
writer
1. Perspective Triads:
~ social balance theory (Cartwright 1956)
2. Effect – Mental State:
is positive towards xj, and xj is positive towards xk, then we expect xi is al
mics hold for the negative case.
Pw!xi = ¬ (Pw!xj Pxi!xj )
ate has a positive effect on one argument, then we expect that the interactio
e. Similar dynamics hold for the negative case.
Exi = Pxj !xi
xi is presupposed as valuable, then we expect that the writer also views x
or the negative case.
Vxi = Pw!xi
icate has a positive effect on an argument xi, then we expect that xi will gain
mics hold for the negative case.
Sxi = Exi
ped Relations: we propose models that parameterize these dynamics using lo
= ¬ (Pw!xj Pxi!xj )
sitive effect on one argument, then we expect that the interaction
ynamics hold for the negative case.
Exi = Pxj !xi
pposed as valuable, then we expect that the writer also views xi
ive case.
Vxi = Pw!xi
ositive effect on an argument xi, then we expect that xi will gain a
or the negative case.
Sxi = Exi
ns: we propose models that parameterize these dynamics using log-linea
ously unseen verbs. For each verb predicate, we in
fer a connotation frame composed of 9 relationshi
aspects that represent: perspective {P(w ! t)
P(w ! a), P(a ! t)}, effect {E(t), E(a)}, valu
69. Connotation Dynamics
agent theme
writer
1. Perspective Triads:
~ social balance theory (Cartwright 1956)
2. Effect – Mental State:
3. Effect – Perspective:
is positive towards xj, and xj is positive towards xk, then we expect xi is al
mics hold for the negative case.
Pw!xi = ¬ (Pw!xj Pxi!xj )
ate has a positive effect on one argument, then we expect that the interactio
e. Similar dynamics hold for the negative case.
Exi = Pxj !xi
xi is presupposed as valuable, then we expect that the writer also views x
or the negative case.
Vxi = Pw!xi
icate has a positive effect on an argument xi, then we expect that xi will gain
mics hold for the negative case.
Sxi = Exi
ped Relations: we propose models that parameterize these dynamics using lo
= ¬ (Pw!xj Pxi!xj )
sitive effect on one argument, then we expect that the interaction
ynamics hold for the negative case.
Exi = Pxj !xi
pposed as valuable, then we expect that the writer also views xi
ive case.
Vxi = Pw!xi
ositive effect on an argument xi, then we expect that xi will gain a
or the negative case.
Sxi = Exi
ns: we propose models that parameterize these dynamics using log-linea
ously unseen verbs. For each verb predicate, we in
fer a connotation frame composed of 9 relationshi
aspects that represent: perspective {P(w ! t)
P(w ! a), P(a ! t)}, effect {E(t), E(a)}, valu
e towards xj, and xj is positive towards xk, then we expect xi is also
for the negative case.
!xi = ¬ (Pw!xj Pxi!xj )
positive effect on one argument, then we expect that the interaction
r dynamics hold for the negative case.
Exi = Pxj !xi
esupposed as valuable, then we expect that the writer also views xi
gative case.
Vxi = Pw!xi
a positive effect on an argument xi, then we expect that xi will gain a
for the negative case.
Sxi = Exi
70. Connotation Dynamics
agent theme
writer
1. Perspective Triads:
~ social balance theory (Cartwright 1956)
2. Effect – Mental State:
3. Effect – Perspective:
4. Value – Perspective:
is positive towards xj, and xj is positive towards xk, then we expect xi is al
mics hold for the negative case.
Pw!xi = ¬ (Pw!xj Pxi!xj )
ate has a positive effect on one argument, then we expect that the interactio
e. Similar dynamics hold for the negative case.
Exi = Pxj !xi
xi is presupposed as valuable, then we expect that the writer also views x
or the negative case.
Vxi = Pw!xi
icate has a positive effect on an argument xi, then we expect that xi will gain
mics hold for the negative case.
Sxi = Exi
ped Relations: we propose models that parameterize these dynamics using lo
= ¬ (Pw!xj Pxi!xj )
sitive effect on one argument, then we expect that the interaction
ynamics hold for the negative case.
Exi = Pxj !xi
pposed as valuable, then we expect that the writer also views xi
ive case.
Vxi = Pw!xi
ositive effect on an argument xi, then we expect that xi will gain a
or the negative case.
Sxi = Exi
ns: we propose models that parameterize these dynamics using log-linea
ously unseen verbs. For each verb predicate, we in
fer a connotation frame composed of 9 relationshi
aspects that represent: perspective {P(w ! t)
P(w ! a), P(a ! t)}, effect {E(t), E(a)}, valu
e towards xj, and xj is positive towards xk, then we expect xi is also
for the negative case.
xi = ¬ (Pw!xj Pxi!xj )
positive effect on one argument, then we expect that the interaction
dynamics hold for the negative case.
Exi = Pxj !xi
supposed as valuable, then we expect that the writer also views xi
gative case.
Vxi = Pw!xi
a positive effect on an argument xi, then we expect that xi will gain a
for the negative case.
Sxi = Exi
e towards xj, and xj is positive towards xk, then we expect xi is also
for the negative case.
!xi = ¬ (Pw!xj Pxi!xj )
positive effect on one argument, then we expect that the interaction
r dynamics hold for the negative case.
Exi = Pxj !xi
esupposed as valuable, then we expect that the writer also views xi
gative case.
Vxi = Pw!xi
a positive effect on an argument xi, then we expect that xi will gain a
for the negative case.
Sxi = Exi
73. Power and Agency Frames
in Modern Films
Maarten Sap et al. (EMNLP 2017 in submission)
74. Bias in Modern Films
• Not a news per say, but more
insights from recent studies
((https://www.google.com/intl/en/about/main/gende
r-equality-films/)
• Bechdel test:
– At least two women
– talk to each other
– about something other than a man.
• Agarwal et al. 2015 report 40% of
the films pass the Bechdel test.
• However, those that pass the test
still have problems…
A character in Dykes to Watch Out For explains
the rules that later came to be known as the
Bechdel test (1985)
75. Character Portrayal in Films
“… Elsa objects, unleashing her powers…
Elsa discards her crown and creates a
palace… Elsa […] freezes Anna … She then
summons… Elsa ends the winter”
“… accidentally injures Anna … Hans
proposes to Anna … She gets lost …
chases Anna … Anna spots Hans … He
locks Anna in a room…”
“Cinderella lives an unhappy life… agrees to let
Cinderella go… he sees Cinderella… Cinderella
hears the clock… allowing the mice to free
Cinderella… Cinderella appears…”
76. Power and Agency Frame
agent theme
writer
readerAgency (x) Agency (x)
Power(x,y)
• Agency (x)
– Powerful, decisive, capable of pushing forward their own storyline
– “waiting” v.s. “searching” for your prince.
• Power (x,y)
– Differential authority level between the verb agent and theme
– “demanding” v.s. “begging” for mercy
77. Frame Verbs
Agency (+) terminate, silence, sue, command, invade,
throw, organize, authorize, persuade, battle
Agency (-) shimmer, remain, consist of, border, plummet,
age, whimper, inherit, recede, resemble
Power (agent) squash, seduce, assign, order, suspend,
regulate,
possess, eject, prohibit, ignite, destroy
Power (theme) obey, dread, ask, worship, belong to, trust,
bow to, serve, implore, require, salute
Collected using Mturk, annotated ~2000 verbs
Power and Agency Frame
78. Power and agency connotation frames run on the Wikipedia
plot summaries of Frozen (2013) and Cinderella (1950)
79. Bias in Modern Films
• Male characters tend to drive
the plot more through high
agency verbs
• Connotation frames provide
more nuanced analysis
compared to existing tests
(Bechdel, Mako Mori, Furiosa,
Sexy Lamp test)
Frame Gender
association
Agency (+) M**
Agency (-) F**
Power (agent) M**
Power (theme) n.s.
Gender association with connotation frames using a
logistic regression, controlling for number of words. **:
p<0.001 (Holm corrected)
772 scripts (Gorinski & Lapata, 2015), 21K characters with automatically extracted gender
80. To Conclude
• Aspects of connotation can be packed into
lexical and frame semantics
agent theme
writer
readerAgency
(x)
Agency
(x)
Power(x,
y)
81. NLP for Social Good!
• Identifying unwanted biases in language
– (Sap et al. @ EMNLP 2017!)
• graded deception detection in media
– www.politifact.org
– (Rashkin et al. @ EMNLP 2017!)
82. Game Plan
1. Commonsense Frame Semantics
– Naive physics
– Social commonsense
2. Modeling the World, not just Language
– Dynamic entities
– Action causalities
83. one vector
You can't cram the meaning of
a whole %&!$# sentence
into a single $&!#* vector!
a sequence of vectors?
by
byoutput
output
token
<S> token
tokendecode
decode
byinput tokenencode token
85. Task Definition
𝑓
• Input:
– Goal: an (optional) title phrase
– Agenda: a list of items to talk about
• Output: long text with many sentences
86. Blueberry Muffins
Ingredients
1 cup milk
1 egg
1/3 cup vegetable oil
2 cups all-purpose flour
2 teaspoons baking powder
1/2 cup white sugar
1/2 cup fresh blueberries
Procedure
1. Preheat oven to 400 degrees F. Line a 12-cup muffin tin with paper liners.
2. In a large bowl, stir together milk, egg, and oil. Add flour, baking powder, sugar, and
blueberries; gently mix the batter with only a few strokes. Spoon batter into cups.
3. Bake for 20 minutes. Serve hot.
http://allrecipes.com/Recipe/Blueberry-Muffins-I/
Cooking Instructions
87. How to generate recipes
planning models
interpretation generation
abstraction planning
recipe plan
recipe text
Is it possible to simplify this?
88. Document-level Graph Parsing
(Kiddon et al., 2015)
Blueberry Muffins
Ingredients
1 cup milk
1 egg
1/3 cup vegetable oil
2 cups all-purpose flour
2 teaspoons baking powder
1/2 cup white sugar
1/2 cup fresh blueberries
Procedure
1. Preheat oven to 400 degrees F (205 degrees C). Line
a 12-cup muffin tin with paper liners.
2. In a large bowl, stir together milk, egg, and oil. Add
flour, baking powder, sugar, and blueberries; gently mix
the batter with only a few strokes. Spoon batter into
cups.
3. Bake for 20 minutes. Serve hot.
1. Predicate argument
structure with implicit
arguments (~ SRL)
2. How entities are
introduced and then flow
through different actions
(~ reference resolution)
3. Discourse-level parsing
89. Recipe generation as machine
translation?
• Encoder-decoder NNs work great for
machine translation (Cho et al.2014, Sutskever et
al. 2014)
by
byoutput
output
token
<S> token
tokendecode
decode
byinput tokenencode token
• We can translate a
recipe title into
the text
90. Encode title - decode recipe
Cut each sandwich in halves.
Sandwiches with sandwiches.
Sandwiches, sandwiches, Sandwiches,
sandwiches, sandwiches
sandwiches, sandwiches, sandwiches,
sandwiches, sandwiches, sandwiches, or
sandwiches or triangles, a griddle, each
sandwich.
Top each with a slice of cheese, tomato,
and cheese.
Top with remaining cheese mixture.
Top with remaining cheese.
Broil until tops are bubbly and cheese is
melted, about 5 minutes.
sausage sandwiches
91. Recipe generation vs machine
translation
by
byrecipe
recipe
token
<S> token
tokendecode
decode
recipe title
ingredient 1
ingredient 2
ingredient 3
ingredient 4
Two input sources
• Only ~6-10% words align
between input and output.
• The rest must be generated
from context (and implicit
knowledge about cooking)
• Contextual switch between
two different input sources
92. Challenges
1. Modeling global coherence and coverage
– May not be possible to compress everything on to
one hidden vector!
2. Need a mechanism to track an “agenda”
-- which ingredients have been already used
-- what need to be yet introduced
95. Neural checklist model
LM
Chop
<S>
hidden state classifier:
non-ingredient
new ingredient
used ingredient
which ingredients
are still available
new hidden state
tomato salsagarlic
100. Hidden state classifier is soft
tomatoes
tomatoes
Add
Add
to
to
.
.
0.90
0.08
0.01
0.01
0.85 1.001.000.85
0.85
1.00
0.020.02
0.02
0.040.04
0.04
0.00
0.00
0.50
0.50
0.85
1.00
0.02
0.04
0.940.05
0.01
101. Neural
Recipe
Example #1
Cook eggplant in boiling water , covered , for 10 min .
Drain and cut in half lengthwise . scoop out insides
leaving 1/2 '' shell . Mash insides with cottage cheese ,
onion , bay leaf , basil , oregano and tomato sauce .
Preheat oven to 350 ^ stuff eggplant halves , place in
casserole dish and bake covered for 15 min .
Add a little water to bottom of pan to keep eggplant
moist . top with provolone cheese .
Bake 5 more min uncovered 1 serving =
In a small bowl , combine the cheese , eggplant ,
basil , oregano , tomato sauce and onion . Mix well .
Shape mixture into 6 patties , each about 3/4-inch
thick.
Place on baking sheet .
Bake at 350 degrees for 30 minutes or until lightly
browned .
Southern living magazine , sometime in 1980 .
Typed for you by nancy coleman .
eggplant
cheese cottage
lowfat
chopped onion
bay ground leaf
basil
oregano
tomato sauce
provolone
title: oven
eggplant
102. Example: skillet chicken rice
Ingredients
2 lb chicken pieces skinned
3 cups mushrooms fresh sliced
4 carrots peeled sliced 1/2”
3/4 cups long grain rice
1/2 c onion chopped
1 tsp poultry seasoning
1 tsp bouillon chicken granules
1/4 tsp salt
In a large skillet, brown chicken in oil.
Add chicken and rice; cook, stirring, until
chicken is no longer pink.
Add rice and broth; bring to a boil.
Reduce heat to low; cover and cook 5
minutes or until rice is tender and liquid is
absorbed. Stir in rice.
Heat to boiling; reduce heat.
Cover and simmer until rice is tender,
about 20 minutes.
Stir in rice. Heat to boiling; reduce heat.
Cover and simmer 20 minutes or until rice
is tender and liquid is absorbed and rice
is tender.
Baseline
In large skillet, heat rice and onion.
Add carrots and mushrooms; cook,
stirring constantly, until hot and
bubbly. Stir in seasonings. Cook
and stir over medium heat until hot
and bubbly. Stir in chicken and salt.
Cover and cook 5 minutes or until
chicken is no longer pink in center.
Serve over rice.
Checklist
Spray a 12-inch skillet with nonstick spray
coating.
Brown chicken pieces on all sides over
medium heat about 15 minutes.
Remove chicken.
Drain fat from skillet, if neccessary.
Add mushrooms, carrots, rice, onion,
bouillon, poultry seasoning, 2 cups
water, salt.
Place chicken atop rice mixture.
Cover; simmer 30 minutes or till chicken
and rice are done.
True recipe
103. Example: chocolate covered potato
chips
Ingredients
8 oz sweet milk chocolate
24 thick sliced ridged potato chips
Preheat oven to 350 degrees.
Grease and flour a 9 x 13 inch pan.
Bake at 350 degrees for 30 minutes.
Baseline
Bake in a 400 degree f oven for 15
minutes or until lightly browned.
Drain on paper towels. Melt
chocolate in top of double boiler
over simmering water. Remove from
heat and cool to room temp. Add
potato mixture to the potato
mixture; mix well. Cover tightly with
plastic wrap. Refrigerate until firm
enough to handle. Shape into
croquettes. Fry in hot oil until
golden brown. Drain on paper
towels. Serve hot.
Checklist
Melt 6 ounces of chocolate. Remove from
heat. While chocolate is still hot, mix the 2
ounces of chocolate into the melted
chocolate, in 3 additions, stirring until each
addition is completely melted before adding
the next. Dip the potato chips, 1 at a time, in
the chocolate. Coat completely and lift with a
small fork. Shake off excess chocolate by
rapping the fork on the edge of the bowl
lightly. Remove any drips from the bottom by
running the fork across the edge of the bowl.
Slide the chips onto a cookie sheet lined with
parchment or wax paper. Allow to cool until
solid. Let chips sit at room temperature or in
the refrigerator.
True recipe
104. Baselines
• Encoder-Decoder baseline (EncDec)
– Encode title and ingredients as sequence, decode
recipe
• Attention model baseline
– Use attention embedding to ingredient list during
output computation at each step
105. Human evaluation
• Amazon Mechanical Turk
0 1 2 3 4 5
Ingredient use
EncDec baseline
Checklist
True recipes
106. Human evaluation
• Amazon Mechanical Turk
0 1 2 3 4 5
Ingredient use
Follows goal
EncDec baseline
Checklist
True recipes
107. Human evaluation
• Amazon Mechanical Turk
0 1 2 3 4 5
Grammar
Ingredient use
Follows goal
EncDec baseline
Checklist
True recipes
108. Natural vs. Neural Language (on dev set)
Human Recipes
Checklist NN
Enc-Dec RNN
|V| = 5206
33% of
|V|
16% of
|V|
109. Neural checklist model for dialogue
generation
• Dialogue system output generation
– Hotel & restaurant informational systems
– Wen et al. 2015
– inform{name(Hotel Stratford),has_internet(no)}
• “Hotel Stratford does not have internet.”
110. Examples
inform(name=piperade;good_for_meal=dinner;food=basque)
piperade is a basque restaurant that is good for dinner.
inform(name='red door cafe';good_for_meal=breakfast;area='cathedral
hill';kids_allowed=no)
the red door cafe is good for breakfast, does not allow children and is in
the cathedral hill area.
inform(name='manna';area='hayes valley or inset';address='845 irving
street')
manna is in the hayes valley or inset area. the address is 845 irving street.
?select(dogs_allowed='yes';dogs_allowed='no')
would you like a hotel that allows dogs?
114. Static vs Dynamic Entity
Representations
• Human: what is your job?
• Machine: i’m a lawyer
• Human: what do you do?
• Machine: i’m a doctor
I am a lawyer
Example from Paperno et al. 2016
…
I am a doctor …
116. What’s missing in the end-to-end…
• Title: deep-fried cauliflower
• Ingredients: cauliflower, frying oil, sauce, salt, pepper.
è
Wash and dry the cauliflower.
Heat the oil in a skillet and fry the sauce until they are
golden brown.
Drain on paper towels.
Add the sauce to the sauce and mix well.
Serve hot or cold.
117. Game Plan - Recap
1. Commonsense Frame Semantics
– naive physics
– Social commonsense
2. Modeling the World, not just Language
– Dynamic entities
– Action causalities
Choose Action:
slice
Choose Entity:
tomato
Read:
Slice the
tomatoes
State Changes:
shape separated
Choose Action:
bake
Choose Entity:
<IMPLICIT>
Read:
Bake in pan for
20 minutes.
State Changes:
cookedness cooked
temperature hot
location oven
Read:
<END_RECIPE>
Action Causality
120. Toward Intelligent Communication
1. Commonsense Frame Semantics
– naive physics
– Social commonsense
– Packing knowledge into lexical/frame
semantics
2. Modeling the World, not just Language
– Dynamic entities
– Action causalities
122. “X stabbing Y with Z”
agent theme
writer
reader=
=
=
Connotation
Frame
Perspective(wàX) = The writer considers X guilty.
Effect(Y) = What happenedis bad for Y.
Value(Y) = Y is inherently valuable.
VerbPhysics Y < Z in RIGID It’s not possible to stab someone
with a cheeseburger.