This document discusses the challenges of semantically annotating adpositions. It notes that adpositions exhibit polysemy and that their meaning depends on construal. It proposes designing an annotation scheme with broad supersenses that can accommodate multiple semantic roles to handle cases where construal results in overlapping roles. The scheme aims for cross-linguistic applicability while maintaining coverage of all adposition tokens.
The pluralization of 'haber' in Puerto Rican SpanishJeroen Claes
This study investigates the pluralization of impersonal haber in a recent sample (March-April 2011) of 24 native speakers of the Spanish of San Juan, Puerto Rico, focusing upon three research questions: (i) What is the linguistic distribution of the pluralization of presentational haber in the Spanish of San Juan, Puerto Rico? (ii) What is the sociolinguistic distribution of the pluralization of presentational haber in the Spanish of San Juan, Puerto Rico? and, (iii) How can these distributions be explained in a psychologically and sociolinguistically adequate manner? In order to answer these interrogatives, against the background of Cognitive Construction Grammar, we propose the hypothesis that the phe-nomenon corresponds to an advanced ongoing language change from below that consists in the substi-tution of the canonical argument-structure construction, in which the NP functions as a direct object, by an innovative schema – identical in meaning, but different in sociolinguistic and stylistic signifi-cance –, in which the NP functions as a subject. The results that were obtained do not all support this model, but the data from the variables ‘Entrenchment of the verb-form in PRES-1’, ‘Degrees of con-ceptual complexity’, ‘Priming’, ‘Gender’, ‘Age’, and ‘Social prestige’ do argue in favor of it.
This document provides an introduction to modern linguistic pragmatics. It discusses neo-Gricean theories of scalar implicatures and the maxims of Horn and Levinson. Scalar implicatures are seen as central to neo-Griceanism and are experimentally testable, though there is no single theory. Alternatives include relevance, substitutability, monotonicity, symmetry problems, and whether alternatives are context-dependent or default. The literature discussed includes works by Chierchia, Fox, Gazdar, Geurts, Hirschberg, Horn, Katzir, and others.
This document discusses the natural language inference system TIL and how it relates to other logical systems like natural logic and fragments of first-order logic. TIL uses concepts, contexts, and roles to represent the meaning of sentences. Contexts are used to model phenomena like negation and propositional attitudes. TIL can handle certain syllogisms and fragments of language, but more work is needed to fully evaluate it against systems like those studied by Moss and Pratt-Hartmann. Further developing TIL's ability to model language constructs like relational syllogisms and sentential negation is suggested as future work.
The document discusses the logical system TIL, which stands for textual inference logic. TIL is a formalization of the AI system Bridge developed at PARC. TIL models natural language using concepts, contexts, and roles, similarly to description logics. It can represent sentences using conceptual and contextual structures. TIL satisfies the proof rules for syllogistic logic involving "all" and "some". It also satisfies proof systems for relational syllogisms and systems with individual names. The document discusses how TIL can be viewed in the context of other logical systems.
Integers and rationals can be expressed in first-order logic using only the less than relation <. The difference between integers and rationals can also be expressed in first-order logic. Specifically, integers have the property that between any two integers, there is no other integer, while this is not true for rationals - between any two rationals there is always another rational.
The document discusses deductive and inductive arguments. It provides examples of valid and invalid deductive arguments using categorical propositions and conditional premises. It also discusses inductive arguments, noting that inductive conclusions generalize from specific premises rather than necessarily following from them. The document then compares deductive and inductive arguments and discusses their uses in everyday life and mathematics. It concludes by introducing some common rules of inference for deductive arguments.
This document presents an approach to syntactic parsing that aims to achieve high accuracy with minimal structural annotation. The approach uses only surface features like the words in a span and the rules used to parse it, without additional annotations like part-of-speech tags or head words. Experimental results on the Penn Treebank show the approach achieves performance comparable to state-of-the-art parsers that use more complex annotations. The approach is also evaluated on multiple languages in the SPMRL 2013 shared task, outperforming baselines on average across nine treebanks.
The pluralization of 'haber' in Puerto Rican SpanishJeroen Claes
This study investigates the pluralization of impersonal haber in a recent sample (March-April 2011) of 24 native speakers of the Spanish of San Juan, Puerto Rico, focusing upon three research questions: (i) What is the linguistic distribution of the pluralization of presentational haber in the Spanish of San Juan, Puerto Rico? (ii) What is the sociolinguistic distribution of the pluralization of presentational haber in the Spanish of San Juan, Puerto Rico? and, (iii) How can these distributions be explained in a psychologically and sociolinguistically adequate manner? In order to answer these interrogatives, against the background of Cognitive Construction Grammar, we propose the hypothesis that the phe-nomenon corresponds to an advanced ongoing language change from below that consists in the substi-tution of the canonical argument-structure construction, in which the NP functions as a direct object, by an innovative schema – identical in meaning, but different in sociolinguistic and stylistic signifi-cance –, in which the NP functions as a subject. The results that were obtained do not all support this model, but the data from the variables ‘Entrenchment of the verb-form in PRES-1’, ‘Degrees of con-ceptual complexity’, ‘Priming’, ‘Gender’, ‘Age’, and ‘Social prestige’ do argue in favor of it.
This document provides an introduction to modern linguistic pragmatics. It discusses neo-Gricean theories of scalar implicatures and the maxims of Horn and Levinson. Scalar implicatures are seen as central to neo-Griceanism and are experimentally testable, though there is no single theory. Alternatives include relevance, substitutability, monotonicity, symmetry problems, and whether alternatives are context-dependent or default. The literature discussed includes works by Chierchia, Fox, Gazdar, Geurts, Hirschberg, Horn, Katzir, and others.
This document discusses the natural language inference system TIL and how it relates to other logical systems like natural logic and fragments of first-order logic. TIL uses concepts, contexts, and roles to represent the meaning of sentences. Contexts are used to model phenomena like negation and propositional attitudes. TIL can handle certain syllogisms and fragments of language, but more work is needed to fully evaluate it against systems like those studied by Moss and Pratt-Hartmann. Further developing TIL's ability to model language constructs like relational syllogisms and sentential negation is suggested as future work.
The document discusses the logical system TIL, which stands for textual inference logic. TIL is a formalization of the AI system Bridge developed at PARC. TIL models natural language using concepts, contexts, and roles, similarly to description logics. It can represent sentences using conceptual and contextual structures. TIL satisfies the proof rules for syllogistic logic involving "all" and "some". It also satisfies proof systems for relational syllogisms and systems with individual names. The document discusses how TIL can be viewed in the context of other logical systems.
Integers and rationals can be expressed in first-order logic using only the less than relation <. The difference between integers and rationals can also be expressed in first-order logic. Specifically, integers have the property that between any two integers, there is no other integer, while this is not true for rationals - between any two rationals there is always another rational.
The document discusses deductive and inductive arguments. It provides examples of valid and invalid deductive arguments using categorical propositions and conditional premises. It also discusses inductive arguments, noting that inductive conclusions generalize from specific premises rather than necessarily following from them. The document then compares deductive and inductive arguments and discusses their uses in everyday life and mathematics. It concludes by introducing some common rules of inference for deductive arguments.
This document presents an approach to syntactic parsing that aims to achieve high accuracy with minimal structural annotation. The approach uses only surface features like the words in a span and the rules used to parse it, without additional annotations like part-of-speech tags or head words. Experimental results on the Penn Treebank show the approach achieves performance comparable to state-of-the-art parsers that use more complex annotations. The approach is also evaluated on multiple languages in the SPMRL 2013 shared task, outperforming baselines on average across nine treebanks.
The document discusses propositional logic as a knowledge representation language. It defines key concepts in propositional logic including: syntax, semantics, validity, satisfiability, interpretation, models, and entailment. It explains that propositional logic uses symbols to represent facts about the world and connectives to combine symbols into sentences. Sentences can then be evaluated based on the truth values assigned to symbols to determine if the overall sentence is true or false. Propositional logic allows new sentences to be deduced from existing sentences through inference rules while maintaining logical validity.
1. The document discusses various concepts related to sentential meaning, including paraphrase, ambiguity, vagueness, tautology, presupposition, entailment, anomaly, contradiction, and analyticity.
2. Paraphrase refers to restating a sentence using synonyms while maintaining the same meaning, and can be lexical or structural. Ambiguity occurs when a sentence has more than one possible interpretation, either from individual words or syntactic structure.
3. Other concepts discussed include vagueness, which is a lack of definite meaning; tautology as unnecessary repetition; presupposition as background information assumed to be shared; and entailment as one sentence requiring the truth of another.
This document discusses ellipsis in English. It begins by defining ellipsis as something understood that is omitted from a clause or sentence but can be inferred from context. It then discusses different types of ellipsis including nominal ellipsis, verbal ellipsis, and clausal ellipsis. For nominal ellipsis, it examines specific deictics, non-specific deictics, post-deictics, numeratives, and epithets. For verbal ellipsis it looks at lexical ellipsis, operator ellipsis, polarity, finiteness and modality, and voice. It concludes by discussing modal and propositional ellipsis that can occur in clauses.
· Persuasive essay rubic
Exemplary
Proficient
Needs Improvement
Not Evident
Introduction and Claim/Thesis Statement
Well-developed introduction and thesis statement. They engage the reader and create interest. They clearly state the topic being analyzed and the persuasive argument.
The introduction and thesis statement are fairly well developed. But they are not very engaging or creative, but they do introduce the persuasive topic.
The introduction and thesis statement include the topic being analyzed, but they do not give a reason for the persuasive argument.
The introduction and thesis statement do not include what is being analyzed, and there is no apparent basis for persuasion.
Main Points and Organization
Main points in the essay are concrete and specific, and they effectively support the persuasive topic. The main points are organized and built upon the persuasive situation.
Main points are concrete and specific, but they are not engaging and interesting. There are sufficient details for the persuasion, but they could be clearer. The structure of the argument is confusing.
There are details, but they are either not concrete and specific, or there are not enough of them to adequately create a persuasive situation. One or more main points are not given sufficient space or details. The structure of the argument is unclear.
Main points are either wrong or lacking, or they are not related to the topic sentence or the analysis. There is no clearly defined structure or organization.
Research and Integration
The required number of resources are used and provide meaningful evidence to support assertions. Quotes are directly related to the argument. There is a strong alignment between argument and research.
The required number of resources are used and most quotes provide evidence to support assertions. Most of the quotes are related to the argument and an attempt has been made to align research and argument.
The required number of resources are not used. The minimum research used provides some support to the assertions. There is an insufficient development of ideas and research and argument are not aligned.
There are no resources used in this paper. It is a persuasive paper based on opinion and insufficient support.
Conclusion
The concluding paragraph effectively unifies the essay around the thesis statement. The reason for the analysis is clearly re-stated and the results are effectively summed-up.
The concluding paragraph unifies the essay around the thesis statement. The reason for the analysis is evident and results are presented as valid.
The concluding paragraph re-states the reason for the analysis, but it does little to unify the essay around the thesis statement.
There is no concluding paragraph, or it does not unify or restate the reason for the analysis.
Writing Mechanics
Writing is smooth, skillful, and coherent. Sentences are strong and expressive with varied structure. Consistent and appropriate tone and word choice is used throu.
1. The document discusses various approaches to semantic analysis including componential analysis and the idea of semantic universals.
2. Componential analysis breaks down meaning into distinct components but has limitations in how it handles relations between concepts.
3. There is debate around whether certain semantic features like gender are truly universal or influenced by other factors like culture and language contact.
Introduction to logic and prolog - Part 1Sabu Francis
The document provides an introduction to logic and Prolog programming. It discusses:
1) Alan Turing's invention of the modern computer to solve complex problems like decoding encrypted messages. This established the concept of algorithms being carried out through linear instruction processing.
2) Prolog programming focuses solely on logic and removes concerns about procedural elements like instruction pointers. It allows programmers to focus only on the problem's logic.
3) Logic is a tool for reasoning that uses concepts like true, false, if-then statements, and, or, etc. It helps clarify reasoning but cannot validate conclusions on its own if premises are flawed.
The Ins and Outs of Preposition Semantics: Challenges in Comprehensive Corpu...Seth Grimes
Presentation by Nathan Scheider, Georgetown University, to the Washington DC Natural Language Processing meetup, October 14, 2019, https://www.meetup.com/DC-NLP/events/264894589/.
Preposition Semantics: Challenges in Comprehensive Corpus Annotation and Auto...Seth Grimes
The document summarizes Nathan Schneider's presentation on preposition semantics. It discusses challenges in annotating prepositions in corpora and approaches to their semantic description and disambiguation. It presents Schneider's work on developing a unified semantic scheme for prepositions and possessives consisting of 50 semantic classes applied to a corpus of English web reviews. Inter-annotator agreement for the new corpus was 78%. Models for preposition disambiguation were evaluated, with the feature-rich linear model achieving the highest accuracy of 80%.
This document provides an overview of syntax and generative grammar. It discusses key concepts like deep and surface structure, structural ambiguity, recursion, phrase structure rules, lexical rules, complement phrases, and transformational rules. Tree diagrams and other symbols are presented to describe syntactic structures. The goal of generative grammar is to have a system of explicit rules that can generate all valid syntactic structures of a language while avoiding invalid ones.
The document discusses various types of definitions that can be used to reduce ambiguity and vagueness in arguments, including: ostensive definitions, which involve directly pointing out the thing being defined; verbal extensional definitions, which list examples of members of the set; intensional definitions, which outline the properties of members; lexical definitions from dictionaries; stipulative definitions for new terms; precising definitions to clarify vague terms; theoretical definitions based on a particular theory; and operational definitions that provide instructions for determining if something fits the definition. It also discusses ambiguity, vagueness, emotive or persuasive definitions, and gives examples and homework assignments related to defining terms clearly.
This document discusses collocations and methods for identifying them. It defines a collocation as a conventional multi-word expression whose meaning cannot be directly derived from the individual words. The document outlines criteria for collocations including non-compositionality, non-substitutability, and non-modifiability. It then describes several computational approaches for finding collocations, including frequency-based methods, measures of mean and variance between words, hypothesis testing using t-tests and chi-square tests, and pointwise mutual information. Each method is explained with examples. The document evaluates the strengths and weaknesses of these different approaches.
The document discusses propositional logic and first-order logic. It states that propositional logic has limited expressive power and cannot represent certain statements involving relationships between objects. First-order logic extends propositional logic by adding predicates and quantifiers, allowing it to more concisely represent natural language statements and relationships between objects. The key characteristics of first-order logic in AI are that it allows logical inference, more accurately represents facts about the real world, and provides a better theoretical foundation for program design.
The document discusses concepts related to thinking and language. It covers topics such as concepts, problem solving, decision making, belief bias, language structure, language development in children, and thinking in different animal species. The relationship between thinking and language is also explored, with language said to influence how people think.
The document discusses representing knowledge using predicate logic. It provides examples of representing simple facts and relationships using predicates, variables, and quantification. It also discusses issues like representing classes and instances, exceptions to general rules, and using computable predicates to efficiently represent relationships that can be computed rather than explicitly stated.
The document discusses sentence patterns and syntax in language. It explains that syntax rules determine how words are organized into phrases and sentences. There are two principles of sentence organization - linear order, where words must be in a specific sequence, and hierarchical structure, where words are grouped into constituents that make up the overall meaning. Syntactic rules show how words are arranged in order and grouped into constituents. These rules can reveal ambiguities in sentences and account for a speaker's ability to understand and produce novel sentences.
By watching this Power Point presentation, you'll acquire the necessary tools as well as basic information that is needed whenever you want to evaluate Vocabulary.
English 6-dlp-4-decoding-meaning-of-unfamiliar-words-using-structurAlice Failano
This module teaches learners how to determine the meaning of unfamiliar words using structural analysis, specifically by breaking words into prefixes and suffixes. It provides examples of common prefixes and their meanings, such as "under-" meaning less or less than average. Learners practice identifying prefixes, suffixes, and their meanings in sentences and words. Activities include forming new words by adding prefixes or suffixes to root words, using the new words in sentences, and self-checking answers.
English 6-dlp-4-decoding-meaning-of-unfamiliar-words-using-structurAlice Failano
This module teaches learners how to determine the meaning of unfamiliar words using structural analysis, specifically by breaking words into prefixes and suffixes. It provides examples of common prefixes and their meanings, such as "under-" meaning less than, and "over-" meaning too much. Learners are given practice identifying prefixes, adding prefixes to change word meanings, and using context clues to determine the meanings of new words formed with suffixes. The module aims to enhance learners' skills in deconstructing words to discern their definitions.
This paper contributes a noun phrase-annotated SMS corpus and proposes a weak semi-Markov CRF model for noun phrase chunking in informal text. The weak semi-CRF model improves training speed over linear-CRF and semi-CRF models while maintaining similar accuracy. Experiments on the SMS corpus show the weak semi-CRF achieves F1 scores comparable to other models but trains faster, especially with larger training data sizes.
This document presents a new method for automatically detecting false friends between Spanish and Portuguese using word embeddings. The method builds word vector spaces for each language using word2vec, finds a linear transformation between the spaces, and measures vector distances to classify word pairs as cognates or false friends. In experiments on a dataset of 710 word pairs, the method achieved state-of-the-art accuracy of 77.28% and high coverage of 97.91%, outperforming previous work. Future work will explore using different word embeddings and fine-grained classifications of partial false friends.
More Related Content
Similar to Jena Hwang - Double Trouble: The Problem of Construal in Semantic Annotation of Adpositions
The document discusses propositional logic as a knowledge representation language. It defines key concepts in propositional logic including: syntax, semantics, validity, satisfiability, interpretation, models, and entailment. It explains that propositional logic uses symbols to represent facts about the world and connectives to combine symbols into sentences. Sentences can then be evaluated based on the truth values assigned to symbols to determine if the overall sentence is true or false. Propositional logic allows new sentences to be deduced from existing sentences through inference rules while maintaining logical validity.
1. The document discusses various concepts related to sentential meaning, including paraphrase, ambiguity, vagueness, tautology, presupposition, entailment, anomaly, contradiction, and analyticity.
2. Paraphrase refers to restating a sentence using synonyms while maintaining the same meaning, and can be lexical or structural. Ambiguity occurs when a sentence has more than one possible interpretation, either from individual words or syntactic structure.
3. Other concepts discussed include vagueness, which is a lack of definite meaning; tautology as unnecessary repetition; presupposition as background information assumed to be shared; and entailment as one sentence requiring the truth of another.
This document discusses ellipsis in English. It begins by defining ellipsis as something understood that is omitted from a clause or sentence but can be inferred from context. It then discusses different types of ellipsis including nominal ellipsis, verbal ellipsis, and clausal ellipsis. For nominal ellipsis, it examines specific deictics, non-specific deictics, post-deictics, numeratives, and epithets. For verbal ellipsis it looks at lexical ellipsis, operator ellipsis, polarity, finiteness and modality, and voice. It concludes by discussing modal and propositional ellipsis that can occur in clauses.
· Persuasive essay rubic
Exemplary
Proficient
Needs Improvement
Not Evident
Introduction and Claim/Thesis Statement
Well-developed introduction and thesis statement. They engage the reader and create interest. They clearly state the topic being analyzed and the persuasive argument.
The introduction and thesis statement are fairly well developed. But they are not very engaging or creative, but they do introduce the persuasive topic.
The introduction and thesis statement include the topic being analyzed, but they do not give a reason for the persuasive argument.
The introduction and thesis statement do not include what is being analyzed, and there is no apparent basis for persuasion.
Main Points and Organization
Main points in the essay are concrete and specific, and they effectively support the persuasive topic. The main points are organized and built upon the persuasive situation.
Main points are concrete and specific, but they are not engaging and interesting. There are sufficient details for the persuasion, but they could be clearer. The structure of the argument is confusing.
There are details, but they are either not concrete and specific, or there are not enough of them to adequately create a persuasive situation. One or more main points are not given sufficient space or details. The structure of the argument is unclear.
Main points are either wrong or lacking, or they are not related to the topic sentence or the analysis. There is no clearly defined structure or organization.
Research and Integration
The required number of resources are used and provide meaningful evidence to support assertions. Quotes are directly related to the argument. There is a strong alignment between argument and research.
The required number of resources are used and most quotes provide evidence to support assertions. Most of the quotes are related to the argument and an attempt has been made to align research and argument.
The required number of resources are not used. The minimum research used provides some support to the assertions. There is an insufficient development of ideas and research and argument are not aligned.
There are no resources used in this paper. It is a persuasive paper based on opinion and insufficient support.
Conclusion
The concluding paragraph effectively unifies the essay around the thesis statement. The reason for the analysis is clearly re-stated and the results are effectively summed-up.
The concluding paragraph unifies the essay around the thesis statement. The reason for the analysis is evident and results are presented as valid.
The concluding paragraph re-states the reason for the analysis, but it does little to unify the essay around the thesis statement.
There is no concluding paragraph, or it does not unify or restate the reason for the analysis.
Writing Mechanics
Writing is smooth, skillful, and coherent. Sentences are strong and expressive with varied structure. Consistent and appropriate tone and word choice is used throu.
1. The document discusses various approaches to semantic analysis including componential analysis and the idea of semantic universals.
2. Componential analysis breaks down meaning into distinct components but has limitations in how it handles relations between concepts.
3. There is debate around whether certain semantic features like gender are truly universal or influenced by other factors like culture and language contact.
Introduction to logic and prolog - Part 1Sabu Francis
The document provides an introduction to logic and Prolog programming. It discusses:
1) Alan Turing's invention of the modern computer to solve complex problems like decoding encrypted messages. This established the concept of algorithms being carried out through linear instruction processing.
2) Prolog programming focuses solely on logic and removes concerns about procedural elements like instruction pointers. It allows programmers to focus only on the problem's logic.
3) Logic is a tool for reasoning that uses concepts like true, false, if-then statements, and, or, etc. It helps clarify reasoning but cannot validate conclusions on its own if premises are flawed.
The Ins and Outs of Preposition Semantics: Challenges in Comprehensive Corpu...Seth Grimes
Presentation by Nathan Scheider, Georgetown University, to the Washington DC Natural Language Processing meetup, October 14, 2019, https://www.meetup.com/DC-NLP/events/264894589/.
Preposition Semantics: Challenges in Comprehensive Corpus Annotation and Auto...Seth Grimes
The document summarizes Nathan Schneider's presentation on preposition semantics. It discusses challenges in annotating prepositions in corpora and approaches to their semantic description and disambiguation. It presents Schneider's work on developing a unified semantic scheme for prepositions and possessives consisting of 50 semantic classes applied to a corpus of English web reviews. Inter-annotator agreement for the new corpus was 78%. Models for preposition disambiguation were evaluated, with the feature-rich linear model achieving the highest accuracy of 80%.
This document provides an overview of syntax and generative grammar. It discusses key concepts like deep and surface structure, structural ambiguity, recursion, phrase structure rules, lexical rules, complement phrases, and transformational rules. Tree diagrams and other symbols are presented to describe syntactic structures. The goal of generative grammar is to have a system of explicit rules that can generate all valid syntactic structures of a language while avoiding invalid ones.
The document discusses various types of definitions that can be used to reduce ambiguity and vagueness in arguments, including: ostensive definitions, which involve directly pointing out the thing being defined; verbal extensional definitions, which list examples of members of the set; intensional definitions, which outline the properties of members; lexical definitions from dictionaries; stipulative definitions for new terms; precising definitions to clarify vague terms; theoretical definitions based on a particular theory; and operational definitions that provide instructions for determining if something fits the definition. It also discusses ambiguity, vagueness, emotive or persuasive definitions, and gives examples and homework assignments related to defining terms clearly.
This document discusses collocations and methods for identifying them. It defines a collocation as a conventional multi-word expression whose meaning cannot be directly derived from the individual words. The document outlines criteria for collocations including non-compositionality, non-substitutability, and non-modifiability. It then describes several computational approaches for finding collocations, including frequency-based methods, measures of mean and variance between words, hypothesis testing using t-tests and chi-square tests, and pointwise mutual information. Each method is explained with examples. The document evaluates the strengths and weaknesses of these different approaches.
The document discusses propositional logic and first-order logic. It states that propositional logic has limited expressive power and cannot represent certain statements involving relationships between objects. First-order logic extends propositional logic by adding predicates and quantifiers, allowing it to more concisely represent natural language statements and relationships between objects. The key characteristics of first-order logic in AI are that it allows logical inference, more accurately represents facts about the real world, and provides a better theoretical foundation for program design.
The document discusses concepts related to thinking and language. It covers topics such as concepts, problem solving, decision making, belief bias, language structure, language development in children, and thinking in different animal species. The relationship between thinking and language is also explored, with language said to influence how people think.
The document discusses representing knowledge using predicate logic. It provides examples of representing simple facts and relationships using predicates, variables, and quantification. It also discusses issues like representing classes and instances, exceptions to general rules, and using computable predicates to efficiently represent relationships that can be computed rather than explicitly stated.
The document discusses sentence patterns and syntax in language. It explains that syntax rules determine how words are organized into phrases and sentences. There are two principles of sentence organization - linear order, where words must be in a specific sequence, and hierarchical structure, where words are grouped into constituents that make up the overall meaning. Syntactic rules show how words are arranged in order and grouped into constituents. These rules can reveal ambiguities in sentences and account for a speaker's ability to understand and produce novel sentences.
By watching this Power Point presentation, you'll acquire the necessary tools as well as basic information that is needed whenever you want to evaluate Vocabulary.
English 6-dlp-4-decoding-meaning-of-unfamiliar-words-using-structurAlice Failano
This module teaches learners how to determine the meaning of unfamiliar words using structural analysis, specifically by breaking words into prefixes and suffixes. It provides examples of common prefixes and their meanings, such as "under-" meaning less or less than average. Learners practice identifying prefixes, suffixes, and their meanings in sentences and words. Activities include forming new words by adding prefixes or suffixes to root words, using the new words in sentences, and self-checking answers.
English 6-dlp-4-decoding-meaning-of-unfamiliar-words-using-structurAlice Failano
This module teaches learners how to determine the meaning of unfamiliar words using structural analysis, specifically by breaking words into prefixes and suffixes. It provides examples of common prefixes and their meanings, such as "under-" meaning less than, and "over-" meaning too much. Learners are given practice identifying prefixes, adding prefixes to change word meanings, and using context clues to determine the meanings of new words formed with suffixes. The module aims to enhance learners' skills in deconstructing words to discern their definitions.
Similar to Jena Hwang - Double Trouble: The Problem of Construal in Semantic Annotation of Adpositions (20)
This paper contributes a noun phrase-annotated SMS corpus and proposes a weak semi-Markov CRF model for noun phrase chunking in informal text. The weak semi-CRF model improves training speed over linear-CRF and semi-CRF models while maintaining similar accuracy. Experiments on the SMS corpus show the weak semi-CRF achieves F1 scores comparable to other models but trains faster, especially with larger training data sizes.
This document presents a new method for automatically detecting false friends between Spanish and Portuguese using word embeddings. The method builds word vector spaces for each language using word2vec, finds a linear transformation between the spaces, and measures vector distances to classify word pairs as cognates or false friends. In experiments on a dataset of 710 word pairs, the method achieved state-of-the-art accuracy of 77.28% and high coverage of 97.91%, outperforming previous work. Future work will explore using different word embeddings and fine-grained classifications of partial false friends.
This document describes a Spanish language corpus for humor analysis that was created through crowd-sourcing annotations. Over 27,000 tweets were collected from humorous accounts and annotated through a web interface. The corpus contains over 100,000 annotations of the tweets' humor and funniness. Inter-annotator agreement was higher for this corpus than a previous Spanish humor corpus. The dataset will help analyze subjectivity in humor and was used in a shared task on humor classification and funniness prediction.
This document discusses position bias in instructor interventions in MOOC discussion forums. It finds that instructors are more likely to intervene in threads that appear higher on the discussion forum user interface due to their recent activity. To address this, it proposes a debiased classifier that weights examples based on their propensity for intervention. It finds this approach identifies intervention opportunities that were overlooked due to position bias. The debiased classifier outperforms a standard classifier on several metrics, demonstrating it can better predict unbiased intervention needs.
The document summarizes the history and current state of the ACL Anthology, a repository of publications from ACL-sponsored conferences. It discusses how the Anthology was established in 2001 and is now maintained by volunteers, containing over 45,000 papers. The presentation calls for community involvement to help future-proof the Anthology through efforts like migrating its infrastructure and improving documentation. It also proposes hosting the Anthology on the main ACL website and recruiting a new editor.
The document presents SAMSA, a new automatic evaluation measure for structural text simplification. SAMSA uses semantic parsing to measure the preservation of semantic structures and relations between an original text and its simplified version. It correlates significantly better with human judgments of meaning preservation and structural simplicity than prior reference-based metrics. SAMSA is the first evaluation method designed specifically for structural simplification operations like sentence splitting.
(1) Sequicity is a framework that simplifies task-oriented dialogue systems using single sequence-to-sequence architectures.
(2) It formalizes dialogues as sequences of belief spans and responses and decodes them in two stages: generating a belief span followed by a response.
(3) An experiment on two datasets found that a two-stage CopyNet instantiation of Sequicity outperformed several baselines in effectiveness, efficiency and handling out-of-vocabulary requests.
The document summarizes a study that explored how people's strategies for giving commands to a robot change over time during a collaborative navigation task. Ten participants each directed a robot for one hour via dialogue. Initially, participants predominantly used metric units like distances in their commands, but over time their commands increasingly referred to environmental landmarks. The study collected audio, text, and robot data to analyze parameters in commands. Future work aims to automate dialogue response generation based on this data.
The document describes a system for estimating emotion intensity in tweets. It takes a lexicon-based and word vector-based approach to create sentence embeddings for tweets. Various regression models are trained and an ensemble is used to predict emotion intensity scores between 0-1 for anger, sadness, joy and fear. The system achieved third place in predicting emotion intensity and second place for intensities over 0.5. Future work involves using contextual sentence embeddings to improve predictions.
This document describes Toshiba's machine translation system submitted to the WAT2015 workshop. It discusses using statistical post-editing (SPE) to improve rule-based machine translation (RBMT) output, as well as combining SPE and SMT systems using reranking with recurrent neural network language models. Experimental results show that the combined system achieved the best BLEU and RIBES scores compared to the individual SPE and SMT systems on several language pairs, including Japanese-English and Chinese-Japanese. However, human evaluation correlations were not entirely clear.
The document describes improvements made to the KyotoEBMT machine translation system. It discusses using forest parsing of input sentences to handle parsing errors and syntactic divergences. It also describes using the Nile alignment tool along with constituent parsing to improve word alignments from the training corpus. New features were added and the reranking was improved by incorporating a neural machine translation-based bilingual language model.
El documento describe el sistema de traducción basado en ejemplos KyotoEBMT. El sistema utiliza análisis de dependencia tanto del idioma origen como del idioma destino y puede manejar ambigüedades en las hipótesis de traducción mediante el uso de reglas de rejilla. Los resultados oficiales del WAT2015 muestran mejoras en las métricas BLEU y RIBES con la reranqueación de traducciones, aunque la reranqueación empeora la evaluación humana para la dirección de traducción japonés-chino. El sistema Ky
This document evaluates several neural machine translation models for English to Japanese translation. It finds that simple neural models outperform statistical machine translation baselines. Soft attention models with LSTM units performed best. However, training these models on pre-reordered data hurt performance. The neural models tended to produce grammatically correct but incomplete translations by omitting information. Replacing unknown words helped some models but more sophisticated solutions are needed for models trained on natural order data.
This document evaluates various neural machine translation models for English to Japanese translation. It compares different network architectures, recurrent units, and training data configurations. Results show that soft-attention models outperformed multi-layer encoder-decoder models, and training on pre-reordered data hurt performance. Neural machine translation models tended to generate grammatically correct but incomplete translations.
This document describes NAVER's machine translation systems for the WAT 2015 evaluation. For English-to-Japanese translation, the best system combined tree-to-string syntax-based machine translation with neural machine translation re-ranking, achieving a BLEU score of 34.60. For Korean-to-Japanese translation, the top system used phrase-based machine translation and neural machine translation re-ranking, obtaining a BLEU score of 71.38. The document also analyzes the effectiveness of character-level tokenization and other techniques for neural machine translation.
Toshiba presented their machine translation system for the WAT2015 workshop. Their system uses statistical post-editing (SPE) to correct rule-based machine translation (RBMT) output. It also combines SPE and phrase-based statistical machine translation (SMT) results by reranking the merged n-best lists using a recurrent neural network language model. Evaluation showed the combined system achieved the best results on most language pairs compared to SPE and SMT individually. Analysis of system selections by the combination found it primarily chose translations from SPE.
The document summarizes research conducted by NICT at the WAT 2015 workshop. They tested simple translation techniques like reverse pre-reordering for Japanese-to-English and character-based translation for Korean-to-Japanese. The techniques were found to work effectively and the researchers encourage wider use of these techniques if confirmed through human evaluation at the workshop.
Neural reranking of machine translation output improves both automatic metrics and subjective human evaluations of translation quality. The document analyzes reranking results from a statistical machine translation system using an attentional neural machine translation model. Reranking corrected errors related to reordering, insertion, deletion, substitution and conjugation. Specifically, it improved phrasal reordering, auxiliary verb insertion/deletion, and coordinate structures. The gains were mainly in grammatical aspects rather than lexical selection. While reranking is shown to be effective, questions remain about comparing it to pure neural machine translation and neural language models.
More from Association for Computational Linguistics (20)
Strategies for Effective Upskilling is a presentation by Chinwendu Peace in a Your Skill Boost Masterclass organisation by the Excellence Foundation for South Sudan on 08th and 09th June 2024 from 1 PM to 3 PM on each day.
LAND USE LAND COVER AND NDVI OF MIRZAPUR DISTRICT, UPRAHUL
This Dissertation explores the particular circumstances of Mirzapur, a region located in the
core of India. Mirzapur, with its varied terrains and abundant biodiversity, offers an optimal
environment for investigating the changes in vegetation cover dynamics. Our study utilizes
advanced technologies such as GIS (Geographic Information Systems) and Remote sensing to
analyze the transformations that have taken place over the course of a decade.
The complex relationship between human activities and the environment has been the focus
of extensive research and worry. As the global community grapples with swift urbanization,
population expansion, and economic progress, the effects on natural ecosystems are becoming
more evident. A crucial element of this impact is the alteration of vegetation cover, which plays a
significant role in maintaining the ecological equilibrium of our planet.Land serves as the foundation for all human activities and provides the necessary materials for
these activities. As the most crucial natural resource, its utilization by humans results in different
'Land uses,' which are determined by both human activities and the physical characteristics of the
land.
The utilization of land is impacted by human needs and environmental factors. In countries
like India, rapid population growth and the emphasis on extensive resource exploitation can lead
to significant land degradation, adversely affecting the region's land cover.
Therefore, human intervention has significantly influenced land use patterns over many
centuries, evolving its structure over time and space. In the present era, these changes have
accelerated due to factors such as agriculture and urbanization. Information regarding land use and
cover is essential for various planning and management tasks related to the Earth's surface,
providing crucial environmental data for scientific, resource management, policy purposes, and
diverse human activities.
Accurate understanding of land use and cover is imperative for the development planning
of any area. Consequently, a wide range of professionals, including earth system scientists, land
and water managers, and urban planners, are interested in obtaining data on land use and cover
changes, conversion trends, and other related patterns. The spatial dimensions of land use and
cover support policymakers and scientists in making well-informed decisions, as alterations in
these patterns indicate shifts in economic and social conditions. Monitoring such changes with the
help of Advanced technologies like Remote Sensing and Geographic Information Systems is
crucial for coordinated efforts across different administrative levels. Advanced technologies like
Remote Sensing and Geographic Information Systems
9
Changes in vegetation cover refer to variations in the distribution, composition, and overall
structure of plant communities across different temporal and spatial scales. These changes can
occur natural.
ISO/IEC 27001, ISO/IEC 42001, and GDPR: Best Practices for Implementation and...PECB
Denis is a dynamic and results-driven Chief Information Officer (CIO) with a distinguished career spanning information systems analysis and technical project management. With a proven track record of spearheading the design and delivery of cutting-edge Information Management solutions, he has consistently elevated business operations, streamlined reporting functions, and maximized process efficiency.
Certified as an ISO/IEC 27001: Information Security Management Systems (ISMS) Lead Implementer, Data Protection Officer, and Cyber Risks Analyst, Denis brings a heightened focus on data security, privacy, and cyber resilience to every endeavor.
His expertise extends across a diverse spectrum of reporting, database, and web development applications, underpinned by an exceptional grasp of data storage and virtualization technologies. His proficiency in application testing, database administration, and data cleansing ensures seamless execution of complex projects.
What sets Denis apart is his comprehensive understanding of Business and Systems Analysis technologies, honed through involvement in all phases of the Software Development Lifecycle (SDLC). From meticulous requirements gathering to precise analysis, innovative design, rigorous development, thorough testing, and successful implementation, he has consistently delivered exceptional results.
Throughout his career, he has taken on multifaceted roles, from leading technical project management teams to owning solutions that drive operational excellence. His conscientious and proactive approach is unwavering, whether he is working independently or collaboratively within a team. His ability to connect with colleagues on a personal level underscores his commitment to fostering a harmonious and productive workplace environment.
Date: May 29, 2024
Tags: Information Security, ISO/IEC 27001, ISO/IEC 42001, Artificial Intelligence, GDPR
-------------------------------------------------------------------------------
Find out more about ISO training and certification services
Training: ISO/IEC 27001 Information Security Management System - EN | PECB
ISO/IEC 42001 Artificial Intelligence Management System - EN | PECB
General Data Protection Regulation (GDPR) - Training Courses - EN | PECB
Webinars: https://pecb.com/webinars
Article: https://pecb.com/article
-------------------------------------------------------------------------------
For more information about PECB:
Website: https://pecb.com/
LinkedIn: https://www.linkedin.com/company/pecb/
Facebook: https://www.facebook.com/PECBInternational/
Slideshare: http://www.slideshare.net/PECBCERTIFICATION
How to Make a Field Mandatory in Odoo 17Celine George
In Odoo, making a field required can be done through both Python code and XML views. When you set the required attribute to True in Python code, it makes the field required across all views where it's used. Conversely, when you set the required attribute in XML views, it makes the field required only in the context of that particular view.
How to Setup Warehouse & Location in Odoo 17 InventoryCeline George
In this slide, we'll explore how to set up warehouses and locations in Odoo 17 Inventory. This will help us manage our stock effectively, track inventory levels, and streamline warehouse operations.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
This document provides an overview of wound healing, its functions, stages, mechanisms, factors affecting it, and complications.
A wound is a break in the integrity of the skin or tissues, which may be associated with disruption of the structure and function.
Healing is the body’s response to injury in an attempt to restore normal structure and functions.
Healing can occur in two ways: Regeneration and Repair
There are 4 phases of wound healing: hemostasis, inflammation, proliferation, and remodeling. This document also describes the mechanism of wound healing. Factors that affect healing include infection, uncontrolled diabetes, poor nutrition, age, anemia, the presence of foreign bodies, etc.
Complications of wound healing like infection, hyperpigmentation of scar, contractures, and keloid formation.
Jena Hwang - Double Trouble: The Problem of Construal in Semantic Annotation of Adpositions
1. August 4, 2017, *SEM, Vancouver
DoubleTrouble:The Problem of Construal
in Semantic Annotation of Adpositions
Jena Hwang
Nathan SchneiderTim O’GormanArchna Bhatia
Na-Rae Han Vivek Srikumar
2. Most languages have
adpositions.
adposition = preposition
| postposition
2
in on at by
for to of with from
about …
kā ko ne se
mẽ par tak …
(n)eun i/ga,
do, (r)eul …
bə- lə- mi-
‘al ‘im …
3. 3
Feature 85A: Order of Adposition and Noun Phrase
Dryer in WALS, http://wals.info/chapter/85
4. We know PPs are challenging
for syntactic parsing.
a talk at the workshop on prepositions
4
But what about the meaning beyond
linking governor & modifier?
11. Shared functions
11
They ran to the roof for a quick escape.
They made for the roof to escape the cops.
DESTINATION PURPOSE
12. Design Principles
1. Coverage: Wicked polysemy, rare senses make
it hard to annotate all tokens in a corpus.
2. Cross-linguistic adequacy: Adpositions/case
markers work differently in different languages.
Ideally, our semantic functions should be
language-independent.
12
13. Design Principles
1. Coverage: Annotate all adposition types and
tokens in a corpus.
2. Cross-linguistic adequacy: Adpositions/case
markers work differently in different languages.
Ideally, our semantic functions should be
language-independent.
13
14. Design Principles
1. Coverage: Annotate all adposition types and
tokens in a corpus.
2. Cross-linguistic adequacy: Our semantic
functions should be as language-independent
as possible.
14
15. Senses vs. Supersenses
15
fine-grained details
lexeme-specific
125The case of over
1.
Protoscene
2.
A-B-C
trajectory
cluster
3.
Covering
4.
Examining
5.
Up cluster
6.
Reflexive
6.A
Repetition
5.A
More
5.A.1
Over-and-above
(excess II)
5.B
Control
5.C
Preference
4.A
Focus-of-
attention
2.A
On-the-
other-side-
of
2.B
Above-and-
beyond
(excess I)
2.C
Completion
2.D
Transfer
ations is potentially recursive and that a distinct sense can be the result of
multiple instances of reanalysis. Moreover, we believe that a complex concep-
tualization, such as the one represented in Figure 5, can be submitted to
multiple reanalyses and thus give rise to several distinct senses. When a
complex conceptualization gives rise to multiple senses, we term the set of
senses a “cluster of senses”. A cluster of senses is denoted in our represen-
tation of a semantic network by an open circle. A single distinct sense is
represented by a dark sphere.
Figure 6. The semantic network for over.
(extensive linguistic & AI research
on space & time)
16. Senses vs. Supersenses
15
fine-grained details
lexeme-specific
125The case of over
1.
Protoscene
2.
A-B-C
trajectory
cluster
3.
Covering
4.
Examining
5.
Up cluster
6.
Reflexive
6.A
Repetition
5.A
More
5.A.1
Over-and-above
(excess II)
5.B
Control
5.C
Preference
4.A
Focus-of-
attention
2.A
On-the-
other-side-
of
2.B
Above-and-
beyond
(excess I)
2.C
Completion
2.D
Transfer
ations is potentially recursive and that a distinct sense can be the result of
multiple instances of reanalysis. Moreover, we believe that a complex concep-
tualization, such as the one represented in Figure 5, can be submitted to
multiple reanalyses and thus give rise to several distinct senses. When a
complex conceptualization gives rise to multiple senses, we term the set of
senses a “cluster of senses”. A cluster of senses is denoted in our represen-
tation of a semantic network by an open circle. A single distinct sense is
represented by a dark sphere.
Figure 6. The semantic network for over.
(extensive linguistic & AI research
on space & time)
N = 4073
Neither
62%
Temporal
13%
Spatial
25%
17. Senses vs. Supersenses
15
fine-grained details
lexeme-specific
cross-lexical classes; coarse;
interpretable names like TOPIC
125The case of over
1.
Protoscene
2.
A-B-C
trajectory
cluster
3.
Covering
4.
Examining
5.
Up cluster
6.
Reflexive
6.A
Repetition
5.A
More
5.A.1
Over-and-above
(excess II)
5.B
Control
5.C
Preference
4.A
Focus-of-
attention
2.A
On-the-
other-side-
of
2.B
Above-and-
beyond
(excess I)
2.C
Completion
2.D
Transfer
ations is potentially recursive and that a distinct sense can be the result of
multiple instances of reanalysis. Moreover, we believe that a complex concep-
tualization, such as the one represented in Figure 5, can be submitted to
multiple reanalyses and thus give rise to several distinct senses. When a
complex conceptualization gives rise to multiple senses, we term the set of
senses a “cluster of senses”. A cluster of senses is denoted in our represen-
tation of a semantic network by an open circle. A single distinct sense is
represented by a dark sphere.
Figure 6. The semantic network for over.
(extensive linguistic & AI research
on space & time)
19. Supersense Hierarchy 1.0
17
StartState
Configuration
Circumstance
Temporal
Place
Whole Elements
Possessor
Species
Instance
Quantity
Superset
Causer
Agent
CreatorCo-Agent
Explanation Attribute
Manner
Reciprocation Purpose
Function
Age Time Frequency
Duration
RelativeTime
EndTimeStartTime ClockTimeCxn
DeicticTime
Path
Locus
Value
Comparison/Contrast
Scalar/Rank
ValueComparison
Approximator
Contour
Direction
Extent
Location Source State
Goal
InitialLocation
Material
Donor/Speaker
Destination
Recipient
EndState
Via
Traversed
1DTrajectory
2DArea 3DMedium
Transit
Instrument
Patient
Co-Patient
Activity
Means
Course
Accompanier
Beneficiary
Theme
Co-Theme Topic
ProfessionalAspect
Undergoer
Co-Participant
Affector
Participant
Experiencer Stimulus
75 preposition supersense categories http://tiny.cc/prepwiki
[LAW 2015]
20. English Annotation in
STREUSLE
• Online reviews corpus previously annotated for
multiword expressions and noun & verb
supersenses. 55,000 words, including 4,250 preps.
• Comprehensive annotation: first dataset with all
prepositions (types+tokens) semantically annotated
‣ Sentences not hand-selected
‣ Sentences fully annotated
‣ Preposition types not constrained by a lexicon (labels
generalize)
‣ All sentences seen by multiple annotators
18
[LAW 2016]
21. Comparing resources
19
P* {P1,P2} Ann P1 < P2 X ~
TPP ✓ (✓) ✓
The Preposition Project
(Litkowski & Hargraves 2005,
SemEval 2007 shared task)
D+ 7 ✓ ✓
TPP senses for 7 preposition
types in PropBank WSJ data
(Dahlmeier et al. 2009)
Tratz 34 (✓) ✓ ✓
Annotator-optimized revised
senses for 34 TPP SemEval
prepositions (Tratz 2011)
S&R 34 ✓
32 hard clusters of TPP senses
for 34 SemEval prepositions
(Srikumar & Roth 2013)
Ours ✓ ✓ ✓ ✓ ✓
Preposition supersenses
(Schneider et al. LAW 2015,
2016)
∞ ∞
P P
P P P
[LAW 2016]
22. A Vexing Problem
20
• Drawing clean boundaries between semantic
categories is always difficult.
• But we were surprised by the frequency of
apparent overlaps between semantic role labels.
• These overlaps proved pervasive in the other
languages we looked at.
23. Destination/Location
21
• The prepositions to, into, onto, and for explicitly encode
DESTINATION.
• DESTINATION masquerading as static LOCATION:
‣ Put the pen in the box. (= into)
‣ He threw his cards on the table. (= onto)
‣ The ball rolled behind the trash can.
• Extremely productive for motion/caused motion!
• We could stipulate one or the other, but annotators
would still get confused.
24. Fictive Motion
22
• In the other direction, we know that static locative
relations can be described using dynamic language
(Talmy 1996):
‣ The road runs through the trees.
‣ I heard him from the room next door.
‣ The school is around the corner.
• In assigning a semantic label, is it sufficient to
“choose sides” between the static nature of the
spatial scene, and the dynamic way that relation is
portrayed by the preposition?
25. Stimulus/Topic
23
• Another conundrum:
‣ I thought about getting my ears pierced.: TOPIC (cf. know, talk,
read)
‣ I feared getting my ears pierced: STIMULUS (cf. see, hurt)
‣ I was scared about getting my ears pierced: ???
• Again, two labels are competing for semantic territory.
• Should we add more categories with double inheritance?
(Problem: Proliferation of categories.)
• Should we just allow annotators to specify multiple labels if
they’re unsure? (Problem: Would create inconsistency.)
26. Construal
• Assumption thus far:
preposition token’s semantics = role in a scene
‣ I thought about getting my ears pierced.
• But it’s not always so simple:
‣ I was scared about getting my ears pierced.
24
…Topic
Topic
Topic
…Stimulus
27. Construal
• Observation: The preposition can frame or
construe the situation in a way that differs from
the predicate or scene.
• Solution: Allow tokens to receive two labels from
the hierarchy, one for the scene role and one for
the preposition’s semantic function, when
warranted.
25
28. Construal
• In fact, Stimulus can be interpreted differently
by different prepositions:
‣ I was scared by the bear.
•
‣ I was scared about getting my ears pierced.
26
Causer
Topic
…Stimulus
…Stimulus
29. Experiencer Dative
• Experiencers can be realized as recipients/datives:
‣ The bear felt scary to me.
• In some languages, this is the main way EXPERIENCERs
are realized:
‣ koev li ha-roš. [Hebrew]
Hurts to.me the-head ‘My head hurts.’
‣ mujh-ko garmii lag rahii hai. [Hindi]
I-DAT head feel PROG PRESS ‘I’m feeling hot.’
27
Recipient
…Experiencer
30. Employment
• The PROFESSIONALASPECT label is used for employer–
employee and other professional relationships.
• It participates in several different preposition construals:
‣ He works for XYZ Inc.
at
‣ He’s from XYZ Inc.
with
28
…ProfAsp
…ProfAsp
Beneficiary
Location
Source
Accompanier
31. Null Functions?
• Sometimes it’s hard to tell whether the
adposition has any semantic contribution:
‣ I’m angry with my mom.
*mad
‣ She’s interested in politics.
*fascinated
29
?…Stimulus
…Topic ?
32. Postposition or Conjunction?
• The Korean marker -wa can have a comitative (ACCOMPANIER)
meaning:
‣ cheolsunun youngmiwa gilul geoleotta
‘Cheolsu walked the streets with Youngmi’
‣ Cheolsunun youngmiwa chalul masyeotta
‘Cheolsu drank tea with Youngmi’
• But it can also mean ‘and’:
‣ keopiwa chalul masija
‘Let’s drink coffee and tea’
• Our semantic inventory is limited to figure–ground relations.
Would require labels for coordination semantics to cover -wa
where it means ‘and’.
30
36. Next Steps
• Annotation:
‣ Updating the English reviews corpus
‣ Monolingual Hebrew, Hindi, Korean data
‣ Parallel data (Little Prince)
• Questions:
‣ What construals are possible in what languages?
‣ Can separating scene role from function better account for
translation?
‣ How well can the role and function be predicted automatically?
34
37. Martha Palmer, Ken Litkowski, Omri Abend, Katie Conger,
Meredith Green, Michael Ellsworth, Paul Portner, Bill Croft
Thanks to
35
https://arnoldzwicky.org/2015/09/12/cartoon-adventures-in-lexical-semantics/