SlideShare a Scribd company logo
Temporal Reasoning Task
San Kim
2021.06.30
1. “Going on a vacation” tasks longer than “Going for a walk”: A Study of Temporal
Commonsense Understanding (EMNLP 19) – MCTACO
2. TORQUE: A Reading Comprehension Dataset of Temporal Ordering Questions (EMNLP 20)
3. Temporal Reasoning on Implicit Events from Distant Supervision (NAACL 21)-TRACIE
MCTACO (Multiple choice temporal common-sense)
Temporal commonsense
Given two events “going on a vacation” and “going for a walk,” most humans would know that a
vacation is typically longer and occurs less often than a walk, but it is still challenging for computers
to understand and reason about temporal commonsense.
5 temporal properties
• Duration (how long an event takes)
• Temporal ordering (typical order of events)
• Typical time (when an event happens)
• Frequency (how often an event occurs)
• Stationarity (whether a state holds for a very long time or
indefinitely)
MCTACO
• MCTACO is comprised of 13k tuples, in the form of (sentence, question, candidate answer).
• The sentences in those tuples are randomly selected from MultiRC
• Collect questions and candidate answers(both correct question and wrong ones) using AMT.
• To ensure the quality of the results, they limit the annotations to native speakers and use
qualification tryouts.
• Step1. Question Generation
• Should ask about one of the five temporal phenomena
the defined earlier
• Should not be solved simply by a word or phrase from
the original sentence
• They also require crowd-sourcers to provide a correct
answer for each of their questions(correct and
incorrect answers)
• Step2. Question verification
• The ask another two crowdsourcers to
check the questions generated in Step 1,
(a) whether the two requirements are
satisfied and (b) whether the question is
grammatically and logically correct.
• For valid questions, they continue to ask
crowdsourcers to give one correct answer
and one incorrect answer
MCTACO
• For those candidates that represent
events, the previously-mentioned token-
level perturbations rarely lead to
interesting and diverse set of candidate
answers. It may lead to invalid phrases
(e.g., “he left the house”  “he walked
the house”.) Therefore, to perturb such
candidates, they create a pool of 60k
event phrases using PropBank and
perturb the candidate answers to be the
most similar ones extracted by an
information retrieval(IR) system.
• Expand the candidate answer set to 20
candidates per question.
• Step3. Candidate answer expansion
• Until this stage, they have collected a small
set of candidate answers (3 positive and 2
negative) for each question.
• Automatically expand this set in three ways
• Use a set of rules to extract numbers
and quantities (“2”, “once”) and temporal
terms (e.g. “a.m.”, “1990”, “afternoon”,
“day”), and then randomly perturb them
based on a list of temporal
units(“second”), adjectives (“early”),
points (“a.m.”) and adverbs (“always”).
( “2 a.m.”  “3 p.m.”, “1 day”  “10
days”, “once a week”  “twice a month”)
• Mask each individual token in a
candidate answer (one at a time) and
use BERT to predict replacements for
each missing term; they rank those
predictions by the confidence level of
BERT and keep the top three.
• Step4. Answer labeling
• Each (sentence, question, answer) tuple
produced earlier is labeled by 4
crowdsourcers, with three options: “likely”,
“unlikely”, or “invalid”.
• A tuple is kept only if all 4 annotators
agree on “likely” or “unlikely”.
MCTACO
TORQUE
• Time is important for understanding events and stories
described in natural language text.
• “he won the championship yesterday” is different from “he
will win the championship tomorrow” (explicit)
• If we read that a woman is “expecting the birth of her first
child”, we know that the birth is in the future, while if she
is “mourning the death of her mother”, the death is in the
past. (implicit)
• These relationships between an event and a time point
(e.g. “won the championship yesterday”) or between two
events (e.g., “expecting” is before “birth” and “mourning”
is after “death”) are called temporal relations.
TORQUE
• Challenges of RC for temporal relationships
1. Reading comprehension works rarely require event understanding. Most datasets largely only
require an understanding of predicates and arguments, and would ask questions like “what was a
woman trapped in?”. But a temporal relation question would be “what started before a woman was
trapped?” To answer it, the system needs to identify events (e.g., LANDSLIDE is an event and “body”
is not), the time of these events (e.g., LANDSLIDE is correct answer, while SAID is not because of
the time when the two events happen), and look at the entire passage rather than the local
predicate-argument structures within a sentence (e.g., SNOW and RAINFALL are correct answers to
the question above).
2. There are many events in a typical passage of text, so tempral relation questions typically query
more than one relationship at the same time. This means that a question can have multiple
answers (e.g., “what happened after the landslide?”), or no answers, because the question may be
beyond the time scope (e.g., “what happened before the snow started?”)
TORQUE
3. Temporal relations queried by natural language questions are often sensitive to a few key words
such as before, after, and start. Those questions can easily be changed to make contrasting
questions with dramatically different answers. Those questions can easily be changed to make
contrasting questions with dramatically different answers. Models that are not sensitive to these small
changes in question words will perform poorly on this task.
landslide
searching, said, found
searching
No answers
Causing, disruption, brining,
flooding, searching, trapped,
landslide, said, found
Landslide, trapped, found, said,
disruption, flooding
Landslide, trapped
said
No answers
TORQUE
• Annotate 3.2k text snippets randomly selected from the TempEval3 dataset.
• TORQUE has 25k events and 21k user-generated and fully answered temporal relation questions.
• RoBERTa-large achieves 51% in exact match on TORQUE after fine-tuning, about 30% behind
human performance.
• Generally speaking, an event involves a predicate and its arguments.
• When studying time, events were defined as actions/states triggered by verbs, adjectives, and
nominals.
• This work follows this line of event definition and uses event and event trigger interchangeably.
• Define an event to be either a verb or a noun.
• In copular constructions, they choose to label the verb as the event, instead of an adjective or
preposition. (for consistent treatment of “she was on the east coast yesterday” and “she was happy”
– easily teach to crowd workers) (Note that from the perspective of data collection, labeling the
copula does not lose information as one can always do post-processing using dependency parsing
or semantic role labeling to recover the connection between “was” and “happy”.)
Events
TORQUE
• Events expressed in text are not always factual. They
can be negated, uncertain, hypothetical or have
associated modalities.
• Prior work dealing with events often tried to
categorize and label these various aspects because
they were crucial for determining temporal relation.
• Simply have people label all events, irrespective of
their modality, and use natural language to describe
relations between them.
Events
TORQUE
Temporal Relations
• The relationship between two events with respect to time, or
between one event and a fixed time point.
• (A, r, B) – A and B are events or time points, and r is a
temporal relation. (e.g. (HAD, happened before, SLEPT) – first
sentence in Fig. 3)
• In previous works, every event is assumed to be associated
with a time interval. When comparing two events, there are
13 possible relation labels.
• There are still many relations that cannot be expressed
because the assumption that every event has a time interval
is inaccurate: The time scope of an event may be fuzzy, an
event can have a non-factual modality, or events can be
repetitive and invoke multiple intervals.
• To better handle these phenomena, they use natural
language to annotate the relationships between events.
TORQUE
Natural Language Annotation of Temporal Relations
• (A, r, B): a temporal relation between two events
• (?, r, B) : a temporal relation question
• (?, happened before, SLEPT): natural language
expression  “what happened before a lion slept?”
• (A, r, B) holds, assuming for any deictic expression A
or B the time point when the passage was written,
and assuming that the passage is true.
TORQUE
Advantages of Natural Language Annotation
• DISRUPTION and FLOODING happened at about
the same time, but we do not know for sure which
one is earlier, so we have to choose vague.
• SNOW and DISRUPTION, we do not know which
one ends earlier and have to choose vague.
• The question-answer (QA) pairs can naturally
capture these fuzzy relations.
TORQUE
Advantages of Natural Language Annotation
• Natural language questions can conveniently
incorporate different modes of events.
• ▲ the relation between “having a meal”, and
“sleeping”
• If we could only choose one label, we must
choose before for all these relations, although
these relations are actually different.
•  a repetitive event may be a series of intervals
rather than a single one, and often before is
very different from before.
TORQUE
Advantages of Natural Language Annotation
• The format of natural language questions
bypasses the need for explicit annotation of
properties of events or other theories.
• The annotator naturally avoids event pairs that
do not have relations.
• “what happened after the service
industries are hardest hit?”
• “what happened after a passerby reported
the body?”
• “what was expected to happen when the
crisis hit America?”
• “what was supposed to happen after a
passerby called the police?”
• It still remains difficult to have a theory
explaining
•  why hit can compare to expected and crisis,
but not to gains.
TORQUE
Penalize Shortcuts by Contrast Sets
• An important problem in building datasets is to
avoid trivial solutions.
• Contrast questions: which slightly modify the
original questions, but dramatically change the
answers
• For an existing question (?, r, B) (e.g., “what
happened after he ate his breakfast?”)
• Keep using B and change r (e.g., “what
happened before/shortly after/… he ate his
breakfast?”)
• Modify it to ask about the start/end time
(e.g., “what happened after he started
eating his breakfast?” or “what would finish
after he ate his breakfast?”)
• Check that the answers to the new question
are different from the original one to avoid
trivial modifications (e.g., changing “what
happened” to “what occurred”)
TORQUE
Data Collection
• Passages that consist of two contiguous
sentences, as this is sufficient to capture the
vast majority of non-trivial temporal relations.
• Create a pool of 26k two-sentence passages
from the TempEval3 workshop (2.8k articles)
• 1. Label all the events
• 2. Repeatedly do the following
• (a) Ask a temporal relation question and
point out all the answers from the list of
events
• Modify the temporal relation to create
one or more new questions and answer
them.
Quality Control
• Qualification: crowd workers were trained and
tested on 3 capabilities: labeling events,
asking temporal relation questions, and
question answering. Crowd workers were
considered level-1 qualified if they could pass
the test within 3 attempts. (1/3 workers
passed the qualification.)
• Pilot: asked level-1 crowd workers to do a
small amount of the real task. They manually
checked the annotations and gave feedback
to them. Roughly 1 out of 3 pilot submissions
received a level-2 qualification. In the end,
there were 63 level-2 annotators, and 60 of
them actually worked on large-scale task.
• Validation: 20% of articles. 5 different level-2
annotators(include original annotator) validate
the event and answers. They intentionally
added noise to the original data for quality
control. They did not do additional validation
for the question because there is no bad
questions in a random sample of 100.
Quality Control
TORQUE
Cost
• 3 passages were presented.
• The crowd worker could decide to use some
or all of the.
• For each passage a worker decided to use,
they needed to label the vents, answer 3 hard-
coded warm-up questions, and them ask and
answer at least 12 questions (including
contrast questions). The final reward is a base
pay of $6 plus $0.5 for each extra question (up
to $4).
• Incentive
• (1) use fewer passages so that they can
do event labeling and warm-up questions
fewer times.
• (2) modify questions instead of asking
from scratch
• (3) ask extra questions in each job.
• In practice, crowd workers on average used 2
passages in each job.
• Validating the events in each passage and the
answers to a specific question both cost $0.1.
• In total, TORQUE cost $15k for an average of
$0.7/question.
statistics
• 3.2k passage annotations (~50 tokens/passage)
• 24.9k events (7.9 events/passage)
• 21.2k user-provided questions (half of them
were labeled by crowd workers as
modifications of existing ones)
• 94 / 200 questions querying about relations
that cannot be directly represented by the
previous single-interval-based labels.
TORQUE
statistics
TORQUE
Experiments
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
MATRES, ACL 18
TempEval1-3, TimeBank-Dense(TB-Dense), EventTimeCorpus
Before, after, equal, vague
Based on tart-points
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• When reading a story, a human can construct
a latent timeline about events’ start and end
times.
• The timeline not only contains the placements
of explicitly mentioned events (e.g., ride a
bicycle), but also accounts for implicit events
(e.g. Farrah was distracted so she looked away).
• The ability to construct such a timeline is
essential for understanding the causal
dynamics of a situation.
• Contributions
• A temporal relation dataset TRACIE
focusing on implicit events
• A distant supervision process for temporal
understanding of implicit events
• A reasoning model that makes end-time
comparisons using predictions of start-
time distances and durations
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Such tests in TRACIE take the form of multi-premise textual entailment (TE)
• Each TRACIE instance contains
• A context story (or premise) consisting of a sequence of explicit narrative events
• An implicit event in the form of a natural language phrase that is unmentioned but has
some role in the story
• An explicit event also in the form of a phrase
• A comparator of either {starts, ends}
• A temporal relation of either {before, after} that marks the relationship in the dimension
defined by the comparator between the implicit-event and the explicit-event
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Such tests in TRACIE take the form of multi-premise textual entailment (TE)
• Premise: context story
• Hypotheses: temporal queries about pair-wise relations between implicit and explicit events
• E.g. “avoids” is implicit-event, “starts” is the comparator, “removed” is explicit-event and “before”
is the temporal-relation.
• Flip the temporal-relation (i.e., “before” to “after” and vice versa) to create
negative(contradiction) instances.
• Use start times of explicit-events as reference points and compare the implicit-event’s start or
end time with them, according to the label definitions (Fig. 3)
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Randomly sample short stories from the
ROCStories dataset
• For each story, one annotator writes 5 implicit
event phrases that are not explicitly mentioned
by the given story, but are inferable and
relevant.
• Additionally rewrites two explicit events closest
to the implicit event’s start and end time,
respectively.
• Build two TRACIE instances (minus the
temporal-relation) per implicit event
Implicit Event Generation
Automatic Instance Generation
• Extract all verbs and relevant arguments with
its semantic role labeling model in AllenNLP
• Construct a pool of explicit events in the form
of short phrases (using verbs and their
arguments)
• Extract all verbs and relevant arguments with
its semantic role labeling model in AllenNLP
• Construct a pool of explicit events in the form
of short phrases (using verbs and their
arguments)
• For each implicit event, randomly select two
{explicit-event, comparator} pairs from the pool.
Label Collection
• For each of the 20 instances per story,
annotate the temporal-relation with four
different annotators.
• Majority agreement as the final label and filter
out unagreeable instances.
• Two authors additionally verify the instances
with ambiguous verbs(e.g., “have”) and
corrected 5% of the end-time instances.
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Distant Supervision
• Within-sentence Extraction
• Collect start time comparisons between pairs of events heuristically from free-text using
“before/after” keywords
• Use AllenNLP’s SRL model to process each input sentence and find verbs with a temporal
argument that starts with either “before” or “after”, and contains at least another verb.
• If there are multiple verbs in the temporal argument, take the one with the largest number
of tokens as arguments.
• 2.8M instances from Wikipedia dump(May 2020)
Pattern-Based Pre-Training
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Distant Supervision
• Cross-sentence Extraction
• The data collected from the within-sentence patterns
does not reveal the relative distance between two
start times.
• Finds direct temporal expressions of hours and dates.
• Because these temporal expressions(e.g., 2021-01-01)
are globally comparable, the compared events can be
anywhere in a document.
• This process collects more supervision signals about
time-point comparisons and their relative distance
on event pairs with trivial causal relation.
• Find exact temporal values by filling unmentioned
elements of a temporal expression with the nearest
previous mention (e.g., add “January to the expression
of “the 10th” in Fig. 4)
Pattern-Based Pre-Training (PTNTIME)
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Cross-sentence Extraction
• Construct supervision instances under the assumption
that the extracted temporal expressions describe the
start times of the associated verbs (e.g., went started
on January 1st )
• Represent the differences between the two start times
as one of seven coarse temporal units: {<=minutes,
hours, days, weeks, months, years, >= decades}
• Go to park is weeks before write review as shown in
Fig. 4
• Couple the specialized temporal pre-training data
described above with additional paragraphs that are used
to perform conventional language model pretraining using
the original denoising task (T5).
• Input sequences of event : [EventA] starts [Relation]
[EventB] . Story: [Paragraph] and output sequences of
answer: [Label] [Distance] . [paragraph]: non empty only
for cross-sentence extractions. [label] is either positive or
negative. [distance] is one of the 7 coarse temporal units
represented with a set of blank tokens [extra_id_N]
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• This model makes end-time comparisons by symbolically combining start time distance and
duration from separate predictions based on some of the components.
• Does not rely on explicit annotations on timepoints, but only relative comparisons between
them.
Symbolic Temporal Reasoning Model (SYSTIME)
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
Symbolic Temporal Reasoning Model (SYSTIME)
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
Duration estimation – pretrain sequence-to-sequence model
𝑟𝑒𝑛𝑑𝑠 𝑒1, 𝑒2 = 𝑏𝑒𝑓𝑜𝑟𝑒 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 + 𝑑𝑢𝑟 𝑒1 < 0
𝑟𝑒𝑛𝑑𝑠 𝑒1, 𝑒2 = 𝑎𝑓𝑡𝑒𝑟 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 + 𝑑𝑢𝑟 𝑒1 > 0
𝑟𝑠𝑡𝑎𝑟𝑡𝑠 𝑒1, 𝑒2 = 𝑏𝑒𝑓𝑜𝑟𝑒 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 < 0
𝑟𝑠𝑡𝑎𝑟𝑡𝑠 𝑒1, 𝑒2 = 𝑎𝑓𝑡𝑒𝑟 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 > 0
• Use duration data from TimeM (1M events and
duration values)
• Input sequence event: [Event] story: [Story]
• Output sequence answer: [Value]
• [Event] represents the tokens of an event with
the trigger verb marked by a special token to its
left
• [Story] represents down tokens from the context
• [Value] is one of the 7unit labels (i.e., {<= minuts,
hours, …})
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
Approximate dist() function using output from PTNTIME
• Input sequences of event : [EventA] starts
[Relation] [EventB] . Story: [Paragraph] and
output sequences of answer: [Label] [Distance] .
• [EventA]: the texture description of e1
• [EventB]: the texture description of e2
• [Paragraph] the context (premise)
• Fix [Relation] to be before.
• By taking the values of the vocabulary indices
corresponding to “positive” and “negative” from
the logits of [Label] and applying a softmax
operation, get P_before, P_after. P = [P_before,
P_after]
• Apply softmax to the logits of [Distance] over the
7words representing the temporal units to obtain
7 values that approximate the prob. of distance.
Place the 7 values in temporal units’ increasing
order in vector d. c = [0, 1, 2, 3, 4, 5, 6]
• To get the direction, apply the tanh function to
the difference between the prob. in p.
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• T5-Large for PTNTIME and the duration model.
• PTNTIME – 45k steps(1.4M instances), duration
model – 80k steps(2.6M instances)
• These pretrained weights in SYSTIME: SYSTIME
ZEROSHOT wich uses no TRACIE supervision.
• Story-wide exact match metric, which is the
percentage of stories with all its related
hypotheses answered correctly.
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Uniform-dist: in the i.i.d. training set, 70% of the examples with the comparator ends and relation after are
positive. – randomly remove instances from the majority classes
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
• Train and evaluate only the instances with a label
of either “before” or “after”, which accounts for
about 80% of all instances.
• OT-NS(original test, no story): train and test with
only the sentences containing the trigger verbs
• OT: train and test with the entire document as an
auxiliary input
• OT-MS(original test, minimal supervision): train
with 1.2k (6%) training instances
• PT(perturbed test): train with the complete
training set and test on a perturbed test set from
Evaluating Models’ Local Decision Boundaries via
Contrast Sets.
TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision

More Related Content

What's hot

Miniaturizing Space: Small-satellites
Miniaturizing Space: Small-satellitesMiniaturizing Space: Small-satellites
Miniaturizing Space: Small-satellites
X. Breogan COSTA
 
Iv defuzzification methods
Iv defuzzification methodsIv defuzzification methods
Iv defuzzification methods
kypameenendranathred
 
Lecture 4 Relationship between pixels
Lecture 4 Relationship between pixelsLecture 4 Relationship between pixels
Lecture 4 Relationship between pixels
VARUN KUMAR
 
Vector Quantization Vs Scalar Quantization
Vector Quantization Vs Scalar Quantization Vector Quantization Vs Scalar Quantization
Vector Quantization Vs Scalar Quantization
ManasiKaur
 
P vs NP
P vs NP P vs NP
P vs NP
Mikel Qafa
 
Computer Networks Module 1 - part 2.pdf
Computer Networks Module 1 - part 2.pdfComputer Networks Module 1 - part 2.pdf
Computer Networks Module 1 - part 2.pdf
ShanthalaKV
 
Frequency domain methods
Frequency domain methods Frequency domain methods
Frequency domain methods
thanhhoang2012
 
Aoa amortized analysis
Aoa amortized analysisAoa amortized analysis
Aoa amortized analysis
Salabat Khan
 
Quantum Computers
Quantum ComputersQuantum Computers
Quantum Computers
NAGUR SHAREEF SHAIK
 
Fuzzy Logic
Fuzzy LogicFuzzy Logic
DiscoRank: optimizing discoverability on SoundCloud
DiscoRank: optimizing discoverability on SoundCloudDiscoRank: optimizing discoverability on SoundCloud
DiscoRank: optimizing discoverability on SoundCloud
Amélie Anglade
 
fuzzy fuzzification and defuzzification
fuzzy fuzzification and defuzzificationfuzzy fuzzification and defuzzification
fuzzy fuzzification and defuzzification
Nourhan Selem Salm
 
Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
 Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S... Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
parmeet834
 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORK
ankita pandey
 
Brain Computer Interface (BCI)
Brain Computer Interface (BCI)Brain Computer Interface (BCI)
Brain Computer Interface (BCI)
Eyakub Sorkar
 
Vector quantization
Vector quantizationVector quantization
Vector quantization
Rajani Sharma
 

What's hot (16)

Miniaturizing Space: Small-satellites
Miniaturizing Space: Small-satellitesMiniaturizing Space: Small-satellites
Miniaturizing Space: Small-satellites
 
Iv defuzzification methods
Iv defuzzification methodsIv defuzzification methods
Iv defuzzification methods
 
Lecture 4 Relationship between pixels
Lecture 4 Relationship between pixelsLecture 4 Relationship between pixels
Lecture 4 Relationship between pixels
 
Vector Quantization Vs Scalar Quantization
Vector Quantization Vs Scalar Quantization Vector Quantization Vs Scalar Quantization
Vector Quantization Vs Scalar Quantization
 
P vs NP
P vs NP P vs NP
P vs NP
 
Computer Networks Module 1 - part 2.pdf
Computer Networks Module 1 - part 2.pdfComputer Networks Module 1 - part 2.pdf
Computer Networks Module 1 - part 2.pdf
 
Frequency domain methods
Frequency domain methods Frequency domain methods
Frequency domain methods
 
Aoa amortized analysis
Aoa amortized analysisAoa amortized analysis
Aoa amortized analysis
 
Quantum Computers
Quantum ComputersQuantum Computers
Quantum Computers
 
Fuzzy Logic
Fuzzy LogicFuzzy Logic
Fuzzy Logic
 
DiscoRank: optimizing discoverability on SoundCloud
DiscoRank: optimizing discoverability on SoundCloudDiscoRank: optimizing discoverability on SoundCloud
DiscoRank: optimizing discoverability on SoundCloud
 
fuzzy fuzzification and defuzzification
fuzzy fuzzification and defuzzificationfuzzy fuzzification and defuzzification
fuzzy fuzzification and defuzzification
 
Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
 Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S... Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
Introduction to the Theory of Computation, Winter 2003 A. Hevia and J. Mao S...
 
HOPFIELD NETWORK
HOPFIELD NETWORKHOPFIELD NETWORK
HOPFIELD NETWORK
 
Brain Computer Interface (BCI)
Brain Computer Interface (BCI)Brain Computer Interface (BCI)
Brain Computer Interface (BCI)
 
Vector quantization
Vector quantizationVector quantization
Vector quantization
 

Similar to Temporal reasoning task

Time: Structural alignment and retrospective duration estimates
Time: Structural alignment and retrospective duration estimatesTime: Structural alignment and retrospective duration estimates
Time: Structural alignment and retrospective duration estimates
John Dennis
 
PhD Research Proposal - Qualifying Exam
PhD Research Proposal - Qualifying ExamPhD Research Proposal - Qualifying Exam
PhD Research Proposal - Qualifying Exam
Paramita Mirza
 
Knowledge representation events in Artificial Intelligence.pptx
Knowledge representation events in Artificial Intelligence.pptxKnowledge representation events in Artificial Intelligence.pptx
Knowledge representation events in Artificial Intelligence.pptx
kitsenthilkumarcse
 
An Analysis of Causality between Events and its Relation to Temporal Information
An Analysis of Causality between Events and its Relation to Temporal InformationAn Analysis of Causality between Events and its Relation to Temporal Information
An Analysis of Causality between Events and its Relation to Temporal Information
Paramita Mirza
 
Lec12 semantic processing
Lec12 semantic processingLec12 semantic processing
Lec12 semantic processing
Manju Rajput
 
AI NOTES ppt 4.pdf
AI NOTES ppt 4.pdfAI NOTES ppt 4.pdf
AI NOTES ppt 4.pdf
ARMANVERMA7
 
NoSQL and Einstein's theory of relativity
NoSQL and Einstein's theory of relativityNoSQL and Einstein's theory of relativity
NoSQL and Einstein's theory of relativity
Lars Marius Garshol
 
NLP_guest_lecture.pdf
NLP_guest_lecture.pdfNLP_guest_lecture.pdf
NLP_guest_lecture.pdf
Soha82
 
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptxLecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
22004788
 
Finding Structure in Time NEURAL NETWORKS
Finding Structure in Time NEURAL NETWORKSFinding Structure in Time NEURAL NETWORKS
Finding Structure in Time NEURAL NETWORKS
ESCOM
 
The Complexity of Data: Computer Simulation and “Everyday” Social Science
The Complexity of Data: Computer Simulation and “Everyday” Social ScienceThe Complexity of Data: Computer Simulation and “Everyday” Social Science
The Complexity of Data: Computer Simulation and “Everyday” Social Science
Edmund Chattoe-Brown
 
Knowledge representation
Knowledge representationKnowledge representation
Knowledge representation
Md. Tanvir Masud
 
Monte Carlo
Monte CarloMonte Carlo
Monte Carlo
guestc990b6
 
Foundations of Knowledge Representation in Artificial Intelligence.pptx
Foundations of Knowledge Representation in Artificial Intelligence.pptxFoundations of Knowledge Representation in Artificial Intelligence.pptx
Foundations of Knowledge Representation in Artificial Intelligence.pptx
kitsenthilkumarcse
 
Debs 2010 context based computing tutorial
Debs 2010 context based computing tutorialDebs 2010 context based computing tutorial
Debs 2010 context based computing tutorial
Opher Etzion
 
Semantic Role Labeling
Semantic Role LabelingSemantic Role Labeling
Semantic Role Labeling
Marina Santini
 
Wither OWL
Wither OWLWither OWL
Wither OWL
James Hendler
 
On the nature of medial temporal lobe contributions to the constructive simul...
On the nature of medial temporal lobe contributions to the constructive simul...On the nature of medial temporal lobe contributions to the constructive simul...
On the nature of medial temporal lobe contributions to the constructive simul...
SPK División Gráfica Digital de Surpack S.A.
 
Event templates for Question answering
Event templates for Question answeringEvent templates for Question answering
Event templates for Question answering
Barbara Starr
 
Learning from Time-to-Event Data from Online Learning Contexts
Learning from Time-to-Event Data from Online Learning Contexts Learning from Time-to-Event Data from Online Learning Contexts
Learning from Time-to-Event Data from Online Learning Contexts
Shalin Hai-Jew
 

Similar to Temporal reasoning task (20)

Time: Structural alignment and retrospective duration estimates
Time: Structural alignment and retrospective duration estimatesTime: Structural alignment and retrospective duration estimates
Time: Structural alignment and retrospective duration estimates
 
PhD Research Proposal - Qualifying Exam
PhD Research Proposal - Qualifying ExamPhD Research Proposal - Qualifying Exam
PhD Research Proposal - Qualifying Exam
 
Knowledge representation events in Artificial Intelligence.pptx
Knowledge representation events in Artificial Intelligence.pptxKnowledge representation events in Artificial Intelligence.pptx
Knowledge representation events in Artificial Intelligence.pptx
 
An Analysis of Causality between Events and its Relation to Temporal Information
An Analysis of Causality between Events and its Relation to Temporal InformationAn Analysis of Causality between Events and its Relation to Temporal Information
An Analysis of Causality between Events and its Relation to Temporal Information
 
Lec12 semantic processing
Lec12 semantic processingLec12 semantic processing
Lec12 semantic processing
 
AI NOTES ppt 4.pdf
AI NOTES ppt 4.pdfAI NOTES ppt 4.pdf
AI NOTES ppt 4.pdf
 
NoSQL and Einstein's theory of relativity
NoSQL and Einstein's theory of relativityNoSQL and Einstein's theory of relativity
NoSQL and Einstein's theory of relativity
 
NLP_guest_lecture.pdf
NLP_guest_lecture.pdfNLP_guest_lecture.pdf
NLP_guest_lecture.pdf
 
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptxLecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
Lecture_6_-_Memory_LTM_W6 (2).pcccccccccccccccccccccccptx
 
Finding Structure in Time NEURAL NETWORKS
Finding Structure in Time NEURAL NETWORKSFinding Structure in Time NEURAL NETWORKS
Finding Structure in Time NEURAL NETWORKS
 
The Complexity of Data: Computer Simulation and “Everyday” Social Science
The Complexity of Data: Computer Simulation and “Everyday” Social ScienceThe Complexity of Data: Computer Simulation and “Everyday” Social Science
The Complexity of Data: Computer Simulation and “Everyday” Social Science
 
Knowledge representation
Knowledge representationKnowledge representation
Knowledge representation
 
Monte Carlo
Monte CarloMonte Carlo
Monte Carlo
 
Foundations of Knowledge Representation in Artificial Intelligence.pptx
Foundations of Knowledge Representation in Artificial Intelligence.pptxFoundations of Knowledge Representation in Artificial Intelligence.pptx
Foundations of Knowledge Representation in Artificial Intelligence.pptx
 
Debs 2010 context based computing tutorial
Debs 2010 context based computing tutorialDebs 2010 context based computing tutorial
Debs 2010 context based computing tutorial
 
Semantic Role Labeling
Semantic Role LabelingSemantic Role Labeling
Semantic Role Labeling
 
Wither OWL
Wither OWLWither OWL
Wither OWL
 
On the nature of medial temporal lobe contributions to the constructive simul...
On the nature of medial temporal lobe contributions to the constructive simul...On the nature of medial temporal lobe contributions to the constructive simul...
On the nature of medial temporal lobe contributions to the constructive simul...
 
Event templates for Question answering
Event templates for Question answeringEvent templates for Question answering
Event templates for Question answering
 
Learning from Time-to-Event Data from Online Learning Contexts
Learning from Time-to-Event Data from Online Learning Contexts Learning from Time-to-Event Data from Online Learning Contexts
Learning from Time-to-Event Data from Online Learning Contexts
 

More from San Kim

20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
San Kim
 
2023 EMNLP day_san.pptx
2023 EMNLP day_san.pptx2023 EMNLP day_san.pptx
2023 EMNLP day_san.pptx
San Kim
 
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptxLongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
San Kim
 
slide-acl2022-combined_san.pptx
slide-acl2022-combined_san.pptxslide-acl2022-combined_san.pptx
slide-acl2022-combined_san.pptx
San Kim
 
Compeition-Level Code Generation with AlphaCode.pptx
Compeition-Level Code Generation with AlphaCode.pptxCompeition-Level Code Generation with AlphaCode.pptx
Compeition-Level Code Generation with AlphaCode.pptx
San Kim
 
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
San Kim
 
AI2 day.pptx
AI2 day.pptxAI2 day.pptx
AI2 day.pptx
San Kim
 
Answering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrievalAnswering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrieval
San Kim
 
Measuring massive multitask language understanding
Measuring massive multitask language understandingMeasuring massive multitask language understanding
Measuring massive multitask language understanding
San Kim
 
Abductive commonsense reasoning
Abductive commonsense reasoningAbductive commonsense reasoning
Abductive commonsense reasoning
San Kim
 
Electra
ElectraElectra
Electra
San Kim
 
XLnet RoBERTa Reformer
XLnet RoBERTa ReformerXLnet RoBERTa Reformer
XLnet RoBERTa Reformer
San Kim
 
Transformer xl
Transformer xlTransformer xl
Transformer xl
San Kim
 
Face recognition v1
Face recognition v1Face recognition v1
Face recognition v1
San Kim
 
Gan seminar
Gan seminarGan seminar
Gan seminar
San Kim
 
Deep learning study 3
Deep learning study 3Deep learning study 3
Deep learning study 3
San Kim
 
Deep learning study 2
Deep learning study 2Deep learning study 2
Deep learning study 2
San Kim
 
Deep learning study 1
Deep learning study 1Deep learning study 1
Deep learning study 1
San Kim
 
Back propagation
Back propagationBack propagation
Back propagation
San Kim
 

More from San Kim (19)

20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
20230419-LLaMA-Adapter_ Efficient Fine-tuning of Language Models with Zero-in...
 
2023 EMNLP day_san.pptx
2023 EMNLP day_san.pptx2023 EMNLP day_san.pptx
2023 EMNLP day_san.pptx
 
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptxLongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
LongT5_Efficient Text-toText Transformer for Long Sequences_san.pptx
 
slide-acl2022-combined_san.pptx
slide-acl2022-combined_san.pptxslide-acl2022-combined_san.pptx
slide-acl2022-combined_san.pptx
 
Compeition-Level Code Generation with AlphaCode.pptx
Compeition-Level Code Generation with AlphaCode.pptxCompeition-Level Code Generation with AlphaCode.pptx
Compeition-Level Code Generation with AlphaCode.pptx
 
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
Intrinsic Dimensionality Explains the Effectiveness of Language Model Fine-Tu...
 
AI2 day.pptx
AI2 day.pptxAI2 day.pptx
AI2 day.pptx
 
Answering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrievalAnswering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrieval
 
Measuring massive multitask language understanding
Measuring massive multitask language understandingMeasuring massive multitask language understanding
Measuring massive multitask language understanding
 
Abductive commonsense reasoning
Abductive commonsense reasoningAbductive commonsense reasoning
Abductive commonsense reasoning
 
Electra
ElectraElectra
Electra
 
XLnet RoBERTa Reformer
XLnet RoBERTa ReformerXLnet RoBERTa Reformer
XLnet RoBERTa Reformer
 
Transformer xl
Transformer xlTransformer xl
Transformer xl
 
Face recognition v1
Face recognition v1Face recognition v1
Face recognition v1
 
Gan seminar
Gan seminarGan seminar
Gan seminar
 
Deep learning study 3
Deep learning study 3Deep learning study 3
Deep learning study 3
 
Deep learning study 2
Deep learning study 2Deep learning study 2
Deep learning study 2
 
Deep learning study 1
Deep learning study 1Deep learning study 1
Deep learning study 1
 
Back propagation
Back propagationBack propagation
Back propagation
 

Recently uploaded

Compexometric titration/Chelatorphy titration/chelating titration
Compexometric titration/Chelatorphy titration/chelating titrationCompexometric titration/Chelatorphy titration/chelating titration
Compexometric titration/Chelatorphy titration/chelating titration
Vandana Devesh Sharma
 
Cytokines and their role in immune regulation.pptx
Cytokines and their role in immune regulation.pptxCytokines and their role in immune regulation.pptx
Cytokines and their role in immune regulation.pptx
Hitesh Sikarwar
 
The binding of cosmological structures by massless topological defects
The binding of cosmological structures by massless topological defectsThe binding of cosmological structures by massless topological defects
The binding of cosmological structures by massless topological defects
Sérgio Sacani
 
Basics of crystallography, crystal systems, classes and different forms
Basics of crystallography, crystal systems, classes and different formsBasics of crystallography, crystal systems, classes and different forms
Basics of crystallography, crystal systems, classes and different forms
MaheshaNanjegowda
 
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
Sérgio Sacani
 
Micronuclei test.M.sc.zoology.fisheries.
Micronuclei test.M.sc.zoology.fisheries.Micronuclei test.M.sc.zoology.fisheries.
Micronuclei test.M.sc.zoology.fisheries.
Aditi Bajpai
 
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxThe use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
MAGOTI ERNEST
 
Sharlene Leurig - Enabling Onsite Water Use with Net Zero Water
Sharlene Leurig - Enabling Onsite Water Use with Net Zero WaterSharlene Leurig - Enabling Onsite Water Use with Net Zero Water
Sharlene Leurig - Enabling Onsite Water Use with Net Zero Water
Texas Alliance of Groundwater Districts
 
Oedema_types_causes_pathophysiology.pptx
Oedema_types_causes_pathophysiology.pptxOedema_types_causes_pathophysiology.pptx
Oedema_types_causes_pathophysiology.pptx
muralinath2
 
NuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyerNuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyer
pablovgd
 
Shallowest Oil Discovery of Turkiye.pptx
Shallowest Oil Discovery of Turkiye.pptxShallowest Oil Discovery of Turkiye.pptx
Shallowest Oil Discovery of Turkiye.pptx
Gokturk Mehmet Dilci
 
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
AbdullaAlAsif1
 
Medical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptxMedical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptx
terusbelajar5
 
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
University of Maribor
 
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
vluwdy49
 
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
Leonel Morgado
 
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
yqqaatn0
 
Thornton ESPP slides UK WW Network 4_6_24.pdf
Thornton ESPP slides UK WW Network 4_6_24.pdfThornton ESPP slides UK WW Network 4_6_24.pdf
Thornton ESPP slides UK WW Network 4_6_24.pdf
European Sustainable Phosphorus Platform
 
ESR spectroscopy in liquid food and beverages.pptx
ESR spectroscopy in liquid food and beverages.pptxESR spectroscopy in liquid food and beverages.pptx
ESR spectroscopy in liquid food and beverages.pptx
PRIYANKA PATEL
 
Phenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvementPhenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvement
IshaGoswami9
 

Recently uploaded (20)

Compexometric titration/Chelatorphy titration/chelating titration
Compexometric titration/Chelatorphy titration/chelating titrationCompexometric titration/Chelatorphy titration/chelating titration
Compexometric titration/Chelatorphy titration/chelating titration
 
Cytokines and their role in immune regulation.pptx
Cytokines and their role in immune regulation.pptxCytokines and their role in immune regulation.pptx
Cytokines and their role in immune regulation.pptx
 
The binding of cosmological structures by massless topological defects
The binding of cosmological structures by massless topological defectsThe binding of cosmological structures by massless topological defects
The binding of cosmological structures by massless topological defects
 
Basics of crystallography, crystal systems, classes and different forms
Basics of crystallography, crystal systems, classes and different formsBasics of crystallography, crystal systems, classes and different forms
Basics of crystallography, crystal systems, classes and different forms
 
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
EWOCS-I: The catalog of X-ray sources in Westerlund 1 from the Extended Weste...
 
Micronuclei test.M.sc.zoology.fisheries.
Micronuclei test.M.sc.zoology.fisheries.Micronuclei test.M.sc.zoology.fisheries.
Micronuclei test.M.sc.zoology.fisheries.
 
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptxThe use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
The use of Nauplii and metanauplii artemia in aquaculture (brine shrimp).pptx
 
Sharlene Leurig - Enabling Onsite Water Use with Net Zero Water
Sharlene Leurig - Enabling Onsite Water Use with Net Zero WaterSharlene Leurig - Enabling Onsite Water Use with Net Zero Water
Sharlene Leurig - Enabling Onsite Water Use with Net Zero Water
 
Oedema_types_causes_pathophysiology.pptx
Oedema_types_causes_pathophysiology.pptxOedema_types_causes_pathophysiology.pptx
Oedema_types_causes_pathophysiology.pptx
 
NuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyerNuGOweek 2024 Ghent programme overview flyer
NuGOweek 2024 Ghent programme overview flyer
 
Shallowest Oil Discovery of Turkiye.pptx
Shallowest Oil Discovery of Turkiye.pptxShallowest Oil Discovery of Turkiye.pptx
Shallowest Oil Discovery of Turkiye.pptx
 
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
Unlocking the mysteries of reproduction: Exploring fecundity and gonadosomati...
 
Medical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptxMedical Orthopedic PowerPoint Templates.pptx
Medical Orthopedic PowerPoint Templates.pptx
 
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
Remote Sensing and Computational, Evolutionary, Supercomputing, and Intellige...
 
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
在线办理(salfor毕业证书)索尔福德大学毕业证毕业完成信一模一样
 
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
Describing and Interpreting an Immersive Learning Case with the Immersion Cub...
 
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
原版制作(carleton毕业证书)卡尔顿大学毕业证硕士文凭原版一模一样
 
Thornton ESPP slides UK WW Network 4_6_24.pdf
Thornton ESPP slides UK WW Network 4_6_24.pdfThornton ESPP slides UK WW Network 4_6_24.pdf
Thornton ESPP slides UK WW Network 4_6_24.pdf
 
ESR spectroscopy in liquid food and beverages.pptx
ESR spectroscopy in liquid food and beverages.pptxESR spectroscopy in liquid food and beverages.pptx
ESR spectroscopy in liquid food and beverages.pptx
 
Phenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvementPhenomics assisted breeding in crop improvement
Phenomics assisted breeding in crop improvement
 

Temporal reasoning task

  • 1. Temporal Reasoning Task San Kim 2021.06.30 1. “Going on a vacation” tasks longer than “Going for a walk”: A Study of Temporal Commonsense Understanding (EMNLP 19) – MCTACO 2. TORQUE: A Reading Comprehension Dataset of Temporal Ordering Questions (EMNLP 20) 3. Temporal Reasoning on Implicit Events from Distant Supervision (NAACL 21)-TRACIE
  • 2. MCTACO (Multiple choice temporal common-sense) Temporal commonsense Given two events “going on a vacation” and “going for a walk,” most humans would know that a vacation is typically longer and occurs less often than a walk, but it is still challenging for computers to understand and reason about temporal commonsense. 5 temporal properties • Duration (how long an event takes) • Temporal ordering (typical order of events) • Typical time (when an event happens) • Frequency (how often an event occurs) • Stationarity (whether a state holds for a very long time or indefinitely)
  • 3. MCTACO • MCTACO is comprised of 13k tuples, in the form of (sentence, question, candidate answer). • The sentences in those tuples are randomly selected from MultiRC • Collect questions and candidate answers(both correct question and wrong ones) using AMT. • To ensure the quality of the results, they limit the annotations to native speakers and use qualification tryouts. • Step1. Question Generation • Should ask about one of the five temporal phenomena the defined earlier • Should not be solved simply by a word or phrase from the original sentence • They also require crowd-sourcers to provide a correct answer for each of their questions(correct and incorrect answers) • Step2. Question verification • The ask another two crowdsourcers to check the questions generated in Step 1, (a) whether the two requirements are satisfied and (b) whether the question is grammatically and logically correct. • For valid questions, they continue to ask crowdsourcers to give one correct answer and one incorrect answer
  • 4. MCTACO • For those candidates that represent events, the previously-mentioned token- level perturbations rarely lead to interesting and diverse set of candidate answers. It may lead to invalid phrases (e.g., “he left the house”  “he walked the house”.) Therefore, to perturb such candidates, they create a pool of 60k event phrases using PropBank and perturb the candidate answers to be the most similar ones extracted by an information retrieval(IR) system. • Expand the candidate answer set to 20 candidates per question. • Step3. Candidate answer expansion • Until this stage, they have collected a small set of candidate answers (3 positive and 2 negative) for each question. • Automatically expand this set in three ways • Use a set of rules to extract numbers and quantities (“2”, “once”) and temporal terms (e.g. “a.m.”, “1990”, “afternoon”, “day”), and then randomly perturb them based on a list of temporal units(“second”), adjectives (“early”), points (“a.m.”) and adverbs (“always”). ( “2 a.m.”  “3 p.m.”, “1 day”  “10 days”, “once a week”  “twice a month”) • Mask each individual token in a candidate answer (one at a time) and use BERT to predict replacements for each missing term; they rank those predictions by the confidence level of BERT and keep the top three. • Step4. Answer labeling • Each (sentence, question, answer) tuple produced earlier is labeled by 4 crowdsourcers, with three options: “likely”, “unlikely”, or “invalid”. • A tuple is kept only if all 4 annotators agree on “likely” or “unlikely”.
  • 6. TORQUE • Time is important for understanding events and stories described in natural language text. • “he won the championship yesterday” is different from “he will win the championship tomorrow” (explicit) • If we read that a woman is “expecting the birth of her first child”, we know that the birth is in the future, while if she is “mourning the death of her mother”, the death is in the past. (implicit) • These relationships between an event and a time point (e.g. “won the championship yesterday”) or between two events (e.g., “expecting” is before “birth” and “mourning” is after “death”) are called temporal relations.
  • 7. TORQUE • Challenges of RC for temporal relationships 1. Reading comprehension works rarely require event understanding. Most datasets largely only require an understanding of predicates and arguments, and would ask questions like “what was a woman trapped in?”. But a temporal relation question would be “what started before a woman was trapped?” To answer it, the system needs to identify events (e.g., LANDSLIDE is an event and “body” is not), the time of these events (e.g., LANDSLIDE is correct answer, while SAID is not because of the time when the two events happen), and look at the entire passage rather than the local predicate-argument structures within a sentence (e.g., SNOW and RAINFALL are correct answers to the question above). 2. There are many events in a typical passage of text, so tempral relation questions typically query more than one relationship at the same time. This means that a question can have multiple answers (e.g., “what happened after the landslide?”), or no answers, because the question may be beyond the time scope (e.g., “what happened before the snow started?”)
  • 8. TORQUE 3. Temporal relations queried by natural language questions are often sensitive to a few key words such as before, after, and start. Those questions can easily be changed to make contrasting questions with dramatically different answers. Those questions can easily be changed to make contrasting questions with dramatically different answers. Models that are not sensitive to these small changes in question words will perform poorly on this task. landslide searching, said, found searching No answers Causing, disruption, brining, flooding, searching, trapped, landslide, said, found Landslide, trapped, found, said, disruption, flooding Landslide, trapped said No answers
  • 9. TORQUE • Annotate 3.2k text snippets randomly selected from the TempEval3 dataset. • TORQUE has 25k events and 21k user-generated and fully answered temporal relation questions. • RoBERTa-large achieves 51% in exact match on TORQUE after fine-tuning, about 30% behind human performance. • Generally speaking, an event involves a predicate and its arguments. • When studying time, events were defined as actions/states triggered by verbs, adjectives, and nominals. • This work follows this line of event definition and uses event and event trigger interchangeably. • Define an event to be either a verb or a noun. • In copular constructions, they choose to label the verb as the event, instead of an adjective or preposition. (for consistent treatment of “she was on the east coast yesterday” and “she was happy” – easily teach to crowd workers) (Note that from the perspective of data collection, labeling the copula does not lose information as one can always do post-processing using dependency parsing or semantic role labeling to recover the connection between “was” and “happy”.) Events
  • 10. TORQUE • Events expressed in text are not always factual. They can be negated, uncertain, hypothetical or have associated modalities. • Prior work dealing with events often tried to categorize and label these various aspects because they were crucial for determining temporal relation. • Simply have people label all events, irrespective of their modality, and use natural language to describe relations between them. Events
  • 11. TORQUE Temporal Relations • The relationship between two events with respect to time, or between one event and a fixed time point. • (A, r, B) – A and B are events or time points, and r is a temporal relation. (e.g. (HAD, happened before, SLEPT) – first sentence in Fig. 3) • In previous works, every event is assumed to be associated with a time interval. When comparing two events, there are 13 possible relation labels. • There are still many relations that cannot be expressed because the assumption that every event has a time interval is inaccurate: The time scope of an event may be fuzzy, an event can have a non-factual modality, or events can be repetitive and invoke multiple intervals. • To better handle these phenomena, they use natural language to annotate the relationships between events.
  • 12. TORQUE Natural Language Annotation of Temporal Relations • (A, r, B): a temporal relation between two events • (?, r, B) : a temporal relation question • (?, happened before, SLEPT): natural language expression  “what happened before a lion slept?” • (A, r, B) holds, assuming for any deictic expression A or B the time point when the passage was written, and assuming that the passage is true.
  • 13. TORQUE Advantages of Natural Language Annotation • DISRUPTION and FLOODING happened at about the same time, but we do not know for sure which one is earlier, so we have to choose vague. • SNOW and DISRUPTION, we do not know which one ends earlier and have to choose vague. • The question-answer (QA) pairs can naturally capture these fuzzy relations.
  • 14. TORQUE Advantages of Natural Language Annotation • Natural language questions can conveniently incorporate different modes of events. • ▲ the relation between “having a meal”, and “sleeping” • If we could only choose one label, we must choose before for all these relations, although these relations are actually different. •  a repetitive event may be a series of intervals rather than a single one, and often before is very different from before.
  • 15. TORQUE Advantages of Natural Language Annotation • The format of natural language questions bypasses the need for explicit annotation of properties of events or other theories. • The annotator naturally avoids event pairs that do not have relations. • “what happened after the service industries are hardest hit?” • “what happened after a passerby reported the body?” • “what was expected to happen when the crisis hit America?” • “what was supposed to happen after a passerby called the police?” • It still remains difficult to have a theory explaining •  why hit can compare to expected and crisis, but not to gains.
  • 16. TORQUE Penalize Shortcuts by Contrast Sets • An important problem in building datasets is to avoid trivial solutions. • Contrast questions: which slightly modify the original questions, but dramatically change the answers • For an existing question (?, r, B) (e.g., “what happened after he ate his breakfast?”) • Keep using B and change r (e.g., “what happened before/shortly after/… he ate his breakfast?”) • Modify it to ask about the start/end time (e.g., “what happened after he started eating his breakfast?” or “what would finish after he ate his breakfast?”) • Check that the answers to the new question are different from the original one to avoid trivial modifications (e.g., changing “what happened” to “what occurred”)
  • 17. TORQUE Data Collection • Passages that consist of two contiguous sentences, as this is sufficient to capture the vast majority of non-trivial temporal relations. • Create a pool of 26k two-sentence passages from the TempEval3 workshop (2.8k articles) • 1. Label all the events • 2. Repeatedly do the following • (a) Ask a temporal relation question and point out all the answers from the list of events • Modify the temporal relation to create one or more new questions and answer them. Quality Control • Qualification: crowd workers were trained and tested on 3 capabilities: labeling events, asking temporal relation questions, and question answering. Crowd workers were considered level-1 qualified if they could pass the test within 3 attempts. (1/3 workers passed the qualification.) • Pilot: asked level-1 crowd workers to do a small amount of the real task. They manually checked the annotations and gave feedback to them. Roughly 1 out of 3 pilot submissions received a level-2 qualification. In the end, there were 63 level-2 annotators, and 60 of them actually worked on large-scale task. • Validation: 20% of articles. 5 different level-2 annotators(include original annotator) validate the event and answers. They intentionally added noise to the original data for quality control. They did not do additional validation for the question because there is no bad questions in a random sample of 100. Quality Control
  • 18. TORQUE Cost • 3 passages were presented. • The crowd worker could decide to use some or all of the. • For each passage a worker decided to use, they needed to label the vents, answer 3 hard- coded warm-up questions, and them ask and answer at least 12 questions (including contrast questions). The final reward is a base pay of $6 plus $0.5 for each extra question (up to $4). • Incentive • (1) use fewer passages so that they can do event labeling and warm-up questions fewer times. • (2) modify questions instead of asking from scratch • (3) ask extra questions in each job. • In practice, crowd workers on average used 2 passages in each job. • Validating the events in each passage and the answers to a specific question both cost $0.1. • In total, TORQUE cost $15k for an average of $0.7/question. statistics • 3.2k passage annotations (~50 tokens/passage) • 24.9k events (7.9 events/passage) • 21.2k user-provided questions (half of them were labeled by crowd workers as modifications of existing ones) • 94 / 200 questions querying about relations that cannot be directly represented by the previous single-interval-based labels.
  • 21. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision MATRES, ACL 18 TempEval1-3, TimeBank-Dense(TB-Dense), EventTimeCorpus Before, after, equal, vague Based on tart-points
  • 22. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision
  • 23. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • When reading a story, a human can construct a latent timeline about events’ start and end times. • The timeline not only contains the placements of explicitly mentioned events (e.g., ride a bicycle), but also accounts for implicit events (e.g. Farrah was distracted so she looked away). • The ability to construct such a timeline is essential for understanding the causal dynamics of a situation. • Contributions • A temporal relation dataset TRACIE focusing on implicit events • A distant supervision process for temporal understanding of implicit events • A reasoning model that makes end-time comparisons using predictions of start- time distances and durations
  • 24. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Such tests in TRACIE take the form of multi-premise textual entailment (TE) • Each TRACIE instance contains • A context story (or premise) consisting of a sequence of explicit narrative events • An implicit event in the form of a natural language phrase that is unmentioned but has some role in the story • An explicit event also in the form of a phrase • A comparator of either {starts, ends} • A temporal relation of either {before, after} that marks the relationship in the dimension defined by the comparator between the implicit-event and the explicit-event
  • 25. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Such tests in TRACIE take the form of multi-premise textual entailment (TE) • Premise: context story • Hypotheses: temporal queries about pair-wise relations between implicit and explicit events • E.g. “avoids” is implicit-event, “starts” is the comparator, “removed” is explicit-event and “before” is the temporal-relation. • Flip the temporal-relation (i.e., “before” to “after” and vice versa) to create negative(contradiction) instances. • Use start times of explicit-events as reference points and compare the implicit-event’s start or end time with them, according to the label definitions (Fig. 3)
  • 26. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Randomly sample short stories from the ROCStories dataset • For each story, one annotator writes 5 implicit event phrases that are not explicitly mentioned by the given story, but are inferable and relevant. • Additionally rewrites two explicit events closest to the implicit event’s start and end time, respectively. • Build two TRACIE instances (minus the temporal-relation) per implicit event Implicit Event Generation Automatic Instance Generation • Extract all verbs and relevant arguments with its semantic role labeling model in AllenNLP • Construct a pool of explicit events in the form of short phrases (using verbs and their arguments) • Extract all verbs and relevant arguments with its semantic role labeling model in AllenNLP • Construct a pool of explicit events in the form of short phrases (using verbs and their arguments) • For each implicit event, randomly select two {explicit-event, comparator} pairs from the pool. Label Collection • For each of the 20 instances per story, annotate the temporal-relation with four different annotators. • Majority agreement as the final label and filter out unagreeable instances. • Two authors additionally verify the instances with ambiguous verbs(e.g., “have”) and corrected 5% of the end-time instances.
  • 27. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Distant Supervision • Within-sentence Extraction • Collect start time comparisons between pairs of events heuristically from free-text using “before/after” keywords • Use AllenNLP’s SRL model to process each input sentence and find verbs with a temporal argument that starts with either “before” or “after”, and contains at least another verb. • If there are multiple verbs in the temporal argument, take the one with the largest number of tokens as arguments. • 2.8M instances from Wikipedia dump(May 2020) Pattern-Based Pre-Training
  • 28. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Distant Supervision • Cross-sentence Extraction • The data collected from the within-sentence patterns does not reveal the relative distance between two start times. • Finds direct temporal expressions of hours and dates. • Because these temporal expressions(e.g., 2021-01-01) are globally comparable, the compared events can be anywhere in a document. • This process collects more supervision signals about time-point comparisons and their relative distance on event pairs with trivial causal relation. • Find exact temporal values by filling unmentioned elements of a temporal expression with the nearest previous mention (e.g., add “January to the expression of “the 10th” in Fig. 4) Pattern-Based Pre-Training (PTNTIME)
  • 29. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Cross-sentence Extraction • Construct supervision instances under the assumption that the extracted temporal expressions describe the start times of the associated verbs (e.g., went started on January 1st ) • Represent the differences between the two start times as one of seven coarse temporal units: {<=minutes, hours, days, weeks, months, years, >= decades} • Go to park is weeks before write review as shown in Fig. 4 • Couple the specialized temporal pre-training data described above with additional paragraphs that are used to perform conventional language model pretraining using the original denoising task (T5). • Input sequences of event : [EventA] starts [Relation] [EventB] . Story: [Paragraph] and output sequences of answer: [Label] [Distance] . [paragraph]: non empty only for cross-sentence extractions. [label] is either positive or negative. [distance] is one of the 7 coarse temporal units represented with a set of blank tokens [extra_id_N]
  • 30. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • This model makes end-time comparisons by symbolically combining start time distance and duration from separate predictions based on some of the components. • Does not rely on explicit annotations on timepoints, but only relative comparisons between them. Symbolic Temporal Reasoning Model (SYSTIME)
  • 31. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision Symbolic Temporal Reasoning Model (SYSTIME)
  • 32. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision Duration estimation – pretrain sequence-to-sequence model 𝑟𝑒𝑛𝑑𝑠 𝑒1, 𝑒2 = 𝑏𝑒𝑓𝑜𝑟𝑒 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 + 𝑑𝑢𝑟 𝑒1 < 0 𝑟𝑒𝑛𝑑𝑠 𝑒1, 𝑒2 = 𝑎𝑓𝑡𝑒𝑟 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 + 𝑑𝑢𝑟 𝑒1 > 0 𝑟𝑠𝑡𝑎𝑟𝑡𝑠 𝑒1, 𝑒2 = 𝑏𝑒𝑓𝑜𝑟𝑒 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 < 0 𝑟𝑠𝑡𝑎𝑟𝑡𝑠 𝑒1, 𝑒2 = 𝑎𝑓𝑡𝑒𝑟 ⇔ 𝑑𝑖𝑠𝑡 𝑒1, 𝑒2 > 0 • Use duration data from TimeM (1M events and duration values) • Input sequence event: [Event] story: [Story] • Output sequence answer: [Value] • [Event] represents the tokens of an event with the trigger verb marked by a special token to its left • [Story] represents down tokens from the context • [Value] is one of the 7unit labels (i.e., {<= minuts, hours, …})
  • 33. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision Approximate dist() function using output from PTNTIME • Input sequences of event : [EventA] starts [Relation] [EventB] . Story: [Paragraph] and output sequences of answer: [Label] [Distance] . • [EventA]: the texture description of e1 • [EventB]: the texture description of e2 • [Paragraph] the context (premise) • Fix [Relation] to be before. • By taking the values of the vocabulary indices corresponding to “positive” and “negative” from the logits of [Label] and applying a softmax operation, get P_before, P_after. P = [P_before, P_after] • Apply softmax to the logits of [Distance] over the 7words representing the temporal units to obtain 7 values that approximate the prob. of distance. Place the 7 values in temporal units’ increasing order in vector d. c = [0, 1, 2, 3, 4, 5, 6] • To get the direction, apply the tanh function to the difference between the prob. in p.
  • 34. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • T5-Large for PTNTIME and the duration model. • PTNTIME – 45k steps(1.4M instances), duration model – 80k steps(2.6M instances) • These pretrained weights in SYSTIME: SYSTIME ZEROSHOT wich uses no TRACIE supervision. • Story-wide exact match metric, which is the percentage of stories with all its related hypotheses answered correctly.
  • 35. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Uniform-dist: in the i.i.d. training set, 70% of the examples with the comparator ends and relation after are positive. – randomly remove instances from the majority classes
  • 36. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision • Train and evaluate only the instances with a label of either “before” or “after”, which accounts for about 80% of all instances. • OT-NS(original test, no story): train and test with only the sentences containing the trigger verbs • OT: train and test with the entire document as an auxiliary input • OT-MS(original test, minimal supervision): train with 1.2k (6%) training instances • PT(perturbed test): train with the complete training set and test on a perturbed test set from Evaluating Models’ Local Decision Boundaries via Contrast Sets.
  • 37. TRACIE: Temporal Reasoning on Implicit Events from Distant Supervision