1) The document discusses various linguistic phenomena including irony, sarcasm, and thwarting. It presents algorithms for detecting sarcasm and thwarting in text.
2) For sarcasm detection, a semi-supervised algorithm uses pattern-based and punctuation-based features to classify sentences, achieving up to 81% accuracy.
3) Thwarting detection compares sentiment across levels of a domain ontology, using either rule-based or machine learning approaches, with the latter approach achieving up to 81% accuracy.
Most existing approaches to Twitter sentiment analysis assume that sentiment is explicitly expressed through affective words. Nevertheless, sentiment is often implicitly expressed via latent semantic relations, patterns and dependencies among words in tweets. In this paper, we propose a novel approach that automatically captures patterns of words of similar contextual semantics and sentiment in tweets. Unlike previous work on sentiment pattern extraction, our proposed approach does not rely on external and fixed sets of syntactical templates/patterns, nor requires deep analyses of the syntactic structure of sentences in tweets.
We evaluate our approach with tweet- and entity-level sentiment analysis tasks by using the extracted semantic patterns as classification features in both tasks. We use 9 Twitter datasets in our evaluation and compare the performance of our patterns against 6 state-of-the-art baselines. Results show that our patterns consistently outperform all other baselines on all datasets by 2.19% at the tweet-level and 7.5% at the entity-level in average F-measure.
Can Deep Learning solve the Sentiment Analysis ProblemMark Cieliebak
Sentiment analysis appears to be one of the easier tasks in the realm of text analytics: given a text like a tweet or product review, decide whether it contains positive or negative opinion. This task is almost trivial for humans, but it turns out to be a true challenge for automated systems. In fact, state-of-the-art sentiment analysis tools are wrong on approx. 4 out of 10 documents.
Current sentiment analysis tools are rule-based, feature-based, or combinations of both. However, recent research uses deep learning on very large sets of documents.
In this talk, we will explain the intrinsic difficulties of automated sentiment analysis; present existing solution approaches and their performance; describe an architecture for a deep learning system; and explore whether deep learning can improve sentiment analysis accuracy.
Most existing approaches to Twitter sentiment analysis assume that sentiment is explicitly expressed through affective words. Nevertheless, sentiment is often implicitly expressed via latent semantic relations, patterns and dependencies among words in tweets. In this paper, we propose a novel approach that automatically captures patterns of words of similar contextual semantics and sentiment in tweets. Unlike previous work on sentiment pattern extraction, our proposed approach does not rely on external and fixed sets of syntactical templates/patterns, nor requires deep analyses of the syntactic structure of sentences in tweets.
We evaluate our approach with tweet- and entity-level sentiment analysis tasks by using the extracted semantic patterns as classification features in both tasks. We use 9 Twitter datasets in our evaluation and compare the performance of our patterns against 6 state-of-the-art baselines. Results show that our patterns consistently outperform all other baselines on all datasets by 2.19% at the tweet-level and 7.5% at the entity-level in average F-measure.
Can Deep Learning solve the Sentiment Analysis ProblemMark Cieliebak
Sentiment analysis appears to be one of the easier tasks in the realm of text analytics: given a text like a tweet or product review, decide whether it contains positive or negative opinion. This task is almost trivial for humans, but it turns out to be a true challenge for automated systems. In fact, state-of-the-art sentiment analysis tools are wrong on approx. 4 out of 10 documents.
Current sentiment analysis tools are rule-based, feature-based, or combinations of both. However, recent research uses deep learning on very large sets of documents.
In this talk, we will explain the intrinsic difficulties of automated sentiment analysis; present existing solution approaches and their performance; describe an architecture for a deep learning system; and explore whether deep learning can improve sentiment analysis accuracy.
NLP Asignment Final Presentation [IIT-Bombay]Sagar Ahire
The final presentation I did with Lekha & Deepali for the Natural Language Processing assignments at IIT-Bombay.
Assignments included:
1: Spelling Correction
2: Part-of-speech Tagging
3: Metaphor Detection
The majority of current approaches attempt to detect the overall polarity of a sentence, paragraph, or text span, regardless of the entities mentioned (e.g., laptops, restaurants) and their aspects (e.g., battery, screen; food, service)
Our task was concerned with aspect based sentiment analysis (ABSA), where the goal was to identify the aspects of given target entities and the sentiment expressed towards each aspect.
Github code: https://github.com/AkshitaJha/IRE
Project Web Page: juhi-ghosh.github.io/IRE
Youtube: https://youtu.be/ksfcodFeVHg
This Power Point presentation will give you the basic guidelines as well the main and most important aspects to be considered when testing and evaluating Grammar among your students.
Lexicon-based approaches to Twitter sentiment analysis are gaining much popularity due to their simplicity, domain independence, and relatively good performance. These approaches rely on sentiment lexicons, where a collection of words are marked with fixed sentiment polarities. However, words' sentiment orientation (positive, neural, negative) and/or sentiment strengths could change depending on context and targeted entities. In this paper we present SentiCircle; a novel lexicon-based approach that takes into account the contextual and conceptual semantics of words when calculating their sentiment orientation and strength in Twitter. We evaluate our approach on three Twitter datasets using three different sentiment lexicons. Results show that our approach significantly outperforms two lexicon baselines. Results are competitive but inconclusive when comparing to state-of-art SentiStrength, and vary from one dataset to another. SentiCircle outperforms SentiStrength in accuracy on average, but falls marginally behind in F-measure.
Sarcasm is a peculiar form of sentiment expression, where the surface sentiment differs from the implied sentiment. The detection of sarcasm in social media platforms has been applied in the past mainly to textual utterances where lexical indicators (such as interjections and intensifiers), linguistic markers, and contextual information (such as user profiles, or past conversations) were used to detect the sar- castic tone. However, modern social media platforms allow to create multimodal messages where audiovisual content is integrated with the text, making the analysis of a mode in isolation partial.
We first study the relationship between the textual and visual aspects in multimodal posts from three major social media platforms, i.e., Instagram, Tumblr and Twitter, and we run a crowdsourcing task to quantify the extent to which images are perceived as necessary by human annotators.
Moreover, we propose two different computational frameworks to detect sarcasm that integrate the textual and visual modalities. Results show the positive effect of combining modalities for the detection of sarcasm across platforms and methods.
Paolo Rosso "On irony detection in social media"AINL Conferences
Каковы лингвистические паттерны, которым следуют пользователи социальных сетей, чтобы высказывать иронию в совсем коротких фразах? Лингвистические средства - такие как неоднозначность, непоследовательность, неожиданность эмоциональный контекст, гораздо более широкий, чем просто негативная или позитивная тональность - играют очень важную роль триггеров иронии. В иронических текстах буквальный смысл сообщения как правило отрицается, но формальные маркеры отрицания отсутствуют. Это делает задачу определения иронии очень сложной. В своем выступлении я опишу как ирония выражается в социальных сетях (Twitter, Amazon, Facebook и др.) и каково современное положение дел в автоматическом определении иронии. Определение иронии очень важно для таких задач анализа текста как определение тональности сообщения, извлечение мнений, или анализ репутаций, и существует определенный интерес исследовательского сообщества к этой теме. На конференции SemEval 2015 будет организована задача-соревнование по определению тональности фигуративного языка в Твиттере (Sentiment Analysis of Figurative Language in Twitter, http://alt.qcri.org/semeval2015/task11/). В конце я коснусь еще более сложной проблемы различения иронии, сатиры и сарказма, например: Если вам тяжело смеяться над собой, я буду счастлив сделать это за вас.
Feature Specific Sentiment Analysis for Product Reviews, Subhabrata Mukherjee and Pushpak Bhattacharyya, In Proceedings of the 13th International Conference on Intelligent Text Processing and Computational Intelligence (CICLING 2012), New Delhi, India, March, 2012 (http://www.cse.iitb.ac.in/~pb/papers/cicling12-feature-specific-sa.pdf)
NLP Asignment Final Presentation [IIT-Bombay]Sagar Ahire
The final presentation I did with Lekha & Deepali for the Natural Language Processing assignments at IIT-Bombay.
Assignments included:
1: Spelling Correction
2: Part-of-speech Tagging
3: Metaphor Detection
The majority of current approaches attempt to detect the overall polarity of a sentence, paragraph, or text span, regardless of the entities mentioned (e.g., laptops, restaurants) and their aspects (e.g., battery, screen; food, service)
Our task was concerned with aspect based sentiment analysis (ABSA), where the goal was to identify the aspects of given target entities and the sentiment expressed towards each aspect.
Github code: https://github.com/AkshitaJha/IRE
Project Web Page: juhi-ghosh.github.io/IRE
Youtube: https://youtu.be/ksfcodFeVHg
This Power Point presentation will give you the basic guidelines as well the main and most important aspects to be considered when testing and evaluating Grammar among your students.
Lexicon-based approaches to Twitter sentiment analysis are gaining much popularity due to their simplicity, domain independence, and relatively good performance. These approaches rely on sentiment lexicons, where a collection of words are marked with fixed sentiment polarities. However, words' sentiment orientation (positive, neural, negative) and/or sentiment strengths could change depending on context and targeted entities. In this paper we present SentiCircle; a novel lexicon-based approach that takes into account the contextual and conceptual semantics of words when calculating their sentiment orientation and strength in Twitter. We evaluate our approach on three Twitter datasets using three different sentiment lexicons. Results show that our approach significantly outperforms two lexicon baselines. Results are competitive but inconclusive when comparing to state-of-art SentiStrength, and vary from one dataset to another. SentiCircle outperforms SentiStrength in accuracy on average, but falls marginally behind in F-measure.
Sarcasm is a peculiar form of sentiment expression, where the surface sentiment differs from the implied sentiment. The detection of sarcasm in social media platforms has been applied in the past mainly to textual utterances where lexical indicators (such as interjections and intensifiers), linguistic markers, and contextual information (such as user profiles, or past conversations) were used to detect the sar- castic tone. However, modern social media platforms allow to create multimodal messages where audiovisual content is integrated with the text, making the analysis of a mode in isolation partial.
We first study the relationship between the textual and visual aspects in multimodal posts from three major social media platforms, i.e., Instagram, Tumblr and Twitter, and we run a crowdsourcing task to quantify the extent to which images are perceived as necessary by human annotators.
Moreover, we propose two different computational frameworks to detect sarcasm that integrate the textual and visual modalities. Results show the positive effect of combining modalities for the detection of sarcasm across platforms and methods.
Paolo Rosso "On irony detection in social media"AINL Conferences
Каковы лингвистические паттерны, которым следуют пользователи социальных сетей, чтобы высказывать иронию в совсем коротких фразах? Лингвистические средства - такие как неоднозначность, непоследовательность, неожиданность эмоциональный контекст, гораздо более широкий, чем просто негативная или позитивная тональность - играют очень важную роль триггеров иронии. В иронических текстах буквальный смысл сообщения как правило отрицается, но формальные маркеры отрицания отсутствуют. Это делает задачу определения иронии очень сложной. В своем выступлении я опишу как ирония выражается в социальных сетях (Twitter, Amazon, Facebook и др.) и каково современное положение дел в автоматическом определении иронии. Определение иронии очень важно для таких задач анализа текста как определение тональности сообщения, извлечение мнений, или анализ репутаций, и существует определенный интерес исследовательского сообщества к этой теме. На конференции SemEval 2015 будет организована задача-соревнование по определению тональности фигуративного языка в Твиттере (Sentiment Analysis of Figurative Language in Twitter, http://alt.qcri.org/semeval2015/task11/). В конце я коснусь еще более сложной проблемы различения иронии, сатиры и сарказма, например: Если вам тяжело смеяться над собой, я буду счастлив сделать это за вас.
Feature Specific Sentiment Analysis for Product Reviews, Subhabrata Mukherjee and Pushpak Bhattacharyya, In Proceedings of the 13th International Conference on Intelligent Text Processing and Computational Intelligence (CICLING 2012), New Delhi, India, March, 2012 (http://www.cse.iitb.ac.in/~pb/papers/cicling12-feature-specific-sa.pdf)
Sentiments Analysis using Python and nltk Ashwin Perti
The presentation contains about how to classify the sentiments or sentiment analysis. Especially there are positive or negative emotions. So to classify them we have used python language by taking the help of nltk package.
Seminar presentation made by me for the topic of 'Resources for Sentiment Analysis' at IIT Bombay. Includes a set of bonus slides for additional information which was not actually presented.
Intrinsic and Extrinsic Evaluations of Word EmbeddingsJinho Choi
In this paper, we first analyze the semantic composition of word embeddings by cross-referencing their clusters with the manual lexical database, WordNet. We then evaluate a variety of word embedding approaches by comparing their contributions to two NLP tasks. Our experiments show that the word embedding clusters give high correlations to the synonym and hyponym sets in WordNet, and give 0.88% and 0.17% absolute improvements in accuracy to named entity recognition and part-of-speech tagging, respectively.
HackYale - Natural Language Processing (Week 1)Nick Hathaway
Slides for a course I taught on Natural Language Processing covering corpus manipulation, word tokenization and text classification tasks using Python's popular Natural Language Toolkit.
Explore the power of Natural Language Processing (NLP) and Data Science in uncovering valuable insights from Flipkart product reviews. This presentation delves into the methodology, tools, and techniques used to analyze customer sentiments, identify trends, and extract actionable intelligence from a vast sea of textual data. From understanding customer preferences to improving product offerings, discover how NLP Data Science is revolutionizing the way businesses leverage consumer feedback on Flipkart. Visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Prepare all students for the rigorous demands of the Next Generation Assessments with one of Mastery Education's signature products, "Measuring Up Core Success"!
This panel discussion presents the experiences of several elementary and middle school ESL teachers as they work with students and colleagues to apply SFL and a genre-based pedagogy to language instruction. The discussion presents successes and challenges, strategies, students’ responses to the approach, collaboration experiences, and student performance data.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
UiPath Test Automation using UiPath Test Suite series, part 4DianaGray10
Welcome to UiPath Test Automation using UiPath Test Suite series part 4. In this session, we will cover Test Manager overview along with SAP heatmap.
The UiPath Test Manager overview with SAP heatmap webinar offers a concise yet comprehensive exploration of the role of a Test Manager within SAP environments, coupled with the utilization of heatmaps for effective testing strategies.
Participants will gain insights into the responsibilities, challenges, and best practices associated with test management in SAP projects. Additionally, the webinar delves into the significance of heatmaps as a visual aid for identifying testing priorities, areas of risk, and resource allocation within SAP landscapes. Through this session, attendees can expect to enhance their understanding of test management principles while learning practical approaches to optimize testing processes in SAP environments using heatmap visualization techniques
What will you get from this session?
1. Insights into SAP testing best practices
2. Heatmap utilization for testing
3. Optimization of testing processes
4. Demo
Topics covered:
Execution from the test manager
Orchestrator execution result
Defect reporting
SAP heatmap example with demo
Speaker:
Deepak Rai, Automation Practice Lead, Boundaryless Group and UiPath MVP
Pushing the limits of ePRTC: 100ns holdover for 100 daysAdtran
At WSTS 2024, Alon Stern explored the topic of parametric holdover and explained how recent research findings can be implemented in real-world PNT networks to achieve 100 nanoseconds of accuracy for up to 100 days.
Climate Impact of Software Testing at Nordic Testing DaysKari Kakkonen
My slides at Nordic Testing Days 6.6.2024
Climate impact / sustainability of software testing discussed on the talk. ICT and testing must carry their part of global responsibility to help with the climat warming. We can minimize the carbon footprint but we can also have a carbon handprint, a positive impact on the climate. Quality characteristics can be added with sustainability, and then measured continuously. Test environments can be used less, and in smaller scale and on demand. Test techniques can be used in optimizing or minimizing number of tests. Test automation can be used to speed up testing.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
The Art of the Pitch: WordPress Relationships and SalesLaura Byrne
Clients don’t know what they don’t know. What web solutions are right for them? How does WordPress come into the picture? How do you make sure you understand scope and timeline? What do you do if sometime changes?
All these questions and more will be explored as we talk about matching clients’ needs with what your agency offers without pulling teeth or pulling your hair out. Practical tips, and strategies for successful relationship building that leads to closing the deal.
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
4. Verbal Irony
“An irony is a figure of speech which implicitly
displays that the utterance situation was surrounded
by an ironic environment.”
There also exists Situational Irony
6. Reasons for Expectation to Fail
Expectation E is caused by an action A
1. E failed because A failed or cannot be
performed because of another action B
2. E failed because A was not performed
Expectation E is not caused by any action
3. E failed by an action B
4. E accidentally failed
Type 1 and 3
have victims
Sarcasm is
irony with
definite victims
and
counterfeited
emotions
7. Properties of an Ironic Environment
An utterance implicitly displays all the three conditions for
ironic environment when it:
1. Alludes to the speaker's expectation E
2. Intentionally violates one of pragmatic principles
3. Implies the speaker's emotional attitude toward the
failure of E
Irony is recognized if any 2 of these 3 are recognized.
Irony conveys the third unidentified property.
8. Allude to Speaker’s Expectation
Deepali baked a pizza to satisfy her hunger. She placed the pizza on the
table and in the meantime Sagar came and gobbled up the whole pizza.
Deepali said to Sagar:
a. I'm not hungry at all
b. Have you seen my pizza on the table?
c. I'll sleep well on a full stomach.
d. I'm really satisfied to eat the pizza.
e. Did you enjoy eating the pizza?
9. Violation of Pragmatic Principles
Sincerity
You make a statement you believe
You ask a question whose answer you don’t know
You offer advice which will benefit the receiver
You thank when you are really grateful
Propositional content
You thank for something that has been done for you
Preparatory condition for Offer
You offer something that you can really give
Maxim of relevance
Politeness principle
Maxim of quantity
10. Emotional Attitude
Tone and expressions
Interjections “Oh! The weather is so nice”
The context implies the emotional attitude of
the speaker
12. Sarcasm
The activity of saying or writing the opposite of
what you mean, or of speaking in a way
intended to make someone feel stupid or show
that you are angry (Macmillan English
Dictionary)
13. Sarcasm manifests in other ways...
● “Love the cover” (book)
● “Be sure to save your purchase receipt”
(Smart Phone)
● “Great idea, now try again with a real
product development team” (e-reader)
● “Where am I?” (GPS device)
14. The Algorithm: Overview
1. Training Set: Sentences manually assigned
scores 1 to 5 where five means clearly
sarcastic and one absence of sarcasm
2. Create feature vectors from the labelled
sentences
3. Use these feature vectors to build a model
and assign scores to unlabelled examples
15. Step 1: Preprocessing of Data
1. Replace each appearance of a
product/company/author by generalized
[product], [company], [author], etc.
2. Remove all HTML tags and special symbols
from review text.
16. Step 2: Creating Feature Vectors
Pattern Based Features:
1. Classify words into High Frequency Words (HFWs) and
Content Words (CWs)
All [product], [company] tags and punctuation marks are
HFWs.
2. A pattern is a sequence of HFWs with slots for CWs.
Example: “Garmin does not care about product quality or
customer support” has patterns “[company] does not CW
about CW CW” or “about CW CW or CW CW”, etc.
17. Pattern Matching
1: Exact Match
: Sparse Match - additional non-matching words can
be inserted between pattern components
: Incomplete Match - only n of N pattern
components appear in sentence, while some
non-matching words can be inserted in
between
18. Punctuation Based Features
●
●
●
●
●
Sentence length in words
Number of “!” characters
Number of “?” characters
Number of quotes
Number of capitalized/all capital words
Features are normalized to be in [0-1] by dividing them by
maximal observed value
19. Step 3: Data Enrichment
● For each sentence in the training set perform
a search engine query containing this
sentence
● Assign similar label to newly extracted
sentence.
20. Step 4: Classification
● Construct feature vectors for each sentence in the
training and test set
● Compute Euclidean Distance to each of matching
vectors in training set
Let ti i=1..k be the k vectors with lowest Euclidean Distance to v.
Then v is classified label l as follows:
Count(l) = Count of vectors in the training set with label l
Label(v) =
21. Star Sentiment Baseline
● From a set of negative reviews (with 1-3
stars) classify those sentences as sarcastic
with strong positive sentiment.
● Positive sentiment words can be eg. “great”,
“best”, “top”, etc.
24. Thwarting?
“The actors were good, the story was great, the
screenplay was a marvel of perfection and the
music was good too, but the movie couldn’t
hold my attention...”
25. Detecting Thwarting: The Big Picture
● Ascertain attributes of entity using ontology
● Find sentiment of each attribute in ontology
and the overall entity
● If there is a contrast, conclude thwarting has
occured
26. Building the Domain Ontology
1. Identify key features of domain from a
corpus
2. Arrange them in a hierarchy
Notes:
● Very human-intensive
● One-time requirement
29. Rule-based Approach
1. Get dependency parse for adjective-noun
dependencies
2. Identify polarities towards all nouns
3. Tag corresponding ontology nodes with
found polarities
4. If a contradiction across levels is found,
conclude that thwarting has taken place
30. Rule-based Approach: Example
Movie
negative
Story Elements
positive
Main Story
positive
Dialogues
positive
Acting
positive
Characters
positive
Music
positive
Songs
positive
Background
Score
negative
32. Learning Weights: Choices
1. Choices for loss function:
a. Linear loss
b. Hinge loss
2. Choices for percolation across ontology
levels:
a. Complete percolation
b. Controlled percolation
33. Classification: Features
● Convert document into a feature vector.
● Examples:
○
○
○
○
○
Document polarity
No of flips of sign
Longest contiguous subsequence of +ve values
Longest contiguous subsequence of -ve values
etc.
35. What’s the catch?
Requires sentiment as input!
Document with
Sentiment Information
Document
Thwarted or Not Thwarted
Current System
Thwarted or Not Thwarted,
Document Sentiment
Ideal System
36. Key Ideas
● Irony indicates presence of an ironic environment,
with 3 properties
● 2 of those 3 are enough to recognize irony
● Sarcasm is irony with victims and counterfeited
emotions
● A semi supervised pattern based algorithm detects
sarcasm well
● Thwarting is the phenomenon of polarity reversal at
a higher level of ontology compared to the polarity
expressed at the lower level
● Rule based and machine learning based
approaches have been attempted for thwarting
37. References
● Akira Utsumi (1996) - A unified theory of irony and its
computational formalization. InCOLING, 962–967.
● Oren Tsur, Dmitry Davidov, Ari Rappoport (2010) ICWSM – A Great Catchy Name: Semi-Supervised
Recognition of Sarcastic Sentences in Online Product
Reviews. In Association for the advancement of Artificial
Intelligence
● Ankit Ramteke, Akshat Malu, Pushpak Bhattacharyya,
J. Saketha Nath (2013) - Detecting Turnarounds in
Sentiment Analysis: Thwarting. In ACL 2013.