SlideShare a Scribd company logo
1 of 24
Download to read offline
FAKE REVIEW DETECTION
What are they saying about you ? Are they real …
Guided By : Dr. Animesh Mukherjee
ONLINE REVIEW
• Captures testimonials of “real” people (unlike advertisements).
• Shapes decision making of customers.
• Positive reviews – financial gains and fame for business.
• Deceptive opinion spamming – to promote or discredit some target product
and services.[1]
• Opinion Spammers admitted for being paid to write fake reviews.(Kost, 2012)
• Yelp.com – “Sting operation” : publically shame business who buy fake
reviews.[2]
[1] Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf
[2] Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-reviews.html
Consumer Alerts !!
• Amazon Mechanical Turk – crowd sourced online workers(turkers) to write fake
reviews ($1 per review) portraying 20 hotels of Chicago in positive light.[1]
• 400 fake positive reviews collected and 400 non-fake reviews on same 20 hotels
using Tripadvisor.com
• Yelp’s filtered and unfiltered reviews collected to understand working of Yelp
review classification algorithm.
• Approach : Linguistic n-gram features and some supervised learning method.[2]
[1] Amazon Mechanical Turk : https://www.mturk.com/
[2] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
Dataset Collection and Approach for Analysis
Linguistic Approach : Results
• Using only Bigram features :
accuracy of 89.6% on AMT data.[1]
• Using same n-gram features :
accuracy of 67.8% on Yelp Data.[2]
• Table 1: Class distribution of Yelp
data is skewed – imbalanced data
produces poor model.(Chawla et al,
2004)
• Good model for imbalanced data –
under sampling. (Drummond and
Holte, 2003)
[1] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
[2] A Mukherjee - 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Linguistic Approach Results : Explained
• For AMT data, word distribution of fake and non-fake reviews are very different,
which explains the high accuracy using n-gram.
• Reason for the different word distribution : Domain knowledge absence and little
gain in writing fake reviews($1 per review).
• Poor performance of n-grams for Yelp data because spammers according to
Yelp filter used very similar language in fake review as non-fake review :
linguistically very similar.[1]
• Inefficiency in linguistic in detecting fake reviews filtered by Yelp encourages to
do behavioral study of reviews.
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
))(/)((2log)()||( iNiFiFNFKL i
Information Theoretic Analysis
• To explain huge accuracy difference : analysis of word distribution of AMT and
Yelp data.
• Good-Turing smoothed unigram language models.
• Computation of word distribution difference across fake & non-fake reviews :
Kullback-Leibler(KL) Divergence :
where F(i) and N(i) are respective probabilities of word i in fake and non-fake reviews.[1]
• Here KL(F||N) gives quantitative estimate of how much fake reviews linguistically
differ from non-fake reviews
[1] Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
Information Theoretic Analysis (Cont..)
• KL Divergence is Asymmetric :
• We have,
• For AMT data : and
• For Yelp data : and
• [1]
   FNKLNFKL |||| 
   FNKLNFKL ||||  0KL
   FNKLNFKL ||||  0KL
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
   FNKLNFKLKL |||| 
Information Theoretic Analysis (Cont..)
• KL(F||N) Definition implies that words having higher probability in F and very low
probability in N contribute most to KL-Divergence.
• To study word-wise contribution to ∆KL, word ∆KL calculated as :
where,
• Contribution of top k words to ∆KL for k= 200 and k=300
)||()||( iiwordiiword
i
Word FNKLNFKLKL 
  






)(
)(
log)||( 2
iN
iF
iFNFKL iiword
Turkers didn’t do a good job at “Faking”!
Word-wise Difference of KL-
Divergence across top 200 words.
Equally Dense : |E| = |G|
• Symmetric Distribution of for top k-words for AMT
data implies the existence of two set of words:
1. set of words E, appearing more in fake reviews than
in non-fake, Ɐi € E, F(i) > N(i) resulting in
2. set of words G, appearing more in non-fake reviews
than in fake, Ɐi € G, N(i) > F(i) resulting in
i
WordKL
0 Ei
WordKL
0 Ei
WordKL
• Additionally top k=200 words only contribute 20% to ∆KL for the AMT data.
There are many words in AMT data having higher probabilities in fake than non-
fake and vice-versa.
• This implies fake and non-fake reviews in AMT data consist of words with very
different frequencies. Turkers didn’t do a good job in faking.[1]
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Yelp Spammers are Smart but Overdid “Faking”!
• Yelp fake review data shows KL(F||N) is much larger than KL(N||F) and ∆KL>1.
• Given graphs from b-e shows among top 200 words which contribute to major
(=70%) most words have and only a few have .
Word-wise Difference of KL-Divergence across top 200 words.[1]
0 i
WordKL 0 i
WordKL
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Yelp Spammers are Smart but Overdid “Faking”! (Cont..)
• Consider A be set of top words contributing most to ∆KL. We partition A as:
where and
• The curve above y=0 is dense and below it is sparse which implies
• clearly indicates that there exists specific words which contribute
most to ∆KL by appearing in fake reviews with much higher frequencies than
in non-fake reviews.
• Spammers made smart effort to ensure that their fake reviews have most
words that also appear in non-fake reviews to sound convincing.
 ))()(,.,(0| iFiNAiieKLiA Ni
Word
N

 ))()(,.,(0| iNiFAiieKLiA Fi
Word
F

NF
AAA   NF
AA
NF
AA 
NF
AA 
Yelp Spammers are Smart but Overdid “Faking”! (Cont..)
• While making their reviews sound convincing, psychologically they
happened to OVERUSE come words resulting higher frequency of certain
words in fake review than non-fake reviews.
• A quick lookup yields {us, price, stay, feel, deal, comfort} in hotel domain and
{options, went, seat, helpful, overall, serve, amount etc.} in restaurants.
• Prior personality work shown that deception/lying usually involves more use
of personal pronouns(eg. us) and associated action(eg. Went, feel) towards
specific targets(eg. option, price, stay) with objective of incorrect
projection(lying or faking).[1]
• Spammers caught by Yelp left behind linguistic footprints which can be
caught by precise behavioral study.
[1] Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-_predicting_deception_from_linguistic_styles.pdf
Spamming Behavior Analysis
1. Maximum Number of Reviews(MNR) : Writing too many reviews in a day is
abnormal.
2. Percentage of Positive Reviews(PR) : Deception words in fake reviews
indicates projection in positive light. CDF of positive(4-5 stars) reviews
among all reviews is plotted to illustrate analysis.
3. Review Length(RL) : While writing fake experiences, there is probably not
much to write and also spammer not want to spend too much time in it.
4. Maximum Content Similarity(MCS) : To examine if some posted reviews are
similar to previous reviews, we computed cosine similarity between two
reviews of a reviewer. Non-spammers mostly write new contents.
Maximum Number of Reviews Percentage of Positive Reviews
Review Length Maximum Content Similarity
Spamming Behavior Analysis CDFs
[1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
Challenges with Supervised Evaluation
• Very difficult to find gold-standard data of fake and non-fake reviews for
model building – too difficult to manually recognize/label fake/non-fake
reviews by mere reading.
• Duplicate and near duplicate assumed to be fake, which is unreliable.
• Usage of manually labeled dataset – reliability issues because it have been
shown that accuracy of human labeling of fake reviews is very poor.[1]
• AMT crowdsourcing fake reviews by paying are fake yet they don’t reflect the
dynamics of fake reviews in commercial website.[2]
• This lack of labeled data, motivates to look after the unsupervised methods of
classification.[3]
[1] Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination.
[2] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
[3] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
Unsupervised Evaluation Model
• Since human labeling for supervised learning is difficult, problem was
proposed by modeling spamicity (degree of spamming) as latent with other
observed behavioral features.
• Unsupervised model – Author Spamicity Model (ASM) proposed.[1]
• Taken a fully Bayesian approach and formulated opinion spam detection as
clustering problem.
• Opinion spammers have different behavioral distributions than non-spammers.
• Causes distributional divergence between latent population distributions of two
clusters: spammers and non-spammers.[1]
• Model inference results in learning the population distributions of two clusters.
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
• Formulates spam detection as an unsupervised clustering problem in
Bayesian setting.
• Belongs to class of Generative models for clustering based on set of
observed features.
• Models spamicity 𝑠 𝑎
(in range[0, 1]) of an author a; and spam label 𝜋 𝑟
of a
review which is the class variable reflecting the cluster membership (two
cluster K=2, spam and non-spam).[1]
• Each author/reviewer and respectively each review has a set of observed
features(behavioral clues).
• Certain characteristics of abnormal behavior defined which likely to link
with spamming and thus can be exploited in model for learning spam and
non-spam clusters. [1]
Author Spamicity Model
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
Author Features
• It have value in range [0, 1] and value close to 1 indicates spamming.
1. Content Similarity : Crafting new review every time is time consuming, spammers
likely to copy reviews across similar product. Choose maximum similarity to
capture worst spamming behavior.[1]
2. Maximum Number of Reviews : Posting many reviews a day is also abnormal.
3. Reviewing Burstiness : Spammers are usually not longtime members of site.
Defined over an activity window(first and last review posting date). If posted over
reasonably long timeframe, it probably a normal activity but all review posted
within short burst likely to be spam.[2]
4. Ratio of First Review : People mostly rely on early reviews and spamming early
impacts hugely on sales. Spammers try to be among first reviewers. [1]
[1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
[2] Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012).
Review Features
• It have five binary review features. Value 1 indicates spamming else 0 is non-spamming.
1. Duplicate /Near Duplicate Reviews: Spammers often post multiple reviews which are
duplicate /near duplicate on same product to boost ratings.
2. Extreme Rating: Spammers mostly like to give extreme ratings(1 or 5) in order to boost
ratings to demote or promote products.
3. Rating Deviation: Spammers usually involve in wrong projection in either positive or
negative and it deviates from average ratings given by other reviewers.
4. Early Time Frame: Early review can greatly impact people’s sentiments on a product.
5. Rating Abuse: Multiple ratings on same product are unusual. Similar to DUP but focuses
on rating dimensions rather than content.
CONCLUSIONS
• We presented an in-depth investigation of nature of fake reviews in Commercial
settings of Yelp.com.
• Our study shows linguistics methods of (Ott et al., 2011) and its high accuracy in
AMT data.
• We presented a behavioral study of spammers for real-life fake reviews.
• We presented a brief introduction to n-gram language model.
• We presented challenges with the supervised evaluation and gave idea about the
unsupervised approach of evaluation.
• We presented a brief introduction to unsupervised Author Spamicity Model (ASM).
REFERENCES
1. Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf
2. Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-
reviews.html
3. MIT N-gram Language Model Tutorial:
http://web.mit.edu/6.863/www/fall2012/readings/ngrampages.pdf
4. Amazon Mechanical Turk : https://www.mturk.com/
5. Ott et al. 2011 :
https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf
6. Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-
submit.pdf
7. Kullback Leibler Divergence :
http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
8. Newman et al , 2003 :
http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-
_predicting_deception_from_linguistic_styles.pdf
9. Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer
Reviews. WWW (2012).
10. Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
11. Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any
Stretch of the Imagination.
Any Question ?
Thank You !!

More Related Content

What's hot

CYBERBULLYING DETECTION USING MACHINE LEARNING-1 (1).pdf
CYBERBULLYING DETECTION USING              MACHINE LEARNING-1 (1).pdfCYBERBULLYING DETECTION USING              MACHINE LEARNING-1 (1).pdf
CYBERBULLYING DETECTION USING MACHINE LEARNING-1 (1).pdfKumbidiGaming
 
ATM System by image processing
ATM System by image processingATM System by image processing
ATM System by image processingshahab islam
 
Credit card fraud detection using machine learning Algorithms
Credit card fraud detection using machine learning AlgorithmsCredit card fraud detection using machine learning Algorithms
Credit card fraud detection using machine learning Algorithmsankit panigrahy
 
Fraud detection with Machine Learning
Fraud detection with Machine LearningFraud detection with Machine Learning
Fraud detection with Machine LearningScaleway
 
BIOMETRICS FINGER PRINT TECHNOLOGY
BIOMETRICS FINGER PRINT TECHNOLOGYBIOMETRICS FINGER PRINT TECHNOLOGY
BIOMETRICS FINGER PRINT TECHNOLOGYsathish sak
 
Detecting Fake News Through NLP
Detecting Fake News Through NLPDetecting Fake News Through NLP
Detecting Fake News Through NLPSakha Global
 
Sentiment Analysis of Twitter Data
Sentiment Analysis of Twitter DataSentiment Analysis of Twitter Data
Sentiment Analysis of Twitter DataSumit Raj
 
Detection of phishing websites
Detection of phishing websitesDetection of phishing websites
Detection of phishing websitesm srikanth
 
Fake Currency detction Using Image Processing
Fake Currency detction Using Image ProcessingFake Currency detction Using Image Processing
Fake Currency detction Using Image ProcessingSavitaHanchinal
 
Wine Quality Analysis Using Machine Learning
Wine Quality Analysis Using Machine LearningWine Quality Analysis Using Machine Learning
Wine Quality Analysis Using Machine LearningMahima -
 
Fingerprint Authentication for ATM
Fingerprint Authentication for ATMFingerprint Authentication for ATM
Fingerprint Authentication for ATMParas Garg
 
Speech recognition final presentation
Speech recognition final presentationSpeech recognition final presentation
Speech recognition final presentationhimanshubhatti
 
final presentation fake news detection.pptx
final presentation fake news detection.pptxfinal presentation fake news detection.pptx
final presentation fake news detection.pptxRudraSaraswat6
 
Machine learning seminar ppt
Machine learning seminar pptMachine learning seminar ppt
Machine learning seminar pptRAHUL DANGWAL
 

What's hot (20)

CYBERBULLYING DETECTION USING MACHINE LEARNING-1 (1).pdf
CYBERBULLYING DETECTION USING              MACHINE LEARNING-1 (1).pdfCYBERBULLYING DETECTION USING              MACHINE LEARNING-1 (1).pdf
CYBERBULLYING DETECTION USING MACHINE LEARNING-1 (1).pdf
 
ATM System by image processing
ATM System by image processingATM System by image processing
ATM System by image processing
 
Credit card fraud detection using machine learning Algorithms
Credit card fraud detection using machine learning AlgorithmsCredit card fraud detection using machine learning Algorithms
Credit card fraud detection using machine learning Algorithms
 
Fraud detection with Machine Learning
Fraud detection with Machine LearningFraud detection with Machine Learning
Fraud detection with Machine Learning
 
BIOMETRICS FINGER PRINT TECHNOLOGY
BIOMETRICS FINGER PRINT TECHNOLOGYBIOMETRICS FINGER PRINT TECHNOLOGY
BIOMETRICS FINGER PRINT TECHNOLOGY
 
Detecting Fake News Through NLP
Detecting Fake News Through NLPDetecting Fake News Through NLP
Detecting Fake News Through NLP
 
Sentiment Analysis of Twitter Data
Sentiment Analysis of Twitter DataSentiment Analysis of Twitter Data
Sentiment Analysis of Twitter Data
 
Sms spam-detection
Sms spam-detectionSms spam-detection
Sms spam-detection
 
Fingerprint recognition
Fingerprint recognitionFingerprint recognition
Fingerprint recognition
 
PPT.pptx
PPT.pptxPPT.pptx
PPT.pptx
 
Detection of phishing websites
Detection of phishing websitesDetection of phishing websites
Detection of phishing websites
 
Fake Currency detction Using Image Processing
Fake Currency detction Using Image ProcessingFake Currency detction Using Image Processing
Fake Currency detction Using Image Processing
 
Wine Quality Analysis Using Machine Learning
Wine Quality Analysis Using Machine LearningWine Quality Analysis Using Machine Learning
Wine Quality Analysis Using Machine Learning
 
Handwritten Character Recognition
Handwritten Character RecognitionHandwritten Character Recognition
Handwritten Character Recognition
 
Seminar Report Mine
Seminar Report MineSeminar Report Mine
Seminar Report Mine
 
Fingerprint Authentication for ATM
Fingerprint Authentication for ATMFingerprint Authentication for ATM
Fingerprint Authentication for ATM
 
Speech recognition final presentation
Speech recognition final presentationSpeech recognition final presentation
Speech recognition final presentation
 
final presentation fake news detection.pptx
final presentation fake news detection.pptxfinal presentation fake news detection.pptx
final presentation fake news detection.pptx
 
Sign language recognizer
Sign language recognizerSign language recognizer
Sign language recognizer
 
Machine learning seminar ppt
Machine learning seminar pptMachine learning seminar ppt
Machine learning seminar ppt
 

Similar to FAKE Review Detection

Fraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning TechniquesFraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning Techniquesijceronline
 
The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...INFOGAIN PUBLICATION
 
Digital technology analysis
Digital technology analysisDigital technology analysis
Digital technology analysis66xiu
 
Yelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer SatisfactionYelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer Satisfactionsunnycmui
 
Spot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsSpot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsYousef Fadila
 
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Theo Mandel, PhD
 
Yelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYao Yao
 
IRJET- Implementation of Review Selection using Deep Learning
IRJET-  	  Implementation of Review Selection using Deep LearningIRJET-  	  Implementation of Review Selection using Deep Learning
IRJET- Implementation of Review Selection using Deep LearningIRJET Journal
 
IRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET Journal
 
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET Journal
 
Digital presentation yelp
Digital presentation yelpDigital presentation yelp
Digital presentation yelpwdgg307
 
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuanePitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuaneAlan Quayle
 
Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Traction Software
 

Similar to FAKE Review Detection (20)

Fraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning TechniquesFraud Detection in Online Reviews using Machine Learning Techniques
Fraud Detection in Online Reviews using Machine Learning Techniques
 
The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...The study of factors affecting customer’s satisfaction with the three star ho...
The study of factors affecting customer’s satisfaction with the three star ho...
 
Digital technology analysis
Digital technology analysisDigital technology analysis
Digital technology analysis
 
CUROP Poster (002)
CUROP Poster (002)CUROP Poster (002)
CUROP Poster (002)
 
Yelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer SatisfactionYelp an Assessment of Customer Satisfaction
Yelp an Assessment of Customer Satisfaction
 
Spot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor ReviewsSpot deceptive TripAdvisor Reviews
Spot deceptive TripAdvisor Reviews
 
Marriott Interviews
Marriott InterviewsMarriott Interviews
Marriott Interviews
 
Mahendra nath
Mahendra nathMahendra nath
Mahendra nath
 
Sm chap 1
Sm chap 1Sm chap 1
Sm chap 1
 
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
Using your User Experience (UX) Super Powers for Good or Evil - Theo Mandel, ...
 
Yelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm PaperYelp's Review Filtering Algorithm Paper
Yelp's Review Filtering Algorithm Paper
 
Can You Identify Fake or Authentic Reviews? An fsQCA Approach
Can You Identify Fake or Authentic Reviews? An fsQCA ApproachCan You Identify Fake or Authentic Reviews? An fsQCA Approach
Can You Identify Fake or Authentic Reviews? An fsQCA Approach
 
IRJET- Implementation of Review Selection using Deep Learning
IRJET-  	  Implementation of Review Selection using Deep LearningIRJET-  	  Implementation of Review Selection using Deep Learning
IRJET- Implementation of Review Selection using Deep Learning
 
IRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion MiningIRJET- Fake Review Detection using Opinion Mining
IRJET- Fake Review Detection using Opinion Mining
 
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
IRJET- Improving Performance of Fake Reviews Detection in Online Review’s usi...
 
Greg Sterling - Advanced Search Summit Napa 2021
Greg Sterling - Advanced Search Summit Napa 2021Greg Sterling - Advanced Search Summit Napa 2021
Greg Sterling - Advanced Search Summit Napa 2021
 
Prevalence of Deception in Online Customer Reviews
Prevalence of Deception in Online Customer ReviewsPrevalence of Deception in Online Customer Reviews
Prevalence of Deception in Online Customer Reviews
 
Digital presentation yelp
Digital presentation yelpDigital presentation yelp
Digital presentation yelp
 
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne RuanePitfalls and potholes of content moderation for chatbots, Elayne Ruane
Pitfalls and potholes of content moderation for chatbots, Elayne Ruane
 
Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....Capturing and managing non-traditional information for Competitive Advantage....
Capturing and managing non-traditional information for Competitive Advantage....
 

FAKE Review Detection

  • 1. FAKE REVIEW DETECTION What are they saying about you ? Are they real … Guided By : Dr. Animesh Mukherjee
  • 2. ONLINE REVIEW • Captures testimonials of “real” people (unlike advertisements). • Shapes decision making of customers. • Positive reviews – financial gains and fame for business. • Deceptive opinion spamming – to promote or discredit some target product and services.[1] • Opinion Spammers admitted for being paid to write fake reviews.(Kost, 2012) • Yelp.com – “Sting operation” : publically shame business who buy fake reviews.[2] [1] Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf [2] Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake-reviews.html
  • 4. • Amazon Mechanical Turk – crowd sourced online workers(turkers) to write fake reviews ($1 per review) portraying 20 hotels of Chicago in positive light.[1] • 400 fake positive reviews collected and 400 non-fake reviews on same 20 hotels using Tripadvisor.com • Yelp’s filtered and unfiltered reviews collected to understand working of Yelp review classification algorithm. • Approach : Linguistic n-gram features and some supervised learning method.[2] [1] Amazon Mechanical Turk : https://www.mturk.com/ [2] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf Dataset Collection and Approach for Analysis
  • 5. Linguistic Approach : Results • Using only Bigram features : accuracy of 89.6% on AMT data.[1] • Using same n-gram features : accuracy of 67.8% on Yelp Data.[2] • Table 1: Class distribution of Yelp data is skewed – imbalanced data produces poor model.(Chawla et al, 2004) • Good model for imbalanced data – under sampling. (Drummond and Holte, 2003) [1] Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf [2] A Mukherjee - 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 6. Linguistic Approach Results : Explained • For AMT data, word distribution of fake and non-fake reviews are very different, which explains the high accuracy using n-gram. • Reason for the different word distribution : Domain knowledge absence and little gain in writing fake reviews($1 per review). • Poor performance of n-grams for Yelp data because spammers according to Yelp filter used very similar language in fake review as non-fake review : linguistically very similar.[1] • Inefficiency in linguistic in detecting fake reviews filtered by Yelp encourages to do behavioral study of reviews. [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 7. ))(/)((2log)()||( iNiFiFNFKL i Information Theoretic Analysis • To explain huge accuracy difference : analysis of word distribution of AMT and Yelp data. • Good-Turing smoothed unigram language models. • Computation of word distribution difference across fake & non-fake reviews : Kullback-Leibler(KL) Divergence : where F(i) and N(i) are respective probabilities of word i in fake and non-fake reviews.[1] • Here KL(F||N) gives quantitative estimate of how much fake reviews linguistically differ from non-fake reviews [1] Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html
  • 8. Information Theoretic Analysis (Cont..) • KL Divergence is Asymmetric : • We have, • For AMT data : and • For Yelp data : and • [1]    FNKLNFKL ||||     FNKLNFKL ||||  0KL    FNKLNFKL ||||  0KL [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf    FNKLNFKLKL |||| 
  • 9. Information Theoretic Analysis (Cont..) • KL(F||N) Definition implies that words having higher probability in F and very low probability in N contribute most to KL-Divergence. • To study word-wise contribution to ∆KL, word ∆KL calculated as : where, • Contribution of top k words to ∆KL for k= 200 and k=300 )||()||( iiwordiiword i Word FNKLNFKLKL           )( )( log)||( 2 iN iF iFNFKL iiword
  • 10. Turkers didn’t do a good job at “Faking”! Word-wise Difference of KL- Divergence across top 200 words. Equally Dense : |E| = |G| • Symmetric Distribution of for top k-words for AMT data implies the existence of two set of words: 1. set of words E, appearing more in fake reviews than in non-fake, Ɐi € E, F(i) > N(i) resulting in 2. set of words G, appearing more in non-fake reviews than in fake, Ɐi € G, N(i) > F(i) resulting in i WordKL 0 Ei WordKL 0 Ei WordKL • Additionally top k=200 words only contribute 20% to ∆KL for the AMT data. There are many words in AMT data having higher probabilities in fake than non- fake and vice-versa. • This implies fake and non-fake reviews in AMT data consist of words with very different frequencies. Turkers didn’t do a good job in faking.[1] [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 11. Yelp Spammers are Smart but Overdid “Faking”! • Yelp fake review data shows KL(F||N) is much larger than KL(N||F) and ∆KL>1. • Given graphs from b-e shows among top 200 words which contribute to major (=70%) most words have and only a few have . Word-wise Difference of KL-Divergence across top 200 words.[1] 0 i WordKL 0 i WordKL [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 12. Yelp Spammers are Smart but Overdid “Faking”! (Cont..) • Consider A be set of top words contributing most to ∆KL. We partition A as: where and • The curve above y=0 is dense and below it is sparse which implies • clearly indicates that there exists specific words which contribute most to ∆KL by appearing in fake reviews with much higher frequencies than in non-fake reviews. • Spammers made smart effort to ensure that their fake reviews have most words that also appear in non-fake reviews to sound convincing.  ))()(,.,(0| iFiNAiieKLiA Ni Word N   ))()(,.,(0| iNiFAiieKLiA Fi Word F  NF AAA   NF AA NF AA  NF AA 
  • 13. Yelp Spammers are Smart but Overdid “Faking”! (Cont..) • While making their reviews sound convincing, psychologically they happened to OVERUSE come words resulting higher frequency of certain words in fake review than non-fake reviews. • A quick lookup yields {us, price, stay, feel, deal, comfort} in hotel domain and {options, went, seat, helpful, overall, serve, amount etc.} in restaurants. • Prior personality work shown that deception/lying usually involves more use of personal pronouns(eg. us) and associated action(eg. Went, feel) towards specific targets(eg. option, price, stay) with objective of incorrect projection(lying or faking).[1] • Spammers caught by Yelp left behind linguistic footprints which can be caught by precise behavioral study. [1] Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words-_predicting_deception_from_linguistic_styles.pdf
  • 14. Spamming Behavior Analysis 1. Maximum Number of Reviews(MNR) : Writing too many reviews in a day is abnormal. 2. Percentage of Positive Reviews(PR) : Deception words in fake reviews indicates projection in positive light. CDF of positive(4-5 stars) reviews among all reviews is plotted to illustrate analysis. 3. Review Length(RL) : While writing fake experiences, there is probably not much to write and also spammer not want to spend too much time in it. 4. Maximum Content Similarity(MCS) : To examine if some posted reviews are similar to previous reviews, we computed cosine similarity between two reviews of a reviewer. Non-spammers mostly write new contents.
  • 15. Maximum Number of Reviews Percentage of Positive Reviews Review Length Maximum Content Similarity Spamming Behavior Analysis CDFs [1] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf
  • 16. Challenges with Supervised Evaluation • Very difficult to find gold-standard data of fake and non-fake reviews for model building – too difficult to manually recognize/label fake/non-fake reviews by mere reading. • Duplicate and near duplicate assumed to be fake, which is unreliable. • Usage of manually labeled dataset – reliability issues because it have been shown that accuracy of human labeling of fake reviews is very poor.[1] • AMT crowdsourcing fake reviews by paying are fake yet they don’t reflect the dynamics of fake reviews in commercial website.[2] • This lack of labeled data, motivates to look after the unsupervised methods of classification.[3] [1] Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination. [2] Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera-submit.pdf [3] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 17. Unsupervised Evaluation Model • Since human labeling for supervised learning is difficult, problem was proposed by modeling spamicity (degree of spamming) as latent with other observed behavioral features. • Unsupervised model – Author Spamicity Model (ASM) proposed.[1] • Taken a fully Bayesian approach and formulated opinion spam detection as clustering problem. • Opinion spammers have different behavioral distributions than non-spammers. • Causes distributional divergence between latent population distributions of two clusters: spammers and non-spammers.[1] • Model inference results in learning the population distributions of two clusters. [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 18. • Formulates spam detection as an unsupervised clustering problem in Bayesian setting. • Belongs to class of Generative models for clustering based on set of observed features. • Models spamicity 𝑠 𝑎 (in range[0, 1]) of an author a; and spam label 𝜋 𝑟 of a review which is the class variable reflecting the cluster membership (two cluster K=2, spam and non-spam).[1] • Each author/reviewer and respectively each review has a set of observed features(behavioral clues). • Certain characteristics of abnormal behavior defined which likely to link with spamming and thus can be exploited in model for learning spam and non-spam clusters. [1] Author Spamicity Model [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf
  • 19. Author Features • It have value in range [0, 1] and value close to 1 indicates spamming. 1. Content Similarity : Crafting new review every time is time consuming, spammers likely to copy reviews across similar product. Choose maximum similarity to capture worst spamming behavior.[1] 2. Maximum Number of Reviews : Posting many reviews a day is also abnormal. 3. Reviewing Burstiness : Spammers are usually not longtime members of site. Defined over an activity window(first and last review posting date). If posted over reasonably long timeframe, it probably a normal activity but all review posted within short burst likely to be spam.[2] 4. Ratio of First Review : People mostly rely on early reviews and spamming early impacts hugely on sales. Spammers try to be among first reviewers. [1] [1] Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf [2] Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012).
  • 20. Review Features • It have five binary review features. Value 1 indicates spamming else 0 is non-spamming. 1. Duplicate /Near Duplicate Reviews: Spammers often post multiple reviews which are duplicate /near duplicate on same product to boost ratings. 2. Extreme Rating: Spammers mostly like to give extreme ratings(1 or 5) in order to boost ratings to demote or promote products. 3. Rating Deviation: Spammers usually involve in wrong projection in either positive or negative and it deviates from average ratings given by other reviewers. 4. Early Time Frame: Early review can greatly impact people’s sentiments on a product. 5. Rating Abuse: Multiple ratings on same product are unusual. Similar to DUP but focuses on rating dimensions rather than content.
  • 21. CONCLUSIONS • We presented an in-depth investigation of nature of fake reviews in Commercial settings of Yelp.com. • Our study shows linguistics methods of (Ott et al., 2011) and its high accuracy in AMT data. • We presented a behavioral study of spammers for real-life fake reviews. • We presented a brief introduction to n-gram language model. • We presented challenges with the supervised evaluation and gave idea about the unsupervised approach of evaluation. • We presented a brief introduction to unsupervised Author Spamicity Model (ASM).
  • 22. REFERENCES 1. Jindal and Liu 2008 : http://www.cs.uic.edu/~liub/FBS/opinion-spam-WSDM-08.pdf 2. Yelp official blog : http://officialblog.yelp.com/2013/05/how-yelp-protects-consumers-from-fake- reviews.html 3. MIT N-gram Language Model Tutorial: http://web.mit.edu/6.863/www/fall2012/readings/ngrampages.pdf 4. Amazon Mechanical Turk : https://www.mturk.com/ 5. Ott et al. 2011 : https://www.cs.cornell.edu/courses/CS4740/2012sp/lectures/op_spamACL2011.pdf 6. Mukherjee et al , 2013 : http://www2.cs.uh.edu/~arjun/papers/ICWSM-Spam_final_camera- submit.pdf 7. Kullback Leibler Divergence : http://www.cs.buap.mx/~dpinto/research/CICLing07_1/Pinto06c/node2.html 8. Newman et al , 2003 : http://www.communicationcache.com/uploads/1/0/8/8/10887248/lying_words- _predicting_deception_from_linguistic_styles.pdf 9. Mukherjee, A., Liu, B. and Glance, N. 2012. Spotting Fake Reviewer Groups in Consumer Reviews. WWW (2012). 10. Mukherjee et al : http://delivery.acm.org/10.1145/2490000/2487580/p632-mukherjee.pdf 11. Ott M, Choi, Y, Cardie, C. and Hancock, J.T. 2011. Finding Deceptive Opinion Spam by Any Stretch of the Imagination.