How Opinions are Received by Online Communities A Case Study on Amazon.com Helpfulness Votes Cristian Danescu-Niculescu-Mizil 1 , Gueorgi Kossinets 2 , Jon Kleinberg 1 , Lillian Lee 1 1 Dept. of Computer Science, Cornell University, 2 Google Inc. WWW 2009 2009. 07. 30. IDS Lab. Hwang Inbeom
31
Embed
How Opinions are Received by Online Communities A Case Study on Amazon.com Helpfulness Votes Cristian Danescu-Niculescu-Mizil 1, Gueorgi Kossinets 2, Jon.
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
How Opinions are Received by Online Communi-tiesA Case Study on Amazon.com Helpfulness Votes
Cristian Danescu-Niculescu-Mizil1, Gueorgi Kossinets2, Jon Kleinberg1, Lillian Lee1
1Dept. of Computer Science, Cornell University, 2Google Inc.
WWW 2009
2009. 07. 30.
IDS Lab.
Hwang Inbeom
Copyright 2009 by CEBT
Outline
Users’ evaluation on online reviews: Helpfulness votes
Observation of behaviors
Making some hypothesis and proving their validity
Coming up with a mathematical model explains these be-haviors
2
Copyright 2009 by CEBT
Introduction
Opinion
What did Y think of X?
3
Copyright 2009 by CEBT
Introduction
Meta-Opinion
What did Z think of Y’s opinion of X?
4
Copyright 2009 by CEBT
The Helpfulness of Reviews
Widely-used web sites include not just reviews, but also evaluations of the helpfulness of the reviews
The helpfulness vote
– “Was this review helpful to you?”
Helpfulness ratio:
– “a out of b people found the review itself helpful”
5
b
a
Copyright 2009 by CEBT
Amazon.com Helpfulness Votes Data
4,000,000 reviews about roughly 700,000 books, includ-ing average star ratings and helpfulness ratios
6
Average star rating
Helpfulness ratio
Copyright 2009 by CEBT
Definitions of “Helpfulness”
Helpfulness in the narrow sense: “Does this review help you in making a purchase
decision?”
Liu’s work: annotation and classification of review helpful-ness
Annotators’ evaluation differed significantly from the help-fulness votes
Helpfulness “in the wild”
The way Amazon users evaluate each others’ reviews
Intertwined with complex social feedback mechanisms
7
Copyright 2009 by CEBT
Flow of Presentation
Hypothe-siz-ing
Verify-ing
Model-ing
8
Copyright 2009 by CEBT
Flow of Presentation
Hypoth-esizing•Con-formity
•Indi-vidual-bias
•Bril-liant-but-cruel
•Qual-ity-only
Verifying Modeling
9
Copyright 2009 by CEBT
Hypotheses: Social Mechanisms underlying
Well-studied hypotheses for how social effects influence group’s reaction to an opinion
The conformity hypothesis
The individual-bias hypothesis
The brilliant-but-cruel hypothesis
The quality-only straw-man hypothesis
10
Copyright 2009 by CEBT
Hypotheses
The conformity hypothesis
Review is evaluated as more helpful when its star rating is closer to the consensus star rating
– Helpfulness ratio will be the highest of which reviews have star rating equal to overall average
The individual-bias hypothesis
When a user considers a review, he or she will rate it more highly if it expresses an opinion that he or she agrees with
11
Copyright 2009 by CEBT
Hypotheses (contd.)
The brilliant-but-cruel hypothesis
Negative reviewers are perceived as more intelligent, com-petent, and expert than positive reviewers
The Quality-only straw-man hypothesis
Helpfulness is being evaluated purely based on the textual content of reviews
Non-textual factors are simply correlates of textual quality
12
Copyright 2009 by CEBT
Flow of Presentation
Hypothe-siz-ing
Verifying•Absolute deviation of helpful-ness ratio
•Signed deviation of helpful-ness ratio
•Variance of star rat-ing and helpful-ness ratio
•Making use of plagiarism
Model-ing
13
Copyright 2009 by CEBT
Hypotheses
Conformity•A review is evaluated as more helpful when its star rating is closer to the average star rating
Individual-bias•A review is evaluated as more helpful when its star rating is closer to evaluator’s opinion
Brilliant-but-cruel•A review is evaluated as more helpful when its star rating is below to the average star rating
Making use of plagiarism is effective way to control for the effect of review text
Definition of plagiarized pair(s) of reviews
Two or more reviews of different products
With near-complete textual overlap
22
Copyright 2009 by CEBT
An Example
Skull-splitting headache guaranteed!•If you enjoy thumping, skull splitting migraine headache, then Sing N Learn is for you.As a longtime language instructor, I agree with the attempt and effort that this series makes, but it is the execution that ultimately weakens Sing N Learn Chinese.To be sure, there are much, much better ways to learn Chi-nese. In fact, I would recommend this title only as a last re-sort and after you’ve thoroughly exhausted traditional ways to learn Chinese …
Migraine Headache at No Extra Charge•If you enjoy a thumping, skull splitting migraine headache, then the Sing N Learn series is for you.As a longtime language instructor, I agree with the effort that this series makes, but it is the execution that ultimately weakens Sing N Learn series. To be sure, there are much, much better ways to learn a foreign language. In fact, I would recommend this title only as a last resort and after you’ve thoroughly exhausted traditional ways to learn Korean …
23
Copyright 2009 by CEBT
Plagiarism (contd.)
Plagiarized reviews
Almost(not exact) same text
– More possibly, same text could be considered as spam reviews
Different non-textual information
If the quality-only straw man hypothesis holds, helpful-ness ratios of documents in each pair should be the same
Possible other methods
Human annotation
– Could be subjective
Classification using machine learning methods
– We cannot guarantee the accuracies of algorithms
24
Copyright 2009 by CEBT
Experiments with Plagiarism
Text quality is not the only explanatory factor
Statistically significant difference between the helpfulness ratios of plagiarized pairs
25
The plagiarized reviews with deviation 1 is significantly more helpful than those with deviation 1.5
Copyright 2009 by CEBT
Hypotheses
Conformity•A review is evaluated as more helpful when its star rating is closer to the average star rating
Individual-bias•A review is evaluated as more helpful when its star rating is closer to evaluator’s opinion
Brilliant-but-cruel•A review is evaluated as more helpful when its star rating is below to the average star rating