Prof. Panos Ipeirotis Search and the New Economy Privacy on the Web.
Panos Ipeirotis Stern School of Business New York University Analyzing User-Generated Content using...
-
Upload
louisa-butler -
Category
Documents
-
view
218 -
download
0
Transcript of Panos Ipeirotis Stern School of Business New York University Analyzing User-Generated Content using...
Panos Ipeirotis
Stern School of Business
New York University
Analyzing User-Generated Content using Econometrics
Comparative Shopping
Comparative Shopping
Are Customers Irrational?
$11.04 (+1.5%)
BuyDig.com gets
Price Premium(customers pay more than
the minimum price)
Price Premiums / Discounts @ Amazon
0
1000
2000
3000
4000
5000
6000
7000
8000
9000
10000
-100 -75 -50 -25 0 25 50 75 100
Price Premium
Nu
mb
er
of
Tra
ns
ac
tio
ns Are Buyers
Irrational (?
)
(paying more)
Are Sellers Irrational (?)(charging less)
Why not Buying the Cheapest?
You buy more than a product
Customers do not pay only for the product
Customers also pay for a set of fulfillment characteristics
Delivery
Packaging
Responsiveness
…
Customers care about reputation of sellers!
Example of a reputation profile
The Idea in a Single Slide
Conjecture: Price premiums measure reputation
Reputation is captured in text feedback
Our contribution: Examine how text affects price premiums
(and learn to rank opinion phrases as a side effect)
ACL 2007
Data: Variables of Interest
Price Premium
Difference of price charged by a seller minus listed price of a competitor
Price Premium = (Seller Price – Competitor Price)
Calculated for each seller-competitor pair, for each transaction
Each transaction generates M observations, (M: number of competing sellers)
Alternative Definitions:
Average Price Premium (one per transaction)
Relative Price Premium (relative to seller price)
Average Relative Price Premium (combination of the above)
Decomposing Reputation
Is reputation just a scalar metric?
Previous studies assumed a “monolithic” reputation
Decompose reputation in individual components
Sellers characterized by a set of fulfillment characteristics(packaging, delivery, and so on)
What are these characteristics (valued by consumers?)
We think of each characteristic as a dimension, represented by a noun, noun phrase, verb or verbal phrase (“shipping”, “packaging”, “delivery”, “arrived”)
We scan the textual feedback to discover these dimensions
Decomposing and Scoring Reputation
Decomposing and scoring reputation
We think of each characteristic as a dimension, represented by a noun or verb phrase (“shipping”, “packaging”, “delivery”, “arrived”)
The sellers are rated on these dimensions by buyers using modifiers (adjectives or adverbs), not numerical scores
“Fast shipping!”
“Great packaging”
“Awesome unresponsiveness”
“Unbelievable delays”
“Unbelievable price”
How can we find out the meaning of these adjectives?
Structuring Feedback Text: Example
What is the reputation score of this feedback?
P1: I was impressed by the speedy delivery! Great Service!
P2: The item arrived in awful packaging, but the delivery was speedy
Deriving reputation score
We assume that a modifier assigns a “score” to a dimension α(μ, k): score associated when modifier μ evaluates the k-th dimension
w(k): weight of the k-th dimension
Thus, the overall (text) reputation score Π(i) is a sum:
Π(i) = 2*α (speedy, delivery) * weight(delivery)+ 1*α (great, service) * weight(service) +
1*α (awful, packaging) * weight(packaging)
unknownunknown?
Sentiment Scoring with Regressions
Scoring the dimensions
Use price premiums as “true” reputation score Π(i) Use regression to assess scores (coefficients)
Regressions
Control for all variables that affect price premiums
Control for all numeric scores of reputation
Examine effect of text: E.g., seller with “fast delivery” has premium $10 over seller with “slow delivery”, everything else being equal
“fast delivery” is $10 better than “slow delivery”
estimated coefficients
Π(i) = 2*α (speedy, delivery) * weight(delivery)+ 1*α (great, service) * weight(service) +
1*α (awful, packaging) * weight(packaging)
PricePremium
Measuring Reputation
• Regress textual reputation against price premiums
• Example for “delivery”:– Fast delivery vs. Slow delivery: +$7.95– So “fast” is better than “slow” by a $7.95 margin
Some Indicative Dollar Values
Positive Negative
Natural method for extracting sentiment strength and polarity
good packaging -$0.56
Naturally captures the pragmatic meaning within the given context
captures misspellings as well
Positive? Negative?
More Results
Further evidence: Who will make the sale?
Classifier that predicts sale given set of sellers
Binary decision between seller and competitor
Used Decision Trees (for interpretability)
Training on data from Oct-Jan, Test on data from Feb-Mar
Only prices and product characteristics: 55%
+ numerical reputation (stars), lifetime: 74%
+ encoded textual information: 89%
text only: 87%
Text carries more information than the numeric metrics
Looking Back
• Comprehensive setting– All information about merchants stored at feedback
profile
• Easy text processing– Large number of feedback postings (100’s and
1000’s of postings common)– Short and concise language
Similar Setting: Word of “Mouse”
• Consumer reviews– Derived from user experience– Describe different product features– Provide subjective evaluations of product features
• Product reviews affect product sales– What is the importance of each product feature?– What is the consumer evaluation of each feature?
Apply the same techniques?
I love virtually everything about this camera....except the lousy picture quality. The camera looks great, feels nice, is easy to use, starts up quickly, and is of course waterproof.
It fits easily in a pocket and the battery lasts for a reasonably long period of time.
Contrast with Reputation
Significant data sparseness• Smaller number of reviews per product
– Typically 30-50 reviews vs. 200-5,000 postings
• Much longer than feedback postings– 2-3 paragraphs each, vs 80-100 characters in reputation
Not an isolated system• Consumers form opinions from many sources
Bayesian Learning Approach
• Consumers perform Bayesian learning of product attributes using signals from reviews– Consumers have prior expectations of quality– Consumers update expectation from new signals
Online shopping as learning
Belief for Image Quality
Updated Belief for Image Quality
“excellent image quality”“fantastic image quality”“superb image quality”
“great image quality”“fantastic image quality”“superb image quality”
Updated Belief for Image Quality
Consumers pick the product that maximizes their expected utility
Expected Utility
• Consumers pick the product that maximizes their expected utility
• Expected utility based on:– Mean of the evaluation and– Uncertainty of the evaluation
Notice: negative reviews may increase sales!
Design Image Quality
+U=
Mean(design)
Var(design)
Mean(img qual)
Variance(img qual)
• Examine changes in demand and infer parameters
Product Reviews and Product Sales
“poor lens”
+3%
“excellent lens”
-1%
“poor photos”
+6%
“excellent photos”
-2%
Feature “photos” is two time more important than “lens” “Excellent” is positive, “poor” is negative “Excellent” is three times stronger than “poor”
Feature Weights for Digital Cameras
0
0.2
0.4
0.6
0.8
1
1.2
SLRPoint & Shoot
New Product Search Approach
• Consumers want the “best product” first
• Best product: Highest value for the money– Maximize (gained) product utility– Minimize (lost) utility of money
Utility of Money
The highest the available income, the lowest the utility of money(i.e., rich people spend easier)
Hotel Search Application
• Transaction data from big travel search website
• Computed “expected utility” for each hotel using:– Reviews– Satellite photos for landscape (beach, downtown,
highway,…)– Location statistics (crime, etc) and points of interest
• Substracted “utility of money” based on its price
• Ranked according to “consumer surplus” (i.e., difference of two)
Hotel Ranking
Percentage of users preferring econometric ranking in blind comparison
Cities
Tripadvisor Travelocity Price
Low to high
Price
High to Low
Hotel
Class
#
Review
# of
Ameni-ties
New York 72% 68% 62% 70% 66% 66% 62%
Los Angeles 68% 66% 64% 84% 88% 80% 64%
San Francisco 84% 62% 68% 72% 66% 66% 70%
Orlando 64% 68% 68% 74% 66% 74% 66%
New Orleans 84% 62% 83% 64% 64% 66% 70%
Salt Lake City 80% 80% 68% 72% 82% 66% 68%
Significance Level
P=0.01
≥ 66%
P=0.001
≥ 72%(Sign Test, N=50)
Other Applications
• Financial news and price/variance prediction
• Measuring (and predicting) importance of political events
• Deriving better keyword bidding, pricing, and ad generation strategies
http://economining.stern.nyu.edu
Other Projects
• SQoUT projectStructured Querying over Unstructured Text
http://sqout.stern.nyu.edu
• Managing Noisy LabelersAmazon Mechanical Turk, “Wisdom of the Crowds”
31
32
SQoUT: Structured Querying over Unstructured Text
Information extraction applications extract structured relations from unstructured text
May 19 1995, Atlanta -- The Centers for Disease Control and Prevention, which is in the front line of the world's response to the deadly Ebola epidemic in Zaire , is finding itself hard pressed to cope with the crisis…
Date Disease Name Location
Jan. 1995 Malaria Ethiopia
July 1995 Mad Cow Disease U.K.
Feb. 1995 Pneumonia U.S.
May 1995 Ebola Zaire
Information Extraction System
(e.g., NYU’s Proteus)
Disease Outbreaks in The New York Times
33
SQoUT: The QuestionsOutput Tokens
…Extraction
System(s)
Text Databases
3. Extract output tuples
2. Process documents
1. Retrieve documents from database/web/archive
Questions: 1. How to we retrieve the documents?2. How to configure the extraction systems?3. What is the execution time? 4. What is the output quality?
SIGMOD’06, TODS’07, ICDE’09, TODS’09
Mechanical Turk Example
34
Motivation Labels can be used in training predictive models
Duplicate detection systems Image recognition Web search
But: labels obtained from above sources are noisy. This directly affects the quality of learning models How can we know the quality of annotators? How can we know the correct answer? How can we use best noisy annotators?
36
40
50
60
70
80
90
100
1 20 40 60 80 100
120
140
160
180
200
220
240
260
280
300
Number of examples (Mushroom)
Acc
ura
cy
Quality and Classification Performance
Labeling quality increases classification quality increases
Q = 0.5
Q = 0.6
Q = 0.8
Q = 1.0
Tradeoffs for Classification
Get more labels Improve label quality Improve classification Get more examples Improve classification
40
50
60
70
80
90
100
1 20 40 60 80 100
120
140
160
180
200
220
240
260
280
300
Number of examples (Mushroom)
Acc
ura
cy
Q = 0.5
Q = 0.6
Q = 0.8
Q = 1.0
KDD 2008
Thank you! Questions?
Price premiums @ Amazon
0
1000
2000
3000
4000
5000
6000
7000
8000
9000
10000
-100 -75 -50 -25 0 25 50 75 100
Price Premium
Nu
mb
er
of
Tra
ns
ac
tio
ns
Average price premiums @ Amazon
0
200
400
600
800
1000
1200
-100 -75 -50 -25 0 25 50 75 100
Average Price Premium
Nu
mb
er
of
Tra
ns
ac
tio
ns
Relative Price Premiums
-1--0.9
-0.9--0.8
-0.8--0.7
-0.7--0.6
-0.6--0.5
-0.5--0.4
-0.4--0.3
-0.3--0.2
-0.2--0.1
-0.1-0.0
0-0.1 0.1-0.2
0.2-0.3
0.3-0.4
0.4-0.5
0.5-0.6
0.6-0.7
0.7-0.8
0.8-0.9
0.9-10
2000
4000
6000
8000
10000
12000
14000
16000
18000
20000
Average Relative Price Premiums
-1--0.9
-0.9--0.8
-0.8--0.7
-0.7--0.6
-0.6--0.5
-0.5--0.4
-0.4--0.3
-0.3--0.2
-0.2--0.1
-0.1-0.0
0-0.1 0.1-0.2
0.2-0.3
0.3-0.4
0.4-0.5
0.5-0.6
0.6-0.7
0.7-0.8
0.8-0.9
0
500
1000
1500
2000
2500
Capturing transactions and “price premiums”
Data: Transactions
Seller ListingItem Price
When item is sold, listing disappears
Capturing transactions and “price premiums”
Data: Transactions
While listing appears, item is still available
time
Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10
Capturing transactions and “price premiums”
Data: Transactions
While listing appears, item is still available
time
Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10
Item still not sold on 1/7
Capturing transactions and “price premiums”
Data: Transactions
When item is sold, listing disappears
time
Item sold on 1/9
Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10
Seller: uCameraSite.com
1. Canon Powershot x300
2. Kodak - EasyShare 5.0MP
3. Nikon - Coolpix 5.1MP
4. Fuji FinePix 5.1
5. Canon PowerShot x900
Reputation Pricing Tool for Sellers
Your last 5 transactions in Cameras
Name of product Price
Seller 1 - $431
Seller 2 - $409
You - $399
Seller 3 - $382
Seller 4-$379
Seller 5-$376
Canon Powershot x300
Your competitive landscapeProduct Price (reputation)
(4.8)
(4.65)
(4.7)
(3.9)
(3.6)
(3.4)
Your Price: $399Your Reputation Price: $419Your Reputation Premium: $20 (5%)
$20
Left on the table
25%
14%
7%
45%
9%
Quantitatively Understand & Manage Seller Reputation
RSI Tool for Seller Reputation Management
How your customers see you relative to other sellers:
35%*
69%
89%
82%
95%
Service
Packaging
Delivery
Overall
Quality
Dimensions of your reputation and the relative importance to your customers:
Service
Packaging
Delivery
Quality
Other* Percentile of all merchants
• RSI Products Automatically Identify the Dimensions of Reputation from Textual Feedback• Dimensions are Quantified Relative to Other Sellers and Relative to Buyer Importance• Sellers can Understand their Key Dimensions of Reputation and Manage them over Time• Arms Sellers with Vital Info to Compete on Reputation Dimensions other than Low Price.
Marketplace Search
Buyer’s Tool
Used Market (ex: Amazon)
Price Range $250-$300
Seller 1 Seller 2
Seller 4 Seller 3
Sort by Price/Service/Delivery/other dimensions
Canon PS SD700
Service
Packaging
Delivery
Price
Dimension Comparison
Seller 1
Price Service Package Delivery
Seller 2
Seller 3
Seller 4
Seller 5
Seller 6
Seller 7
Data
Overview
Panel of 280 software products sold by Amazon.com X 180 days
Data from “used goods” market
Amazon Web services facilitate capturing transactions
We do not use any proprietary Amazon data (Details in the paper)
Data: Secondary Marketplace
Data: Capturing Transactions
time
Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8
We repeatedly “crawl” the marketplace using Amazon Web Services
While listing appears item is still available no sale
Data: Capturing Transactions
time
Jan 1 Jan 2 Jan 3 Jan 4 Jan 5 Jan 6 Jan 7 Jan 8 Jan 9 Jan 10
We repeatedly “crawl” the marketplace using Amazon Web Services
When listing disappears item sold
Weights of Hotel Characteristics Based on Different Travel Purposes
Consumers with different travel purposes assign different weight distributions on the same set of hotel characteristics.
Sensitivity to Rating and Review Count Based on Different Age Groups
Age 18-34 pay more attention to online reviews compared to other age groups.
User Study
Experiment 1: Blind pair-wise comparisons, 100 anonymous AMT users;
8 existing baselines:
-Price low to high
-Price high to low
-Online review count
-Hotel class
-Hotel size (number of rooms
-Number of internal amenities
-TripAdvisor reviewer rating
-Travelocity reviewer rating
Conclusion: CS-based ranking is overwhelmingly preferred.
Reasoning: Diversity, satisfies consumers’ multidimensional preferences
User Study
Experiment 2: Blind pair-wise comparisons, 100 anonymous AMT users;
baseline: generalized CS-based ranking (for an average consumer).E.g., Business trip and family trip AMT user study results in the NYC experiment.
Conclusion: Personalized CS-based ranking is overwhelmingly preferred.
80%
20%
BusinessPersonal-ized
87%
13%
FamilyPersonal-ized
Reasoning: Capture consumers’ specific expectations, dovetail with their real purchase motivation.
Estimation Results Capture Real Motivation
e.g., Business travelers indicated that they prefer quiet inner environment and easy access to highway or public transportation. This was fully captured in our estimation results, see (b).