CS 572: Information Retrieval - Emory University

76
1 CS 572: Information Retrieval Personalized Search Acknowledgements Some slides in this lecture adapted from various IR researchers, credits noted on each slide. CS 572: Information Retrieval. Spring 2010 4/13/2016

Transcript of CS 572: Information Retrieval - Emory University

Page 1: CS 572: Information Retrieval - Emory University

1

CS 572: Information Retrieval

Personalized Search

Acknowledgements

Some slides in this lecture adapted from various IR researchers, credits noted on

each slide.

CS 572: Information Retrieval. Spring 20104/13/2016

Page 2: CS 572: Information Retrieval - Emory University

Today's Plan

Search Personalization:

– Why?

– Who is a “person”?

– When to personalize (or not)

– How to personalize?

– How to evaluate?

CS 572: Information Retrieval. Spring 2010 24/13/2016

Page 3: CS 572: Information Retrieval - Emory University

Personalization is now default

• Google:

https://googleblog.blogspot.com/2009/12/personalized-search-for-everyone.html

• Bing:

http://searchengineland.com/bing-results-get-localized-personalized-64284

4/13/2016 CS 572: Information Retrieval. Spring 2010 3

Page 4: CS 572: Information Retrieval - Emory University

CFP

Paper

4/13/2016 4CS 572: Information Retrieval. Spring 2010

Page 5: CS 572: Information Retrieval - Emory University

Why Personalize

• How good are search results?

• Do people want the same results for a query?

• How to capture variation in user intent?

– Explicitly

– Implicitly

• How can we use what we learn?

4/13/2016 5CS 572: Information Retrieval. Spring 2010

Page 6: CS 572: Information Retrieval - Emory University

Questions

• How good are search results?

• Do people want the same results for a query?

• How to capture variation in user intent?

– Explicitly

– Implicitly

• How can we use what we learn?

Teevan et al., TOCHI 2010

4/13/2016 6CS 572: Information Retrieval. Spring 2010

Page 7: CS 572: Information Retrieval - Emory University

How Good Are Search Results?

0

0.7

1 2 3 4 5 6 7 8 9 10

No

rmal

ized

Gai

n

Rank

Explicit Behavior Content

Lots of relevant

results ranked low

Teevan et al., TOCHI 2010

4/13/2016 7CS 572: Information Retrieval. Spring 2010

Page 8: CS 572: Information Retrieval - Emory University

How Good Are Search Results?

0

0.7

1 2 3 4 5 6 7 8 9 10

No

rmal

ized

Gai

n

Rank

Explicit Behavior Content

Lots of relevant

results ranked low

Behavior data

has presentation

bias

Teevan et al., TOCHI 2010

4/13/2016 8CS 572: Information Retrieval. Spring 2010

Page 9: CS 572: Information Retrieval - Emory University

How Good Are Search Results?

0

0.7

1 2 3 4 5 6 7 8 9 10

No

rmal

ized

Gai

n

Rank

Explicit Behavior Content

Lots of relevant

results ranked low

Content data also

identifies low

results

Behavior data

has presentation

bias

Teevan et al., TOCHI 2010

4/13/2016 9CS 572: Information Retrieval. Spring 2010

Page 10: CS 572: Information Retrieval - Emory University

Do People Want the Same Results?

• What’s best for

– For you?

– For everyone?

• When it’s just you, can rank perfectly

• With many people, ranking must be a compromise

personalization research?

Teevan et al., TOCHI 2010

4/13/2016 10CS 572: Information Retrieval. Spring 2010

Page 11: CS 572: Information Retrieval - Emory University

Do People Want the Same Results?

0.55

0.7

0.85

1

1 2 3 4 5 6

No

rmal

ized

DC

G

Number of People in Group

Group Individual Web

Potential for

Personalization

Teevan et al., TOCHI 2010

4/13/2016 11CS 572: Information Retrieval. Spring 2010

Page 12: CS 572: Information Retrieval - Emory University

How to Capture Variation (1)?

0.55

0.7

0.85

1

1 2 3 4 5 6

No

rmal

ized

DC

G

Number of People in Group

Explicit Behavior Content

Behavior gap

smaller because

of presentation

bias

Teevan et al., TOCHI 2010

4/13/2016 12CS 572: Information Retrieval. Spring 2010

Page 13: CS 572: Information Retrieval - Emory University

How to Capture Variation (2)?

0.55

0.7

0.85

1

1 2 3 4 5 6

No

rmal

ized

DC

G

Number of People in Group

Explicit Behavior Content

Content data

shows more

variation than

explicit judgments

Behavior gap

smaller because

of presentation

bias

Teevan et al., TOCHI 2010

4/13/2016 13CS 572: Information Retrieval. Spring 2010

Page 14: CS 572: Information Retrieval - Emory University

What to personalize?

• Identify ambiguous queries

• Solicit more information about need

• Personalize search

– Using content and behavior-based measures

0.52

0.54

0.56

0.58

0.6

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1

No

rmal

ized

DC

G

Content Behavior

Web

Personalized

4/13/2016 14CS 572: Information Retrieval. Spring 2010

Page 15: CS 572: Information Retrieval - Emory University

Authors Tutorials New content: WSDM 2012

Attending Workshops to be held February 9-12 in

Sponsors Conference Venue Seattle, WA.

www.wsdm2011.org

33% queries repeated

73% of those are

navigational[Teevan et al. SIGIR 2007]

Page 16: CS 572: Information Retrieval - Emory University

Identifying Personal Navigation

• Repeat queries are often navigational

• The same navigation used over and over again

• Was there a unique click on the same result the last 2 times the person issued the query?

wsdmwsdm

hong kong

wsdm

wsdm

sheraton

sigir

wsdm cfp

Page 17: CS 572: Information Retrieval - Emory University

Understanding Personal Navigation

• Identified millions of navigation queries

– Most occur fewer than 25 times in the logs

– 15% of the query volume

• Queries more ambiguous

– Rarely contain a URL fragment

– Click entropy the same as for general Web queries

– enquirer (multiple meanings)

– bed bugs (found navigation)

– etsy (serendipitous encounters)

National Enquirer

Cincinnati Enquirerhttp://www.medicinenet.com/bed_bugs/article.ht

m

[Informational]

Etsy.com

Regretsy.com

(parody)

Page 18: CS 572: Information Retrieval - Emory University

People Express Things Differently

• Differences can be a challenge for Web search

– Picture of a man handing over a key.

– Oil painting of the surrender of Breda.

Page 19: CS 572: Information Retrieval - Emory University

People Express Things Differently

• Differences can be a challenge for Web search

– Picture of a man handing over a key.

– Oil painting of the surrender of Breda.

• Personalization

– Closes the gap using more about the person

• Groupization

– Closes the gap using more about the group

Page 20: CS 572: Information Retrieval - Emory University

Person ~= Group

4/13/2016 CS 572: Information Retrieval. Spring 2010 20

Page 21: CS 572: Information Retrieval - Emory University

How to Take Advantage of Groups?

• Who do we share interests with?

• Do we talk about things similarly?

• What algorithms should we use?

Page 22: CS 572: Information Retrieval - Emory University

Interested in Many Group Types

• Group longevity

– Task-based

– Trait-based

• Group identification

– Explicit

– Implicit

Task-based Trait-

based Longevity

Identificatio

nIm

plic

itE

xp

licit Task

AgeGende

rJob team

Job role

Location Interest

groupRelevance judgments

Query

selection

Desktop content

Page 23: CS 572: Information Retrieval - Emory University

People Studied

Trait-based dataset

• 110 people

– Work

– Interests

– Demographics

• Microsoft employees

Task-based dataset

• 10 groups x 3 (= 30)

• Know each other

• Have common task

– “Find economic pros and cons of telecommuting”

– “Search for information about companies offering learning services to corporate customers”

Page 24: CS 572: Information Retrieval - Emory University

Queries Studied

Trait-based dataset

• Challenge

– Overlapping queries

– Natural motivation

• Queries picked from 12

– Workc# delegates, live meeting

– Interests

bread recipes, toilet train dog

Task-based dataset

• Common task

– Telecommuting v. office

pros and cons of working in an office

social comparison telecommuting versus office

telecommuting

working at home cost benefit

Page 25: CS 572: Information Retrieval - Emory University

Data Collected

• Queries evaluated

• Explicit relevance judgments

– 20 - 40 results

– Personal relevance

• Highly relevant

• Relevant

• Not relevant

• User profile: Desktop index

Page 26: CS 572: Information Retrieval - Emory University

Answering the Questions

• Who do we share interests with?

• Do we talk about things similarly?

• What algorithms should we use?

Page 27: CS 572: Information Retrieval - Emory University

Who do we share interests with?

• Variation in query selection

– Work groups selected similar work queries

– Social groups selected similar social queries

• Variation in relevance judgments

– Judgments varied greatly (κ=0.08)

– Task-based groups most similar

– Similar for one query ≠ similar for another

Page 28: CS 572: Information Retrieval - Emory University

In task group Not in group Difference

0.42 0.31 34%

In task group Not in group Difference

All queries 0.42 0.31 34%

Group queries 0.77 0.35 120%

Do we talk about things similarly?

• Group profile similarity

– Members more similar to each other than others

– Most similar for aspects related to the group

• Clustering profiles recreates groups

• Index similarity ≠ judgment similarity

– Correlation coefficient of 0.09

Page 29: CS 572: Information Retrieval - Emory University

What algorithms should we use?

• Calculate personalized score for each member

– Content: User profile as relevance feedback

– Behavior: Previously visited URLs and domains

– [Teevan et al. 2005]

• Sum personalized scores across group

• Produces same ranking for all members

(ri+0.5)(N-ni-R+ri+0.5)

(ni-ri+0.5)(R-ri+0.5)tfi

logΣterms i

Page 30: CS 572: Information Retrieval - Emory University

Performance: Task-Based Groups

• Personalization improves on Web

• Groupization gains +5%

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Web GroupizationWeb Personalized

Groupized

Page 31: CS 572: Information Retrieval - Emory University

Performance: Task-Based Groups

• Personalization improves on Web

• Groupization gains +5%

• Split by query type

– On-task v. off-task

– Groupization the same as personalization for off-task queries

– 11% improvement for on-task queries

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

Web Groupization

Off-t

ask q

ueries

On-t

ask q

ueries

Web Personalized

Groupized

Page 32: CS 572: Information Retrieval - Emory University

Performance: Trait-Based Groups

0.45

0.5

0.55

0.6

0.65

0.7

0.75

No

rmal

ize

d D

CG

Groupization

Personalizatio

n

Interest

s

Work

Page 33: CS 572: Information Retrieval - Emory University

Performance: Trait-Based Groups

0.45

0.5

0.55

0.6

0.65

0.7

0.75

No

rmal

ize

d D

CG

Groupization

Personalizatio

n

Work

queries

Interes

t

querie

s

Interest

s

Work

Page 34: CS 572: Information Retrieval - Emory University

Performance: Trait-Based Groups

0.45

0.5

0.55

0.6

0.65

0.7

0.75

No

rmal

ize

d D

CG

Groupization

Personalizatio

n

Work

queries

Interes

t

querie

s

Interest

s

Work

Page 35: CS 572: Information Retrieval - Emory University

Local Search (Geographical Personalization)

35

• Location provides personalization info & context

• Local search uses geographic information to modify the ranking of search results

– location derived from the query text

– location of the device where the query originated

• e.g.,

– “underworld 3 cape cod”

– “underworld 3” from mobile device in Hyannis

4/13/2016 CS 572: Information Retrieval. Spring 2010

Page 36: CS 572: Information Retrieval - Emory University

“Pizza Emory”

Geography and Query Intent

IP address / profile zip

“Pizza Lavista Road, Atlanta, GA”query1

Distance 1:

home–query intent Distance 2:

Reformulation

distance

query2Location 2: Home address

Location 1: query location

Location 3: query location

[ Baeza-Yates and Jones] 2008

4/13/2016 36CS 572: Information Retrieval. Spring 2010

Page 37: CS 572: Information Retrieval - Emory University

37

Topic-Distance Profiles

• 20 bins– 0 distance

– Equal fractions of the rest of the data

• Does distribution into distance bins topics vary by topic?Movie theater Distant places Near-by

movie theater maps

[ Baeza-Yates and Jones] 2008

restaurant

4/13/2016 CS 572: Information Retrieval. Spring 2010

Page 38: CS 572: Information Retrieval - Emory University

How: Approaches

1. Pitkow et al., 2002

2. Qiu et al., 2006

3. Jeh et al., 2003

4. Teevan et al., 2005

5. Das et al., 2007

38

Figure adapted from: Personalized search on the world wide web, by

Micarelli, A. and Gasparetti, F. and Sciarrone, F. and Gauch, S., LNCS 2007

1

2 4

5

3

Teevan et al., TOCHI 2010

4/13/2016 CS 572: Information Retrieval. Spring 2010

Page 39: CS 572: Information Retrieval - Emory University

How: System Perspective

4/13/2016 CS 572: Information Retrieval. Spring 2010 39

Jin-young Kim

Page 40: CS 572: Information Retrieval - Emory University

Profile Info

4/13/2016 CS 572: Information Retrieval. Spring 2010 40

Page 41: CS 572: Information Retrieval - Emory University

What’s in User Profile?

4/13/2016 CS 572: Information Retrieval. Spring 2010 41

Page 42: CS 572: Information Retrieval - Emory University

Profile: Cont’d

4/13/2016 CS 572: Information Retrieval. Spring 2010 42

Page 43: CS 572: Information Retrieval - Emory University

Profile Representation: Details

4/13/2016 CS 572: Information Retrieval. Spring 2010 43

Page 44: CS 572: Information Retrieval - Emory University

Server-Side vs. Client-Side Profile

4/13/2016 CS 572: Information Retrieval. Spring 2010 44

Page 45: CS 572: Information Retrieval - Emory University

Using Profiles: OverviewP. Brusilovsky

Page 46: CS 572: Information Retrieval - Emory University

Pre-Process: Query Expansion

• User profile is applied to add terms to the query– Popular terms could be added to introduce context

– Similar terms could be added to resolve indexer-user mismatch

– Related terms could be added to resolve ambiguity

– Works with any IR model or search engine

Page 47: CS 572: Information Retrieval - Emory University

User related documents

(desktop documents)

containing the query

Score and extract

keywords

Top

query-dependent,

user-biased

keywords

Extract

query expansion or

re-ranking terms

[ Chirita, Firan, Nejdl. Summarizing local context to personalize

global web search. CIKM 2006 ]

Simple Context-based Query Expansion: Chirita et al. 2006

Page 48: CS 572: Information Retrieval - Emory University

How: Approaches

1. Pitkow et al., 2002

2. Qiu et al., 2006

3. Jeh et al., 2003

4. Teevan et al., 2005

5. Das et al., 2007

48

Figure adapted from: Personalized search on the world wide web, by

Micarelli, A. and Gasparetti, F. and Sciarrone, F. and Gauch, S., LNCS 2007

1

2 4

5

3

Teevan et al., TOCHI 2010

4/13/2016 CS 572: Information Retrieval. Spring 2010

Page 49: CS 572: Information Retrieval - Emory University

PS Search Engine:Teevan et al., SIGIR 2005

Page 50: CS 572: Information Retrieval - Emory University

PS Search Engine

dog cat

monkey

banana

food

baby

infant

child boy

girl

forest

hiking

walking

gorpbaby

infant

child boy

girl

csail mit

artificial

research

robotweb

search

retrieval ir

hunt

Page 51: CS 572: Information Retrieval - Emory University

1.6 0.2

6.0

0.2 2.7

1.3

PS Search Engine

Search

results

page

web

search

retrieval

ir hunt

1.3

Page 52: CS 572: Information Retrieval - Emory University

Calculating a Document’s Score

• Based on standard tf.idf

Score = Σ tfi * wi

web

search

retrieval

ir hunt

1.3

Page 53: CS 572: Information Retrieval - Emory University

Calculating a Document’s Score

• Based on standard tf.idf

Score = Σ tfi * wi

Σ0.1

0.05

0.5 0.35

0.3

1.3

Page 54: CS 572: Information Retrieval - Emory University

0.3 0.7

0.1 0.23

0.6 0.6

0.002

0.7 0.1

0.01 0.6

0.2 0.8

0.1 0.001

0.3 0.4

0.1 0.7

0.001

0.23 0.6

0.1 0.7

0.001

0.23 0.6

0.1

0.05

0.5 0.35

0.3

0.1

0.05

0.5 0.35

0.3

N

ni

Calculating a Document’s Score

• Based on standard tf.idf

Score = Σ tfi * wi

World(N)

(ni)wi = log

Σ 1.3

Page 55: CS 572: Information Retrieval - Emory University

0.002

0.7 0.1

0.01 0.6

0.002

0.7 0.1

0.01 0.6

0.3 0.7

0.1 0.23

0.6 0.6

0.002

0.7 0.1

0.01 0.6

0.2 0.8

0.1 0.001

0.3 0.4

0.1 0.7

0.001

0.23 0.6

0.1 0.7

0.001

0.23 0.6

0.1

0.05

0.5 0.35

0.3

N

ni

Calculating a Document’s Score

• Based on standard tf.idf

Score = Σ tfi * wi

(N)

(ni)wi = logWorld

ri

R

(ri+0.5)(N-ni-R+ri+0.5)

(ni-ri+0.5)(R-ri+0.5)wi = log

† From Sparck Jones, Walker

and Roberson, 1998 [21].

Where: N = N+R, ni = ni+ri’’

Client

(ri+0.5)(N-ni-R+ri+0.5)

(ni-ri+0.5)(R-ri+0.5)wi = log

Page 56: CS 572: Information Retrieval - Emory University

Finding the Parameter Values

• Corpus representation (N, ni)– How common is the term in general?

– Web vs. result set

• User representation (R, ri)– How well does it represent the user’s interest?

– All vs. recent vs. Web vs. queries vs. none

• Document representation– What terms to sum over?

– Full document vs. snippet

web

search

retrieval

ir hunt

Page 57: CS 572: Information Retrieval - Emory University

Building a Test Bed

• 15 evaluators x ~10 queries

– 131 queries total

• Personally meaningful queries

– Selected from a list

– Queries issued earlier (kept diary)

• Evaluate 50 results for each query

– Highly relevant / relevant / irrelevant

• Index of personal information

Page 58: CS 572: Information Retrieval - Emory University

Evaluating Personalized Search

• Measure algorithm quality

– DCG(i) = {

• Look at one parameter at a time

– 67 different parameter combinations!

– Hold other parameters constant and vary one

• Look at best parameter combination

– Compare with various baselines

Gain(i),

DCG(i–1) + Gain(i)/log(i),

if i = 1

otherwise

Page 59: CS 572: Information Retrieval - Emory University

Analysis of Parameters

0.27

0.28

0.29

0.3

0.31

0.32

0.33

0.34

0.35F

ull t

ex

t

We

b

Sn

ipp

et

No

ne

Qu

ery

We

b

Re

ce

nt

All

Sn

ipp

et

Fu

ll t

ex

t

User

Page 60: CS 572: Information Retrieval - Emory University

Analysis of Parameters

0.27

0.28

0.29

0.3

0.31

0.32

0.33

0.34

0.35F

ull t

ex

t

We

b

Sn

ipp

et

No

ne

Qu

ery

We

b

Re

ce

nt

All

Sn

ipp

et

Fu

ll t

ex

t

Corpus User Document

Page 61: CS 572: Information Retrieval - Emory University

PS Improves Text Retrieval

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

No RF PS Web Combo

DC

G

No model

Relevance

Feedback

Personalized

Search

0.37

0.41

0.46

Page 62: CS 572: Information Retrieval - Emory University

Text Features Not Enough

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

No RF PS Web Combo

DC

G

0.37

0.41

0.46

0.56

Page 63: CS 572: Information Retrieval - Emory University

Take Advantage of Web Ranking

0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

No RF PS Web Combo

DC

G

0.37

0.41

0.46

0.560.58

PS+Web

Page 64: CS 572: Information Retrieval - Emory University

Summary

• Personalization of Web search

– Result re-ranking

– User’s documents as relevance feedback

• Rich representations important

– Rich user profile particularly important

– Efficiency hacks possible

– Need to incorporate features beyond text

Page 65: CS 572: Information Retrieval - Emory University

Much Room for Improvement

• Group ranking

– Best improves on Web by 23%

– More people Less improvement

• Personal ranking

– Best improves on Web by 38%

– Remains constant

0.8

0.85

0.9

0.95

1

1.05

1 2 3 4 5 6

Number of People

DC

G

Personalized Group

Potential for

Personalization

Page 66: CS 572: Information Retrieval - Emory University

User Interface Issues

• Make personalization transparent

• Give user control over personalization

– Slider between Web and personalized results

– Allows for background computation

• Exacerbates problem with re-finding

– Results change as user model changes

– Thesis research – Re:Search Engine

Page 67: CS 572: Information Retrieval - Emory University

Post-Filter: Annotations

• The result could be relevant to the user in several aspects. Fusing this relevance with query relevance is error prone and leads to a loss of data

• Results are ranked by the query relevance, but annotated with visual cues reflecting other kinds of relevance– User interests - Syskill and Webert, group

interests - KnowledgeSea

Page 68: CS 572: Information Retrieval - Emory University

Example: Syskill and Webert

• First example of annotation

• Post-filter to Lycos

• Hot, cold, lukewarm• Pazzani, M., Muramatsu, J., and Billsus, D. (1996)

Syskill & Webert: Identifying interesting Web sites. In: Proceedings of the Thirteen National Conference on Artificial Intelligence, AAAI'96, Portland, OR, August, 1996, AAAI Press /MIT Press, pp. 54-61, also available at http://www.ics.uci.edu/~pazzani/Syskill.html.

Page 69: CS 572: Information Retrieval - Emory University

YourNews: Adaptive Search and Filtering with Open User Profile

Page 70: CS 572: Information Retrieval - Emory University

User Control in Personalization (RF)

70

J-S. Ahn, P. Brusilovsky, D. He, and S.Y. Syn. Open user profiles for

adaptive news systems: Help or harm? WWW 2007

4/13/2016 CS 572: Information Retrieval. Spring 2010

Page 71: CS 572: Information Retrieval - Emory University

TaskSieve: User-Controled Adaptive Search

Page 72: CS 572: Information Retrieval - Emory University

TaskSieve: Adaptive Snippets

• The goal of a usual snippet is to show query relevance. TaskSieve applies adaptive snippets to show profile relevance as well

• Selects top 3 relevant sentences combining query relevance and task relevance to sentences

• Applies color coding by query/profile

Page 73: CS 572: Information Retrieval - Emory University

Evaluating Personalized Search

• Explicit judgments (offline and in situ)– Evaluate components before system

– NOTE: What’s relevant for you

• Deploy system– Verbatim feedback, Questionnaires, etc.

– Measure behavioral interactions (e.g., click, reformulation, abandonment, etc.)

– Click biases –order, presentation, etc.

– Interleaving for unbiased clicks

• Link implicit and explicit (Curious Browser toolbar)

• From single query to search sessions and beyond

734/13/2016 CS 572: Information Retrieval. Spring 2010

Page 74: CS 572: Information Retrieval - Emory University

Evaluating Personalized Search: Active Area of Research

74

Page 75: CS 572: Information Retrieval - Emory University

Personalization Summary

• Lots of relevant content ranked low

• Potential for personalization high

• Implicit measures capture explicit variation

– Behavior-based: Highly accurate

– Content-based: Lots of variation

• Example: Personalized Search

– Behavior + content work best together

– Improves search result click through

4/13/2016 75CS 572: Information Retrieval. Spring 2010

Page 76: CS 572: Information Retrieval - Emory University

Resources

• Diane Kelly and Jaime Teevan. Implicit Feedback for Inferring User Preference: A Bibliographyhttp://www.csail.mit.edu/~teevan/work/publications/papers/sigir-forum03.pdf

• Jaime Teevan, Susan T. Dumais and Eric Horvitz. Potential for Personalization. ACM Transactions on Computer-Human Interaction (TOCHI), 2010.http://people.csail.mit.edu/teevan/work/publications/papers/tochi10.pdf

• Web Spam Taxonomy (paper):http://infolab.stanford.edu/~zoltan/publications/gyongyi2005web.pdf

• SciGen: http://pdos.csail.mit.edu/scigen/

CS 572: Information Retrieval. Spring 2010 764/13/2016