+ All Categories
Home > Documents > Lecture 16: IR Components 2

Lecture 16: IR Components 2

Date post: 20-Jan-2016
Category:
Upload: chipo
View: 27 times
Download: 0 times
Share this document with a friend
Description:
Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00 pm Spring 2007 http://courses.ischool.berkeley.edu/i240/s07. Lecture 16: IR Components 2. Principles of Information Retrieval. Overview. Review IR Components - PowerPoint PPT Presentation
77
2007.03.15 - SLIDE 1 IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00 pm Spring 2007 http://courses.ischool.berkeley.edu/i240/s07 Principles of Information Retrieval Lecture 16: IR Components 2
Transcript
Page 1: Lecture 16: IR Components 2

2007.03.15 - SLIDE 1IS 240 – Spring 2007

Prof. Ray Larson University of California, Berkeley

School of InformationTuesday and Thursday 10:30 am - 12:00 pm

Spring 2007http://courses.ischool.berkeley.edu/i240/s07

Principles of Information Retrieval

Lecture 16: IR Components 2

Page 2: Lecture 16: IR Components 2

2007.03.15 - SLIDE 2IS 240 – Spring 2007

Overview

• Review– IR Components– Text Processing and Stemming

• Relevance Feedback

Page 3: Lecture 16: IR Components 2

2007.03.15 - SLIDE 3IS 240 – Spring 2007

Stemming and Morphological Analysis

• Goal: “normalize” similar words

• Morphology (“form” of words)– Inflectional Morphology

• E.g,. inflect verb endings and noun number• Never change grammatical class

– dog, dogs– tengo, tienes, tiene, tenemos, tienen

– Derivational Morphology • Derive one word from another, • Often change grammatical class

– build, building; health, healthy

Page 4: Lecture 16: IR Components 2

2007.03.15 - SLIDE 4IS 240 – Spring 2007

Simple “S” stemming

• IF a word ends in “ies”, but not “eies” or “aies”– THEN “ies” “y”

• IF a word ends in “es”, but not “aes”, “ees”, or “oes”– THEN “es” “e”

• IF a word ends in “s”, but not “us” or “ss”– THEN “s” NULL

Harman, JASIS 1991

Page 5: Lecture 16: IR Components 2

2007.03.15 - SLIDE 5IS 240 – Spring 2007

Stemmer Examples

The SMART stemmer The Porter stemmer The IAGO! stemmer% tstem ateate% tstem applesappl% tstem formulaeformul% tstem appendicesappendix% tstem implementationimple% tstem glassesglass%

% pstemmer ateat% pstemmer applesappl% pstemmer formulaeformula% pstemmer appendicesappendic% pstemmer implementationimplement% pstemmer glassesglass%

% stemate|2eat|2apples|1apple|1formulae|1formula|1appendices|1appendix|1implementation|1implementation|1glasses|1glasses|1%

Page 6: Lecture 16: IR Components 2

2007.03.15 - SLIDE 6IS 240 – Spring 2007

Errors Generated by Porter Stemmer (Krovetz 93)

Too Aggressive Too Timidorganization/ organ european/ europe

policy/ police cylinder/ cylindrical

execute/ executive create/ creation

arm/ army search/ searcher

Page 7: Lecture 16: IR Components 2

2007.03.15 - SLIDE 7IS 240 – Spring 2007

Automated Methods

• Stemmers:– Very dumb rules work well (for English)– Porter Stemmer: Iteratively remove suffixes– Improvement: pass results through a lexicon

• Newer stemmers are configurable (Snowball)• Powerful multilingual tools exist for

morphological analysis– PCKimmo, Xerox Lexical technology– Require a grammar and dictionary– Use “two-level” automata– Wordnet “morpher”

Page 8: Lecture 16: IR Components 2

2007.03.15 - SLIDE 8IS 240 – Spring 2007

Wordnet

• Type “wn word” on irony.

• Large exception dictionary:

• Demo

aardwolves aardwolf abaci abacus abacuses abacus abbacies abbacy abhenries abhenry abilities ability abkhaz abkhaz abnormalities abnormality aboideaus aboideau aboideaux aboideau aboiteaus aboiteau aboiteaux aboiteau abos abo abscissae abscissa abscissas abscissa absurdities absurdity…

Page 9: Lecture 16: IR Components 2

2007.03.15 - SLIDE 9IS 240 – Spring 2007

Using NLP

• Strzalkowski (in Reader)

Text NLP represDbasesearch

TAGGERNLP: PARSER TERMS

Page 10: Lecture 16: IR Components 2

2007.03.15 - SLIDE 10IS 240 – Spring 2007

Using NLP

INPUT SENTENCEThe former Soviet President has been a local hero ever sincea Russian tank invaded Wisconsin.

TAGGED SENTENCEThe/dt former/jj Soviet/jj President/nn has/vbz been/vbn a/dt local/jj hero/nn ever/rb since/in a/dt Russian/jj tank/nn invaded/vbd Wisconsin/np ./per

Page 11: Lecture 16: IR Components 2

2007.03.15 - SLIDE 11IS 240 – Spring 2007

Using NLP

TAGGED & STEMMED SENTENCEthe/dt former/jj soviet/jj president/nn have/vbz be/vbn a/dt local/jj hero/nn ever/rb since/in a/dt russian/jj tank/nn invade/vbd wisconsin/np ./per

Page 12: Lecture 16: IR Components 2

2007.03.15 - SLIDE 12IS 240 – Spring 2007

Using NLP

PARSED SENTENCE

[assert

[[perf [have]][[verb[BE]]

[subject [np[n PRESIDENT][t_pos THE]

[adj[FORMER]][adj[SOVIET]]]]

[adv EVER]

[sub_ord[SINCE [[verb[INVADE]]

[subject [np [n TANK][t_pos A]

[adj [RUSSIAN]]]]

[object [np [name [WISCONSIN]]]]]]]]]

Page 13: Lecture 16: IR Components 2

2007.03.15 - SLIDE 13IS 240 – Spring 2007

Using NLP

EXTRACTED TERMS & WEIGHTS

President 2.623519 soviet 5.416102

President+soviet 11.556747 president+former 14.594883

Hero 7.896426 hero+local 14.314775

Invade 8.435012 tank 6.848128

Tank+invade 17.402237 tank+russian 16.030809

Russian 7.383342 wisconsin 7.785689

Page 14: Lecture 16: IR Components 2

2007.03.15 - SLIDE 14IS 240 – Spring 2007

Same Sentence, different sys

INPUT SENTENCEThe former Soviet President has been a local hero ever sincea Russian tank invaded Wisconsin.

TAGGED SENTENCE (using uptagger from Tsujii)The/DT former/JJ Soviet/NNP President/NNP has/VBZ been/VBN a/DT local/JJ hero/NN ever/RB since/IN a/DT Russian/JJ tank/NN invaded/VBD Wisconsin/NNP ./.

Page 15: Lecture 16: IR Components 2

2007.03.15 - SLIDE 15IS 240 – Spring 2007

Same Sentence, different sys

CHUNKED Sentence (chunkparser – Tsujii)(TOP (S (NP (DT The) (JJ former) (NNP Soviet) (NNP President) ) (VP (VBZ has) (VP (VBN been) (NP (DT a) (JJ local) (NN hero) ) (ADVP (RB ever) ) (SBAR (IN since) (S (NP (DT a) (JJ Russian) (NN tank) ) (VP (VBD invaded) (NP (NNP Wisconsin) ) ) ) ) ) ) (. .) ) )

Page 16: Lecture 16: IR Components 2

2007.03.15 - SLIDE 16IS 240 – Spring 2007

Same Sentence, different sys

Enju ParserROOT ROOT ROOT ROOT -1 ROOT been be VBN VB 5been be VBN VB 5 ARG1 President president NNP NNP 3been be VBN VB 5 ARG2 hero hero NN NN 8a a DT DT 6 ARG1 hero hero NN NN 8a a DT DT 11 ARG1 tank tank NN NN 13local local JJ JJ 7 ARG1 hero hero NN NN 8The the DT DT 0 ARG1 President president NNP NNP 3former former JJ JJ 1 ARG1 President president NNP NNP 3Russian russian JJ JJ 12 ARG1 tank tank NN NN 13Soviet soviet NNP NNP 2 MOD President president NNP NNP 3invaded invade VBD VB 14 ARG1 tank tank NN NN 13invaded invade VBD VB 14 ARG2 Wisconsin wisconsin NNP NNP 15has have VBZ VB 4 ARG1 President president NNP NNP 3has have VBZ VB 4 ARG2 been be VBN VB 5since since IN IN 10 MOD been be VBN VB 5since since IN IN 10 ARG1 invaded invade VBD VB 14ever ever RB RB 9 ARG1 since since IN IN 10

Page 17: Lecture 16: IR Components 2

2007.03.15 - SLIDE 17IS 240 – Spring 2007

Assumptions in IR

• Statistical independence of terms

• Dependence approximations

Page 18: Lecture 16: IR Components 2

2007.03.15 - SLIDE 18IS 240 – Spring 2007

Statistical Independence

Two events x and y are statistically independent if the product of their probability of their happening individually equals their probability of happening together.

),()()( yxPyPxP

Page 19: Lecture 16: IR Components 2

2007.03.15 - SLIDE 19IS 240 – Spring 2007

Statistical Independence and Dependence

• What are examples of things that are statistically independent?

• What are examples of things that are statistically dependent?

Page 20: Lecture 16: IR Components 2

2007.03.15 - SLIDE 20IS 240 – Spring 2007

Statistical Independence vs. Statistical Dependence

• How likely is a red car to drive by given we’ve seen a black one?

• How likely is the word “ambulence” to appear, given that we’ve seen “car accident”?

• Color of cars driving by are independent (although more frequent colors are more likely)

• Words in text are not independent (although again more frequent words are more likely)

Page 21: Lecture 16: IR Components 2

2007.03.15 - SLIDE 21IS 240 – Spring 2007

Lexical Associations

• Subjects write first word that comes to mind– doctor/nurse; black/white (Palermo & Jenkins 64)

• Text Corpora yield similar associations• One measure: Mutual Information (Church and

Hanks 89)

• If word occurrences were independent, the numerator and denominator would be equal (if measured across a large collection)

)()(

),(log),( 2 yPxP

yxPyxI

Page 22: Lecture 16: IR Components 2

2007.03.15 - SLIDE 22IS 240 – Spring 2007

Interesting Associations with “Doctor”

I(x,y) f(x,y) f(x) x f(y) y11.3 12 111 Honorary 621 Doctor

11.3 8 1105 Doctors 44 Dentists

10.7 30 1105 Doctors 241 Nurses

9.4 8 1105 Doctors 154 Treating

9.0 6 275 Examined 621 Doctor

8.9 11 1105 Doctors 317 Treat

8.7 25 621 Doctor 1407 Bills

(AP Corpus, N=15 million, Church & Hanks 89)

Page 23: Lecture 16: IR Components 2

2007.03.15 - SLIDE 23IS 240 – Spring 2007

I(x,y) f(x,y) f(x) x f(y) y0.96 6 621 doctor 73785 with

0.95 41 284690 a 1105 doctors

0.93 12 84716 is 1105 doctors

These associations were likely to happen because the non-doctor words shown here are very commonand therefore likely to co-occur with any noun.

Un-Interesting Associations with “Doctor”

Page 24: Lecture 16: IR Components 2

2007.03.15 - SLIDE 24IS 240 – Spring 2007

Today

• Relevance Feedback– aka query modification– aka “more like this”

Page 25: Lecture 16: IR Components 2

2007.03.15 - SLIDE 25IS 240 – Spring 2007

IR Components

• A number of techniques have been shown to be potentially important or useful for effective IR (in TREC-like evaluations)

• Today and over the next couple weeks (except for Spring Break) we will look at these components of IR systems and their effects on retrieval

• These include: Relevance Feedback, Latent Semantic Indexing, clustering, and application of NLP techniques in term extraction and normalization

Page 26: Lecture 16: IR Components 2

2007.03.15 - SLIDE 26IS 240 – Spring 2007

Querying in IR System

Interest profiles& Queries

Documents & data

Rules of the game =Rules for subject indexing +

Thesaurus (which consists of

Lead-InVocabulary

andIndexing

Language

StorageLine

Potentially Relevant

Documents

Comparison/Matching

Store1: Profiles/Search requests

Store2: Documentrepresentations

Indexing (Descriptive and

Subject)

Formulating query in terms of

descriptors

Storage of profiles

Storage of Documents

Information Storage and Retrieval System

Page 27: Lecture 16: IR Components 2

2007.03.15 - SLIDE 27IS 240 – Spring 2007

Relevance Feedback in an IR System

Interest profiles& Queries

Documents & data

Rules of the game =Rules for subject indexing +

Thesaurus (which consists of

Lead-InVocabulary

andIndexing

Language

StorageLine

Potentially Relevant

Documents

Comparison/Matching

Store1: Profiles/Search requests

Store2: Documentrepresentations

Indexing (Descriptive and

Subject)

Formulating query in terms of

descriptors

Storage of profiles

Storage of Documents

Information Storage and Retrieval System

Selected relevant docs

Page 28: Lecture 16: IR Components 2

2007.03.15 - SLIDE 28IS 240 – Spring 2007

Query Modification

• Changing or Expanding a query can lead to better results

• Problem: how to reformulate the query?– Thesaurus expansion:

• Suggest terms similar to query terms

– Relevance feedback:• Suggest terms (and documents) similar to

retrieved documents that have been judged to be relevant

Page 29: Lecture 16: IR Components 2

2007.03.15 - SLIDE 29IS 240 – Spring 2007

Relevance Feedback

• Main Idea:– Modify existing query based on relevance

judgements• Extract terms from relevant documents and add

them to the query• and/or re-weight the terms already in the query

– Two main approaches:• Automatic (psuedo-relevance feedback)• Users select relevant documents

– Users/system select terms from an automatically-generated list

Page 30: Lecture 16: IR Components 2

2007.03.15 - SLIDE 30IS 240 – Spring 2007

Relevance Feedback

• Usually do both:– Expand query with new terms– Re-weight terms in query

• There are many variations– Usually positive weights for terms from

relevant docs– Sometimes negative weights for terms from

non-relevant docs– Remove terms ONLY in non-relevant

documents

Page 31: Lecture 16: IR Components 2

2007.03.15 - SLIDE 31IS 240 – Spring 2007

Rocchio Method

0.25) to and 0.75 to set best to studies some(in terms

t nonrelevan andrelevant of importance thetune and ,

chosen documentsrelevant -non ofnumber the

chosen documentsrelevant ofnumber the

document relevant -non for the vector the

document relevant for the vector the

query initial for the vector the

2

1

0

121101

21

n

n

iS

iR

Q

where

Sn

Rn

QQ

i

i

i

n

i

n

ii

Page 32: Lecture 16: IR Components 2

2007.03.15 - SLIDE 32IS 240 – Spring 2007

Rocchio/Vector Illustration

Retrieval

Information

0.5

1.0

0 0.5 1.0

D1

D2

Q0

Q’

Q”

Q0 = retrieval of information = (0.7,0.3)D1 = information science = (0.2,0.8)D2 = retrieval systems = (0.9,0.1)

Q’ = ½*Q0+ ½ * D1 = (0.45,0.55)Q” = ½*Q0+ ½ * D2 = (0.80,0.20)

Page 33: Lecture 16: IR Components 2

2007.03.15 - SLIDE 33IS 240 – Spring 2007

Example Rocchio Calculation

)04.1,033.0,488.0,022.0,527.0,01.0,002.0,000875.0,011.0(

12

25.0

75.0

1

)950,.00.0,450,.00.0,500,.00.0,00.0,00.0,00.0(

)00.0,020,.00.0,025,.005,.00.0,020,.010,.030(.

)120,.100,.100,.025,.050,.002,.020,.009,.020(.

)120,.00.0,00.0,050,.025,.025,.00.0,00.0,030(.

121

1

2

1

new

new

Q

SRRQQ

Q

S

R

R

Relevantdocs

Non-rel doc

Original Query

Constants

Rocchio Calculation

Resulting feedback query

Page 34: Lecture 16: IR Components 2

2007.03.15 - SLIDE 34IS 240 – Spring 2007

Rocchio Method

• Rocchio automatically– re-weights terms– adds in new terms (from relevant docs)

• have to be careful when using negative terms• Rocchio is not a machine learning algorithm

• Most methods perform similarly– results heavily dependent on test collection

• Machine learning methods are proving to work better than standard IR approaches like Rocchio

Page 35: Lecture 16: IR Components 2

2007.03.15 - SLIDE 35IS 240 – Spring 2007

Probabilistic Relevance Feedback

Document Relevance

Documentindexing

Given a query term t

+ -

+ r n-r n

- R-r N-n-R+r N-n

R N-R N

Where N is the number of documents seenRobertson & Sparck Jones

Page 36: Lecture 16: IR Components 2

2007.03.15 - SLIDE 36IS 240 – Spring 2007

Robertson-Spark Jones Weights

• Retrospective formulation --

rRnNrnrR

r

wnewt log

Page 37: Lecture 16: IR Components 2

2007.03.15 - SLIDE 37IS 240 – Spring 2007

Robertson-Sparck Jones Weights

5.05.05.0

5.0

log)1(

rRnNrnrR

r

w

Predictive formulation

Page 38: Lecture 16: IR Components 2

2007.03.15 - SLIDE 38IS 240 – Spring 2007

Using Relevance Feedback

• Known to improve results– in TREC-like conditions (no user involved)– So-called “Blind Relevance Feedback”

typically uses the Rocchio algorithm with the assumption that the top N documents in an initial retrieval are relevant

• What about with a user in the loop?– How might you measure this?– Let’s examine a user study of relevance

feedback by Koenneman & Belkin 1996.

Page 39: Lecture 16: IR Components 2

2007.03.15 - SLIDE 39IS 240 – Spring 2007

Questions being InvestigatedKoenemann & Belkin 96

• How well do users work with statistical ranking on full text?

• Does relevance feedback improve results?

• Is user control over operation of relevance feedback helpful?

• How do different levels of user control effect results?

Page 40: Lecture 16: IR Components 2

2007.03.15 - SLIDE 40IS 240 – Spring 2007

How much of the guts should the user see?

• Opaque (black box) – (like web search engines)

• Transparent – (see available terms after the r.f. )

• Penetrable – (see suggested terms before the r.f.)

• Which do you think worked best?

Page 41: Lecture 16: IR Components 2

2007.03.15 - SLIDE 41IS 240 – Spring 2007

Page 42: Lecture 16: IR Components 2

2007.03.15 - SLIDE 42IS 240 – Spring 2007

Penetrable…

Terms available for relevance feedback made visible

(from Koenemann & Belkin)

Page 43: Lecture 16: IR Components 2

2007.03.15 - SLIDE 43IS 240 – Spring 2007

Details on User StudyKoenemann & Belkin 96

• Subjects have a tutorial session to learn the system

• Their goal is to keep modifying the query until they’ve developed one that gets high precision

• This is an example of a routing query (as opposed to ad hoc)

• Reweighting:– They did not reweight query terms– Instead, only term expansion

• pool all terms in rel docs• take top N terms, where • n = 3 + (number-marked-relevant-docs*2)• (the more marked docs, the more terms added to the query)

Page 44: Lecture 16: IR Components 2

2007.03.15 - SLIDE 44IS 240 – Spring 2007

Details on User StudyKoenemann & Belkin 96

• 64 novice searchers– 43 female, 21 male, native English

• TREC test bed– Wall Street Journal subset

• Two search topics– Automobile Recalls– Tobacco Advertising and the Young

• Relevance judgements from TREC and experimenter

• System was INQUERY (Inference net system using (mostly) vector methods)

Page 45: Lecture 16: IR Components 2

2007.03.15 - SLIDE 45IS 240 – Spring 2007

Sample TREC query

Page 46: Lecture 16: IR Components 2

2007.03.15 - SLIDE 46IS 240 – Spring 2007

Evaluation

• Precision at 30 documents• Baseline: (Trial 1)

– How well does initial search go?– One topic has more relevant docs than the

other

• Experimental condition (Trial 2)– Subjects get tutorial on relevance feedback– Modify query in one of four modes

• no r.f., opaque, transparent, penetration

Page 47: Lecture 16: IR Components 2

2007.03.15 - SLIDE 47IS 240 – Spring 2007

Precision vs. RF condition (from Koenemann & Belkin 96)

Page 48: Lecture 16: IR Components 2

2007.03.15 - SLIDE 48IS 240 – Spring 2007

Effectiveness Results

• Subjects with R.F. did 17-34% better performance than no R.F.

• Subjects with penetration case did 15% better as a group than those in opaque and transparent cases.

Page 49: Lecture 16: IR Components 2

2007.03.15 - SLIDE 49IS 240 – Spring 2007

Number of iterations in formulating queries (from Koenemann & Belkin 96)

Page 50: Lecture 16: IR Components 2

2007.03.15 - SLIDE 50IS 240 – Spring 2007

Number of terms in created queries (from Koenemann & Belkin 96)

Page 51: Lecture 16: IR Components 2

2007.03.15 - SLIDE 51IS 240 – Spring 2007

Behavior Results

• Search times approximately equal• Precision increased in first few iterations • Penetration case required fewer iterations to

make a good query than transparent and opaque

• R.F. queries much longer– but fewer terms in penetrable case -- users were

more selective about which terms were added in.

Page 52: Lecture 16: IR Components 2

2007.03.15 - SLIDE 52IS 240 – Spring 2007

Relevance Feedback Summary

• Iterative query modification can improve precision and recall for a standing query

• In at least one study, users were able to make good choices by seeing which terms were suggested for R.F. and selecting among them

• So … “more like this” can be useful!

Page 53: Lecture 16: IR Components 2

2007.03.15 - SLIDE 53IS 240 – Spring 2007

Alternative Notions of Relevance Feedback

• Find people whose taste is “similar” to yours. Will you like what they like?

• Follow a users’ actions in the background. Can this be used to predict what the user will want to see next?

• Track what lots of people are doing. Does this implicitly indicate what they think is good and not good?

Page 54: Lecture 16: IR Components 2

2007.03.15 - SLIDE 54IS 240 – Spring 2007

Alternative Notions of Relevance Feedback

• Several different criteria to consider:– Implicit vs. Explicit judgements – Individual vs. Group judgements– Standing vs. Dynamic topics– Similarity of the items being judged vs.

similarity of the judges themselves

Page 55: Lecture 16: IR Components 2

2007.03.15 - SLIDE 55

Collaborative Filtering (social filtering)

• If Pam liked the paper, I’ll like the paper

• If you liked Star Wars, you’ll like Independence Day

• Rating based on ratings of similar people– Ignores the text, so works on text, sound,

pictures etc.– But: Initial users can bias ratings of future

users Sally Bob Chris Lynn KarenStar Wars 7 7 3 4 7Jurassic Park 6 4 7 4 4Terminator II 3 4 7 6 3Independence Day 7 7 2 2 ?

Page 56: Lecture 16: IR Components 2

2007.03.15 - SLIDE 56

Ringo Collaborative Filtering (Shardanand & Maes 95)

• Users rate musical artists from like to dislike– 1 = detest 7 = can’t live without 4 = ambivalent– There is a normal distribution around 4– However, what matters are the extremes

• Nearest Neighbors Strategy: Find similar users and predicted (weighted) average of user ratings

• Pearson r algorithm: weight by degree of correlation between user U and user J– 1 means very similar, 0 means no correlation, -1 dissimilar– Works better to compare against the ambivalent rating (4), rather

than the individual’s average score

22 )()(

))((

JJUU

JJUUrUJ

Page 57: Lecture 16: IR Components 2

2007.03.15 - SLIDE 57IS 240 – Spring 2007

Social Filtering

• Ignores the content, only looks at who judges things similarly

• Works well on data relating to “taste”– something that people are good at predicting

about each other too

• Does it work for topic? – GroupLens results suggest otherwise

(preliminary)– Perhaps for quality assessments– What about for assessing if a document is

about a topic?

Page 58: Lecture 16: IR Components 2

2007.03.15 - SLIDE 58

Learning interface agents

• Add agents in the UI, delegate tasks to them• Use machine learning to improve performance

– learn user behavior, preferences

• Useful when:– 1) past behavior is a useful predictor of the future– 2) wide variety of behaviors amongst users

• Examples: – mail clerk: sort incoming messages in right mailboxes– calendar manager: automatically schedule meeting

times?

Page 59: Lecture 16: IR Components 2

2007.03.15 - SLIDE 59IS 240 – Spring 2007

Example Systems

• Example Systems– Newsweeder– Letizia– WebWatcher– Syskill and Webert

• Vary according to– User states topic or not– User rates pages or not

Page 60: Lecture 16: IR Components 2

2007.03.15 - SLIDE 60

NewsWeeder (Lang & Mitchell)

• A netnews-filtering system

• Allows the user to rate each article read from one to five

• Learns a user profile based on these ratings

• Use this profile to find unread news that interests the user.

Page 61: Lecture 16: IR Components 2

2007.03.15 - SLIDE 61

Letizia (Lieberman 95)

• Recommends web pages during browsing based on user profile

• Learns user profile using simple heuristics • Passive observation, recommend on request• Provides relative ordering of link interestingness

• Assumes recommendations “near” current page are more valuable than others

user letizia

user profile

heuristics recommendations

Page 62: Lecture 16: IR Components 2

2007.03.15 - SLIDE 62IS 240 – Spring 2007

Letizia (Lieberman 95)

• Infers user preferences from behavior• Interesting pages

– record in hot list– save as a file– follow several links from pages– returning several times to a document

• Not Interesting– spend a short time on document– return to previous document without following links– passing over a link to document (selecting links above

and below document)

Page 63: Lecture 16: IR Components 2

2007.03.15 - SLIDE 63

WebWatcher (Freitag et al.)

• A "tour guide" agent for the WWW. – User tells it what kind of information is wanted– System tracks web actions– Highlights hyperlinks that it computes will be

of interest.

• Strategy for giving advice is learned from feedback from earlier tours. – Uses WINNOW as a learning algorithm

Page 64: Lecture 16: IR Components 2

2007.03.15 - SLIDE 64

Page 65: Lecture 16: IR Components 2

2007.03.15 - SLIDE 65

Syskill & Webert (Pazzani et al 96)

• User defines topic page for each topic• User rates pages (cold or hot) • Syskill & Webert creates profile with

Bayesian classifier– accurate– incremental– probabilities can be used for ranking of

documents– operates on same data structure as picking

informative features• Syskill & Webert rates unseen pages

Page 66: Lecture 16: IR Components 2

2007.03.15 - SLIDE 66

Rating Pages

Page 67: Lecture 16: IR Components 2

2007.03.15 - SLIDE 67

Advantages

• Less work for user and application writer– compare w/ other agent approaches

• no user programming• significant a priori domain-specific and user

knowledge not required

• Adaptive behavior– agent learns user behavior, preferences over

time

• Model built gradually

Page 68: Lecture 16: IR Components 2

2007.03.15 - SLIDE 68

Consequences of passiveness

• Weak heuristics– click through multiple uninteresting pages en

route to interestingness– user browses to uninteresting page, heads to

nefeli for a coffee– hierarchies tend to get more hits near root

• No ability to fine-tune profile or express interest without visiting “appropriate” pages

Page 69: Lecture 16: IR Components 2

2007.03.15 - SLIDE 69

Open issues

• How far can passive observation get you?– for what types of applications is passiveness

sufficient?

• Profiles are maintained internally and used only by the application. some possibilities:– expose to the user (e.g. fine tune profile) ?– expose to other applications (e.g. reinforce belief)?– expose to other users/agents (e.g. collaborative

filtering)?– expose to web server (e.g. cnn.com custom news)?

• Personalization vs. closed applications• Others?

Page 70: Lecture 16: IR Components 2

2007.03.15 - SLIDE 70IS 240 – Spring 2007

Relevance Feedback on Non-Textual Information

• Image Retrieval

• Time-series Patterns

Page 71: Lecture 16: IR Components 2

2007.03.15 - SLIDE 71

MARS (Riu et al. 97)

Relevance feedback based on image similarity

Page 72: Lecture 16: IR Components 2

2007.03.15 - SLIDE 72IS 240 – Spring 2007

BlobWorld (Carson, et al.)

Page 73: Lecture 16: IR Components 2

2007.03.15 - SLIDE 73

Time Series R.F. (Keogh & Pazzani 98)

Page 74: Lecture 16: IR Components 2

2007.03.15 - SLIDE 74IS 240 – Spring 2007

Classifying R.F. Systems

• Standard Relevance Feedback– Individual, explicit, dynamic, item

comparison

• Standard Filtering (NewsWeeder)– Individual, explicit, standing profile, item

comparison

• Standard Routing– “Community” (gold standard), explicit,

standing profile, item comparison

Page 75: Lecture 16: IR Components 2

2007.03.15 - SLIDE 75IS 240 – Spring 2007

Classifying R.F. Systems

• Letizia and WebWatcher– Individual, implicit, dynamic, item comparison

• Ringo and GroupLens:– Group, explicit, standing query, judge-based

comparison

Page 76: Lecture 16: IR Components 2

2007.03.15 - SLIDE 76IS 240 – Spring 2007

Classifying R.F. Systems

• Syskill & Webert:– Individual, explicit, dynamic + standing, item

comparison

• Alexa: (?)– Community, implicit, standing, item

comparison, similar items

• Amazon (?):– Community, implicit, standing, judges + items,

similar items

Page 77: Lecture 16: IR Components 2

2007.03.15 - SLIDE 77IS 240 – Spring 2007

Summary

• Relevance feedback is an effective means for user-directed query modification.

• Modification can be done with either direct or indirect user input

• Modification can be done based on an individual’s or a group’s past input.


Recommended