Presentation is loading. Please wait.

Presentation is loading. Please wait.

SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000.

Similar presentations


Presentation on theme: "SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000."— Presentation transcript:

1 SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000

2 Today l Relevance Feedback (RF) –Review the Rocchio formula –Evaluation of an RF interface –(this is an intro to user interfaces in IR) –The Guides System

3 Information need Index Pre-process Parse Collections Rank Query text input

4 Information need Index Pre-process Parse Collections Rank Query text input Query Modification

5 l Problem: how to reformulate the query? –Thesaurus expansion: »Suggest terms similar to query terms –Relevance feedback: »Suggest terms (and documents) similar to retrieved documents that have been judged to be relevant

6 Relevance Feedback Also known as: –query modification – “more like this”

7 Relevance Feedback l Main Idea: –Modify existing query based on relevance judgements »Extract terms from relevant documents and add them to the query »and/or re-weight the terms already in the query –Two main approaches: »Automatic (psuedo-relevance feedback) »Users select relevant documents –Users/system select terms from an automatically-generated list

8 Relevance Feedback l Usually do both: –expand query with new terms –re-weight terms in query l There are many variations –usually positive weights for terms from relevant docs –sometimes negative weights for terms from non-relevant docs

9 Rocchio Method

10 l Rocchio automatically –re-weights terms –adds in new terms (from relevant docs) »have to be careful when using negative terms »Rocchio is not a machine learning algorithm l Most methods perform similarly –results heavily dependent on test collection l Machine learning methods are proving to work better than standard IR approaches like Rocchio

11 Using Relevance Feedback l Known to improve results –in TREC-like conditions (no user involved) l What about with a user in the loop? –How might you measure this? –Let’s examine a user study of relevance feedback by Koenneman & Belkin 1996.

12 Questions being Investigated Koenemann & Belkin 96 l How well do users work with statistical ranking on full text? l Does relevance feedback improve results? l Is user control over operation of relevance feedback helpful? l How do different levels of user control effect results?

13 How much of the guts should the user see? l Opaque (black box) –(like web search engines) l Transparent –(see available terms after the r.f. ) l Penetrable –(see suggested terms before the r.f.) l Which do you think worked best?

14

15 Terms available for relevance feedback made visible (from Koenemann & Belkin)

16 Details on User Study Koenemann & Belkin 96 l Subjects have a tutorial session to learn the system l Their goal is to keep modifying the query until they’ve developed one that gets high precision l This is an example of a routing query (as opposed to ad hoc) l Reweighting: –They did not reweight query terms –Instead, only term expansion »pool all terms in rel docs »take top N terms, where »n = 3 + (number-marked-relevant-docs*2) »(the more marked docs, the more terms added to the query)

17 Details on User Study Koenemann & Belkin 96 l 64 novice searchers –43 female, 21 male, native English l TREC test bed –Wall Street Journal subset l Two search topics –Automobile Recalls –Tobacco Advertising and the Young l Relevance judgements from TREC and experimenter l System was INQUERY (vector space with some bells and whistles)

18 Sample TREC query

19 Evaluation l Precision at 30 documents l Baseline: (Trial 1) –How well does initial search go? –One topic has more relevant docs than the other l Experimental condition (Trial 2) –Subjects get tutorial on relevance feedback –Modify query in one of four modes »no r.f., opaque, transparent, penetration

20 Precision vs. RF condition (from Koenemann & Belkin 96)

21 Effectiveness Results l Subjects with R.F. did 17-34% better performance than no R.F. l Subjects with penetration case did 15% better as a group than those in opaque and transparent cases.

22 Number of iterations in formulating queries ( from Koenemann & Belkin 96)

23 Number of terms in created queries (from Koenemann & Belkin 96)

24 Behavior Results l Search times approximately equal l Precision increased in first few iterations l Penetration case required fewer iterations to make a good query than transparent and opaque l R.F. queries much longer –but fewer terms in penetrable case -- users were more selective about which terms were added in.

25 The Guides System l An innovative search interface –Character-oriented relevance feedback –Note use of the vector space model »Each character gets a vector »User’s activity defines the user’s vector »The character whose vector is closest at any time becomes the active one –(Show Video) –Developed at Apple in the 80’s by Abbe Don, Tim Oren, and Brenda Laurel –http://www.abbedon.com/Project/guides.html

26 Relevance Feedback Summary l Iterative query modification can improve precision and recall for a standing query l In at least one study, users were able to make good choices by seeing which terms were suggested for R.F. and selecting among them l So … “more like this” can be useful!


Download ppt "SIMS 202 Information Organization and Retrieval Prof. Marti Hearst and Prof. Ray Larson UC Berkeley SIMS Tues/Thurs 9:30-11:00am Fall 2000."

Similar presentations


Ads by Google