Personalizing Information Search: Understanding Users and their Interests Diane Kelly School of Information & Library Science University of North Carolina.

Slides:



Advertisements
Similar presentations
Haystack: Per-User Information Environment 1999 Conference on Information and Knowledge Management Eytan Adar et al Presented by Xiao Hu CS491CXZ.
Advertisements

1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Developing and Evaluating a Query Recommendation Feature to Assist Users with Online Information Seeking & Retrieval With graduate students: Karl Gyllstrom,
1 Retrieval Performance Evaluation Modern Information Retrieval by R. Baeza-Yates and B. Ribeiro-Neto Addison-Wesley, (Chapter 3)
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Information Retrieval Visualization CPSC 533c Class Presentation Qixing Zheng March 22, 2004.
Evaluating Search Engine
Search Engines and Information Retrieval
WebMiningResearch ASurvey Web Mining Research: A Survey Raymond Kosala and Hendrik Blockeel ACM SIGKDD, July 2000 Presented by Shan Huang, 4/24/2007.
Information Retrieval in Practice
INFO 624 Week 3 Retrieval System Evaluation
© Tefko Saracevic, Rutgers University 1 EVALUATION in searching IR systems Digital libraries Reference sources Web sources.
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
WebMiningResearchASurvey Web Mining Research: A Survey Raymond Kosala and Hendrik Blockeel ACM SIGKDD, July 2000 Presented by Shan Huang, 4/24/2007 Revised.
Web Logs and Question Answering Richard Sutcliffe 1, Udo Kruschwitz 2, Thomas Mandl University of Limerick, Ireland 2 - University of Essex, UK 3.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Personalizing the Digital Library Experience Nicholas J. Belkin, Jacek Gwizdka, Xiangmin Zhang SCILS, Rutgers University
WXGB6106 INFORMATION RETRIEVAL Week 3 RETRIEVAL EVALUATION.
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
Personalization of the Digital Library Experience: Progress and Prospects Nicholas J. Belkin Rutgers University, USA
Search Engines and Information Retrieval Chapter 1.
1 Information Filtering & Recommender Systems (Lecture for CS410 Text Info Systems) ChengXiang Zhai Department of Computer Science University of Illinois,
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
Gradual Adaption Model for Estimation of User Information Access Behavior J. Chen, R.Y. Shtykh and Q. Jin Graduate School of Human Sciences, Waseda University,
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
User Models for Personalization Josh Alspector Chief Technology Officer.
Query Expansion By: Sean McGettrick. What is Query Expansion? Query Expansion is the term given when a search engine adding search terms to a user’s weighted.
WebMining Web Mining By- Pawan Singh Piyush Arora Pooja Mansharamani Pramod Singh Praveen Kumar 1.
Heuristic evaluation Functionality: Visual Design: Efficiency:
Mining the Web to Create Minority Language Corpora Rayid Ghani Accenture Technology Labs - Research Rosie Jones Carnegie Mellon University Dunja Mladenic.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
Presented By :Ayesha Khan. Content Introduction Everyday Examples of Collaborative Filtering Traditional Collaborative Filtering Socially Collaborative.
Effects of Popularity and Quality on the Usage of Query Suggestions during Information Search Can users be induced to take bad query suggestions because.
Personalized Search Xiao Liu
Toward A Session-Based Search Engine Smitha Sriram, Xuehua Shen, ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Collaborative Information Retrieval - Collaborative Filtering systems - Recommender systems - Information Filtering Why do we need CIR? - IR system augmentation.
 Examine two basic sources for implicit relevance feedback on the segment level for search personalization. Eye tracking Display time.
Personalized Course Navigation Based on Grey Relational Analysis Han-Ming Lee, Chi-Chun Huang, Tzu- Ting Kao (Dept. of Computer Science and Information.
Personalized Interaction With Semantic Information Portals Eric Schwarzkopf DFKI
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Information Retrieval in Context of Digital Libraries - or DL in Context of IR Peter Ingwersen Royal School of LIS Denmark –
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Information Retrieval
Xinyu Xing, Wei Meng, Dan Doozan, Georgia Institute of Technology Alex C. Snoeren, UC San Diego Nick Feamster, and Wenke Lee, Georgia Institute of Technology.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
User Modeling and Recommender Systems: Introduction to recommender systems Adolfo Ruiz Calleja 06/09/2014.
Topic by Topic Performance of Information Retrieval Systems Walter Liggett National Institute of Standards and Technology TREC-7 (1999)
WIRED Future Quick review of Everything What I do when searching, seeking and retrieving Questions? Projects and Courses in the Fall Course Evaluation.
AQUAINT AQUAINT Evaluation Overview Ellen M. Voorhees.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Evaluation of Information Retrieval Systems Xiangming Mu.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
KMS & Collaborative Filtering Why CF in KMS? CF is the first type of application to leverage tacit knowledge People-centric view of data Preferences matter.
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Text Information Management ChengXiang Zhai, Tao Tao, Xuehua Shen, Hui Fang, Azadeh Shakery, Jing Jiang.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
UNIVERSITY UTARA MALAYSIA COLLEGE OF ARTS & SCIENCES.
Information Retrieval in Practice
Evaluation Anisio Lacerda.
Information Retrieval (in Practice)
Presentation transcript:

Personalizing Information Search: Understanding Users and their Interests Diane Kelly School of Information & Library Science University of North Carolina IPAM | 04 October 2007

 What is IR?  Who works on problems in IR?  Where can I find the most recent work in IR? Where can I find the most recent work in IR?  A TREC primer A TREC primer Background: IR and TREC

 Personalization is a process where retrieval is customized to the individual (not one-size-fits-all searching)  Hans Peter Luhn was one of the first people to personalize IR through selective dissemination of information (SDI) (now called ‘filtering’)  Profiles and user models are often employed to ‘house’ data about users and represent their interests  Figuring out how to populate and maintain the profile or user model is a hard problem Background: Personalization

 Explicit Feedback  Implicit Feedback  User’s desktop Major Approaches

Explicit Feedback

 Term relevance feedback is one of the most widely used and studied explicit feedback techniques  Typical relevance feedback scenarios (examples)examples  Systems-centered research has found that relevance feedback works (including pseudo- relevance feedback)  User-centered research has found mixed results about its effectiveness Explicit Feedback

 Terms are not presented in context so it may be hard for users to understand how they can help  Quality of terms suggested is not always good  Users don’t have the additional cognitive resources to engage in explicit feedback  Users are too lazy to provide feedback  Questions about the sustainability of explicit feedback for long-term modeling Explicit Feedback

Examples

BACK

Query Elicitation Study  Users typically pose very short queries  This may be because  users have a difficult time articulating their information needs  traditional search interfaces encourage short queries  Polyrepresentative extraction of information needs suggests obtaining multiple representations of a single information need (reference interview)

Motivation  Research has demonstrated that a positive relationship exists between query length and performance in batch-mode experimental IR  Query expansion is an effective technique for increasing query length, but research has demonstrated that users have some difficulty with traditional term relevance feedback features

Elicitation Form [Why Know] [Already Know] [Keywords]

Results: Number of Terms Already Know Why Keywords N=

Experimental Runs Source of TermsRun ID Baselinebaseline Baseline + Pseudo Relevance Feedback pseudo05, pseudo10, pseudo20, pseudo50 Baseline + Elicitation Form Q2Q2 Baseline + Elicitation Form Q3Q3 Baseline + Elicitation Form Q4Q4 Baseline + Combination of Elicitation Form Questions Q3Q4, Q2Q3, Q2Q4, Q234

Overall Performance

Query Length and Performance y = (x), p=.000

Major Findings  Users provided lengthy responses to some of the questions  There were large differences in the length of users’ responses to each question  In most cases responses significantly improved retrieval  Query length and performance were significantly related

Implicit Feedback

 What is it? Information about users, their needs and document preferences that can be obtained unobtrusively, by watching users’ interactions and behaviors with systems  What are some examples?  Examine: Select, View, Listen, Scroll, Find, Query, Cumulative measures  Retain: Print, Save, Bookmark, Purchase,  Reference: Link, Cite  Annotate/Create: Mark up, Type, Edit, Organize, Label Implicit Feedback

 Why is it important?  It is generally believed that users are unwilling to engage in explicit relevance feedback  It is unlikely that users can maintain their profiles over time  Users generate large amounts of data each time the engage in online information-seeking activities and the things in which they are ‘interested’ is in this data somewhere Implicit Feedback

 What do we “know” about it?  There seems to be a positive correlation between selection (click-through) and relevance  There seems to be a positive correlation between display time and relevance  What is problematic about it?  Much of the research has been based on incomplete data and general behavior  And has not considered the impact of contextual variables – such as task and a user’s familiarity with a topic – on behaviors Implicit Feedback

Implicit Feedback Study  To investigate:  the relationship between behaviors and relevance  the relationship between behaviors and context  To develop a method for studying and measuring behaviors, context and relevance in a natural setting, over time

Method  Approach: naturalistic and longitudinal, but some control  Subjects/Cases: 7 Ph.D. students  Study period: 14 weeks  Compensation: new laptops and printers

Data Collection Document Context Tasks Topics Persistence Familiarity Endurance Frequency Stage Behaviors Display Time PrintingSaving Relevance Usefulness

Protocol STARTEND 14 weeks Context Evaluation Document Evaluations Context Evaluation; Document Evaluations Client- & Server-side Logging Week 1Week 13

Results: Description of Data Subject Client2.6 MB6.8 MB3.9 MB2.0 MB1.5 MB21.7 MB4.9 MB Proxy1.7 GB83 MB39 MB42 MB48 MB2.9 GB2.1 GB URLs Requested 15,4995,3193,1573,2053,40414,58611,657 Docs Evaluated 870 (5%) 802 (14%) 384 (12%) 353 (11%) 200 (6%) 1,328 (8%) 1,160 (10%) Tasks Topics

Relevance: Usefulness 4.8 (1.65) 6.1 (2.00) 5.3 (2.20) 6.0 (0.80) 5.3 (2.40) 4.6 (0.80) 5.0 (2.40)

Relevance: Usefulness

Display Time

Display Time & Usefulness

Display Time & Task Tasks 1. Researching Dissertation 2. Shopping 3. Read News 4. Movie Reviews & Schedules 5. Preparing Course 6. Entertainment

Major Findings  Behaviors differed for each subject, but in general  most display times were low  most usefulness ratings were high  not much printing or saving  No direct relationship between display time and usefulness

Major Findings  Main effects for display time and all contextual variables:  Task (5 subjects)  Topic (6 subjects)  Familiarity (5 subjects)  Lower levels of familiarity associated with higher display times  No clear interaction effects among behaviors, context and relevance

Personalizing Search  Using the display time, task and relevance information from the study, we evaluated the effectiveness of a set of personalized retrieval algorithms  Four algorithms for using display time as implicit feedback were tested: 1.User 2.Task 3.User + Task 4.General

Results Iteration MAP

Major Findings  Tailoring display time thresholds based on task information improved performance, but doing so based on user information did not  There was a lot of variability between subjects, with the user-centered algorithms performing well for some and poorly for others  The effectiveness of most of the algorithms increased with time (and more data)

Some Problems

Relevance  What are we modeling? Does click = relevance?  Relevance is multi-dimensional and dynamic  A single measure does to adequately reflect ‘relevance’  Most pages are likely to be rated as useful, even if the value or importance of the information differs

Definition Recipe

Weather Forecast Information about Rocky Mountain Spotted Fever

Paper about Personalization

Page Structure  Some behaviors are more likely to occur on some types of pages  A more ‘intelligent’ modeling function would know when and what to observe and expect  The structure of pages encourage/inhibit certain behaviors  Not all pages are equally as useful for modeling a user’s interests

What types of behaviors do you expect here? And here?

The Future

Future  New interaction styles and systems create new opportunities for explicit and implicit feedback  Collaborative search features and query recommendation  Features/Systems that support the entire search process (e.g., saving, organizing, etc.)  QA systems  New types of feedback  Negative  Physiological

Diane Kelly WEB : Collaborators: Nick Belkin, Xin Fu, Vijay Dollu, Ryen White Thank You

TREC [Text REtrieval Conference] It’s not this …

What is TREC?  TREC is a workshop series sponsored by the National Institute of Standards and Technology (NIST) and the US Department of Defense.  It’s purpose is to build infrastructure for large-scale evaluation of text retrieval technology.  TREC collections and evaluation measures are the de facto standard for evaluation in IR.  TREC is comprised of different tracks each of which focuses on different issues (e.g., question answering, filtering).

TREC Collections  Central to each TREC Track is a collection, which consists of three major components: 1.A corpus of documents (typically newswire) 2.A set of information needs (called topics)topics 3.A set of relevance judgments.  Each Track also adopts particular evaluation measures  Precision and Recall; F-measure  Average Precision (AP) and Mean AP (MAP)

Comparison of Measures List 1List 2 1R1/1 = 11NR 2R2/2 = 22NR 3R3/3 = 33NR 4R4/4 = 44NR 5R5/5 = 55NR 6 6R1/6 =.167 7NR7R2/7 =.286 8NR8R3/8 =.375 9NR9R4/9 = NR10R5/10 =.50 AP1.0AP.354

Learn more about TREC   Voorhees, E. M., & Harman, D. K. (2005). TREC: Experiment and Evaluation in Information Retrieval, Cambridge, MA: MIT Press. BACK

Example Topic BACK

Learn more about IR  ACM SIGIR Conference  Sparck-Jones, K., & Willett, P. (1997). Readings in Information Retrieval. Morgan-Kaufman Publishers.  Baeza-Yates, R., & Ribeiro-Neto, B. (1999). Modern information retrieval. New York, NY: ACM Press.  Grossman, D. A., Frieder, O. (2004). Information retrieval: Algorithms and Heuristics. The Netherlands: Springer. BACK