Download presentation
Presentation is loading. Please wait.
Published byKatherine Jacobs Modified over 9 years ago
1
Retrieval and Evaluation Techniques for Personal Information Jin Young Kim 7/26 Ph.D Dissertation Seminar
2
Personal Information Retrieval (PIR) 2 The practice and the study of supporting users to retrieve personal information effectively
3
Personal Information Retrieval in the Wild Everyone has unique information & practices Different information and information needs Different preference and behavior Many existing software solutions Platform-level: desktop search, folder structure Application-level: email, calendar, office suites 3
4
Previous Work in PIR (Desktop Search) Focus User interface issues [Dumais03,06] Desktop-specific features [Solus06] [Cohen08] Limitations Each based on different environment and user group None of them performed comparative evaluation Research findings do not accumulate over the years 4
5
Our Approach Develop general techniques for PIR Start from essential characteristics of PIR Applicable regardless of users and information types Make contributions to related areas Structured document retrieval Simulated evaluation for known-item finding Build a platform for sustainable progress Develop repeatable evaluation techniques Share the research findings and the data 5
6
Essential Characteristics of PIR Many document types Unique metadata for each type People combine search and browsing [Teevan04] Long-term interactions with a single user People mostly find known-items [Elsweiler07] Privacy concern for the data set 6 Field-based Search Models Associative Browsing Model Simulated Evaluation Methods
7
Challenge Users may remember different things about the document How can we present effective results for both cases? Search and Browsing Retrieval Models Registration James User’s Memory Query Retrieval Results Search Browsing Lexical Memory Associative Memory 1. 2. 3. 4. 5. 7
8
Information Seeking Scenario in PIR Registration James Registration 2011 User Input System Output 2011 Search Browsing Search A user initiate a session with a keyword query The user switches to browsing by clicking on a email document The user switches to back to search with a different query
9
Challenge User’s query originates from what she remembers. How can we simulate user’s querying behavior realistically? Simulated Evaluation Techniques Registration James User’s Memory Query Retrieval Results Lexical Memory Associative Memory 1. 2. 3. 4. 5. Search Browsing 9
10
Research Questions Field-based Search Models How can we improve the retrieval effectiveness in PIR? How can we improve the type prediction quality? Associative Browsing Model How can we enable the browsing support for PIR? How can we improve the suggestions for browsing? Simulated Evaluation Methods How can we evaluate a complex PIR system by simulation? How can we establish the validity of simulated evaluation? 10
11
Field-based Search Models
12
Searching for Personal Information An example of desktop search 12
13
Field-based Search Framework for PIR Type-specific Ranking Rank documents in each document collection (type) Type Prediction Predict the document type relevant to user’s query Final Results Generation Merge into a single ranked list 13
14
Type-specific Ranking for PIR Individual collection has type-specific features Thread-based features for emails Path-based features for documents Most of these documents have rich metadata Email: Document: Calendar: We focus on developing general retrieval techniques for structured documents
15
Structured Document Retrieval Field Operator / Advanced Search Interface User’s search terms are found in multiple fields 15 Understanding Re-finding Behavior in Naturalistic Email Interaction Logs. Elsweiler, D, Harvey, M, Hacker., M [SIGIR'11]
16
Structured Document Retrieval: Models Document-based Retrieval Model Score each document as a whole Field-based Retrieval Model Combine evidences from each field q 1 q 2... q m Document-based Scoring Field-based Scoring f1f1 f1f1 f2f2 f2f2 fnfn fnfn... q 1 q 2... q m f1f1 f1f1 f2f2 f2f2 fnfn fnfn... f1f1 f1f1 f2f2 f2f2 fnfn fnfn w1w1 w2w2 wnwn w1w1 w2w2 wnwn 16
17
17 1 1 2 2 1 2 Field Relevance Different fields are important for different query terms ‘james’ is relevant when it occurs in ‘registration’ is relevant when it occurs in Field Relevance Model for Structured IR
18
Estimating the Field Relevance: Overview If User Provides Feedback Relevant document provides sufficient information If No Feedback is Available Combine field-level term statistics from multiple sources 18 content title from/to Relevant Docs content title from/to Collection content title from/to Top-k Docs + ≅
19
Assume a user who marked D R as relevant Estimate field relevance from the field-level term dist. of D R We can personalize the results accordingly Rank higher docs with similar field-level term distribution This weight is provably optimal under LM retrieval framework Estimating Field Relevance using Feedback 19 DRDR - To is relevant for ‘james’ - Content is relevant for ‘registration’ Field Relevance:
20
Linear Combination of Multiple Sources Weights estimated using training queries Features Field-level term distribution of the collection Unigram and Bigram LM Field-level term distribution of top-k docs Unigram and Bigram LM A priori importance of each field (w j ) Estimated using held-out training queries Estimating Field Relevance without Feedback 20 Unigram is the same to PRM-S Similar to MFLM and BM25F Pseudo-relevance Feedback
21
21 Retrieval Using the Field Relevance Comparison with Previous Work Ranking in the Field Relevance Model q 1 q 2... q m f1f1 f1f1 f2f2 f2f2 fnfn fnfn... f1f1 f1f1 f2f2 f2f2 fnfn fnfn w1w1 w2w2 wnwn w1w1 w2w2 wnwn q 1 q 2... q m f1f1 f1f1 f2f2 f2f2 fnfn fnfn... f1f1 f1f1 f2f2 f2f2 fnfn fnfn P(F 1 |q 1 ) P(F 2 |q 1 ) P(F n |q 1 ) P(F 1 |q m ) P(F 2 |q m ) P(F n |q m ) Per-term Field Weight Per-term Field Score sum multiply
22
Retrieval Effectiveness (Metric: Mean Reciprocal Rank) Evaluating the Field Relevance Model 22 Fixed Field Weights Per-term Field Weights
23
Type Prediction Methods Field-based collection Query-Likelihood (FQL) Calculate QL score for each field of a collection Combine field-level scores into a collection score Feature-based Method Combine existing type-prediction methods Grid Search / SVM for finding combination weights 23
24
Type Prediction Performance Pseudo-desktop Collections CS Collection FQL improves performance over CQL Combining features improves the performance further 24 (% of queries with correct prediction)
25
Summary So Far… Field relevance model for structured document retrieval Enables relevance feedback through field weighting Improves performance using linear feature-based estimation Type prediction methods for PIR Field-based type prediction method (FQL) Combination of features improve the performance further We move onto associative browsing model What happens when users can’t recall good search terms?
26
Associative Browsing Model
27
Recap: Retrieval Framework for PIR Registration James Keyword SearchAssociative Browsing 27
28
User Interaction for Associative Browsing Users enter a concept or document page by search The system provides a list of suggestions for browsing Data ModelUser Interface
29
How can we build associations? 29 Manually? Participants wouldn’t create associations beyon d simple tagging operations - Sauermann et al. 2005 Participants wouldn’t create associations beyon d simple tagging operations - Sauermann et al. 2005 Automatically? How would it match user’s preference?
30
Building the Associative Browsing Model 30 2. Concept Extraction 3. Link Extraction 4. Link Refinement 1. Document Collection Term Similarity Temporal Similarity Co-occurrence Click-based Training
31
Concept: Search Engine Link Extraction and Refinement 31 Link Scoring Combination of link type scores S(c 1,c 2 ) = Σ i [ w i × Link i (c 1,c 2 ) ] Link Presentation Ranked list of suggested items Users click on them for browsing Link Refinement (training w i ) Maximize click-based relevance Grid Search : Maximize retrieval effectiveness (MRR) RankSVM : Minimize error in pairwise preference ConceptsDocuments Term Vector Similarity Temporal Similarity Tag Similarity String SimilarityPath / Type Similarity Co-occurrenceConcept Similarity
32
Evaluating Associative Browsing Model Data set: CS Collection Collect public documents in UMass CS department CS dept. people competed in known-item finding tasks Value of browsing for known-item finding % of sessions browsing was used % of sessions browsing was used & led to success Quality of browsing suggestions Mean Reciprocal Rank using clicks as judgments 10-fold cross validation over the click data collected 32
33
Value of Browsing for Known-item Finding Comparison with Simulation Results Roughly matches in terms of overall usage and success ratio The Value of Associative Browsing Browsing was used in 30% of all sessions Browsing saved 75% of sessions when used Evaluation TypeTotal (#sessions) Browsing usedSuccessful outcome Simulation63,2609,410 (14.8%)3,957 (42.0%) User Study (1)29042 (14.5%)15 (35.7%) User Study (2)14243 (30.2%)32 (74.4%) Document Only Document + Concept
34
Quality of Browsing Suggestions Concept Browsing (MRR) Document Browsing (MRR) 34
35
Simulated Evaluation Methods
36
Challenges in PIR Evaluation Hard to create a ‘test-collection’ Each user has different documents and habits People will not donate their documents and queries for research Limitations of user study Experimenting with a working system is costly Experimental control is hard with real users and tasks Data is not reusable by third parties 36
37
Simulate components of evaluation Collection: user’s documents with metadata Task: search topics and relevance judgments Interaction: query and click data Our Approach: Simulated Evaluation 37
38
Simulated Evaluation Overview Simulated document collections Pseudo-desktop Collections Subsets of W3C mailing list + Other document types CS Collection UMass CS mailing list / Calendar items / Crawl of homepage Evaluation Methods Controlled User StudySimulated Interaction Field-based Search DocTrack Search GameQuery Generation Methods Associative Browsing DocTrack Search + Browsing Game Probabilistic User Modeling
39
Controlled User Study: DocTrack Game Procedure Collect public documents in UMass CS dept. (CS Collection) Build a web interface where participants can find documents People in CS department participated DocTrack search game 20 participants / 66 games played 984 queries collected for 882 target documents DocTrack search+browsing game 30 participants / 53 games played 290 +142 search sessions collected 39
40
DocTrack Game 40 *Users can use search and browsing for DocTrack search+browsing game
41
Query Generation for Evaluating PIR Known-item finding for PIR A target document represents an information need Users would take terms from the target document Query Generation for PIR Randomly select a target document Algorithmically take terms from the document Parameters of Query Generation Choice of extent : Document [Azzopardi07] vs. Field Choice of term : Uniform vs. TF vs. IDF vs. TF-IDF [Azzopardi07] 41
42
Validating of Generated Queries Basic Idea Use the set of human-generated queries for validation Compare at the level of query terms and retrieval scores Validation by Comparing Query-terms The generation probability of manual query q from P term Validation by Compare Retrieval Scores [Azzopardi07] Two-sided Kolmogorov-Smirnov test 42
43
Validation Results for Generated Queries Validation based on query terms Validation based on retrieval score distribution
44
Probabilistic User Model for PIR Query generation model Term selection from a target document State transition model Use browsing when result looks marginally relevant Link selection model Click on browsing suggestions based on perceived relevance 44
45
A User Model for Link Selection User’s level of knowledge Random : randomly click on a ranked list Informed : more likely to click on more relevant item Oracle : always click on the most relevant item Relevance estimated using the position of the target item 45
46
Success Ratio of Browsing Varying the level of knowledge and fan-out for simulation Exploration is valuable for users with low knowledge level More Exploration 46
47
Community Efforts using the Data Sets 47
48
Conclusions & Future Work
49
Major Contributions Field-based Search Models Field relevance model for structured document retrieval Field-based and combination-based type prediction method Associative Browsing Model An adaptive technique for generating browsing suggestions Evaluation of associative browsing in known-item finding Simulated Evaluation Methods for Known-item Finding DocTrack game for controlled user study Probabilistic user model for generating simulated interaction 49
50
Field Relevance for Complex Structures Current work assumes documents with flat structure Field Relevance for Complex Structures? XML documents with hierarchical structure Joined Database Relations with graph structure
51
Cognitive Model of Query Generation Current query generation methods assume: Queries are generated from the complete document Query-terms are chosen independently from one another Relaxing these assumptions Model the user’s degradation in memory Model the dependency in query term selection Ongoing work Graph-based representation of documents Query terms can be chosen by random walk
52
Thank you for your attention! Special thanks to my advisor, coauthors, and all of you here! Are we closer to the superhuman now?
53
One More Slide: What I Learned… Start from what’s happening from user’s mind Field relevance / query generation, … Balance user input and algorithmic support Generating suggestions for associative browsing Learn from your peers & make contributions Query generation method / DocTrack game Simulated test collections & workshop
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.