Integrating Multiple Resources for Diversified Query Expansion Arbi Bouchoucha, Xiaohua Liu, and Jian-Yun Nie Dept. of Computer Science and Operations.

Slides:



Advertisements
Similar presentations
Term Level Search Result Diversification DATE : 2013/09/11 SOURCE : SIGIR’13 AUTHORS : VAN DANG, W. BRUCE CROFT ADVISOR : DR.JIA-LING, KOH SPEAKER : SHUN-CHEN,
Advertisements

DQR : A Probabilistic Approach to Diversified Query recommendation Date: 2013/05/20 Author: Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Source:
Query Chain Focused Summarization Tal Baumel, Rafi Cohen, Michael Elhadad Jan 2014.
Chapter 5: Introduction to Information Retrieval
Diversified Retrieval as Structured Prediction Redundancy, Diversity, and Interdependent Document Relevance (IDR ’09) SIGIR 2009 Workshop Yisong Yue Cornell.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
A Machine Learning Approach for Improved BM25 Retrieval
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
DOMAIN DEPENDENT QUERY REFORMULATION FOR WEB SEARCH Date : 2013/06/17 Author : Van Dang, Giridhar Kumaran, Adam Troy Source : CIKM’12 Advisor : Dr. Jia-Ling.
WSCD INTRODUCTION  Query suggestion has often been described as the process of making a user query resemble more closely the documents it is expected.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
Evaluating Search Engine
Query Operations: Automatic Local Analysis. Introduction Difficulty of formulating user queries –Insufficient knowledge of the collection –Insufficient.
Context-Aware Query Classification Huanhuan Cao 1, Derek Hao Hu 2, Dou Shen 3, Daxin Jiang 4, Jian-Tao Sun 4, Enhong Chen 1 and Qiang Yang 2 1 University.
Semantic text features from small world graphs Jure Leskovec, IJS + CMU John Shawe-Taylor, Southampton.
Queensland University of Technology An Ontology-based Mining Approach for User Search Intent Discovery Yan Shen, Yuefeng Li, Yue Xu, Renato Iannella, Abdulmohsen.
Affinity Rank Yi Liu, Benyu Zhang, Zheng Chen MSRA.
An investigation of query expansion terms Gheorghe Muresan Rutgers University, School of Communication, Information and Library Science 4 Huntington St.,
Chapter 5: Information Retrieval and Web Search
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 30, (2014) BERLIN CHEN, YI-WEN CHEN, KUAN-YU CHEN, HSIN-MIN WANG2 AND KUEN-TYNG YU Department of Computer.
Longbiao Kang, Baotian Hu, Xiangping Wu, Qingcai Chen, and Yan He Intelligent Computing Research Center, School of Computer Science and Technology, Harbin.
Modern Retrieval Evaluations Hongning Wang
An Integrated Approach to Extracting Ontological Structures from Folksonomies Huairen Lin, Joseph Davis, Ying Zhou ESWC 2009 Hyewon Lim October 9 th, 2009.
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Exploiting Wikipedia as External Knowledge for Document Clustering Sakyasingha Dasgupta, Pradeep Ghosh Data Mining and Exploration-Presentation School.
Leveraging Conceptual Lexicon : Query Disambiguation using Proximity Information for Patent Retrieval Date : 2013/10/30 Author : Parvaz Mahdabi, Shima.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
Focused Matrix Factorization for Audience Selection in Display Advertising BHARGAV KANAGAL, AMR AHMED, SANDEEP PANDEY, VANJA JOSIFOVSKI, LLUIS GARCIA-PUEYO,
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
Linking Wikipedia to the Web Antonio Flores Bernal Department of Computer Sciencies San Pablo Catholic University 2010.
A Simple Unsupervised Query Categorizer for Web Search Engines Prashant Ullegaddi and Vasudeva Varma Search and Information Extraction Lab Language Technologies.
Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Speaker: Ruirui Li 1 The University of Hong Kong.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Estimating Topical Context by Diverging from External Resources SIGIR’13, July 28–August 1, 2013, Dublin, Ireland. Presenter: SHIH, KAI WUN Romain Deveaud.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Context-Sensitive Information Retrieval Using Implicit Feedback Xuehua Shen : department of Computer Science University of Illinois at Urbana-Champaign.
Chapter 6: Information Retrieval and Web Search
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Improving Web Search Results Using Affinity Graph Benyu Zhang, Hua Li, Yi Liu, Lei Ji, Wensi Xi, Weiguo Fan, Zheng Chen, Wei-Ying Ma Microsoft Research.
Modeling term relevancies in information retrieval using Graph Laplacian Kernels Shuguang Wang Joint work with Saeed Amizadeh and Milos Hauskrecht.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
Wikipedia as Sense Inventory to Improve Diversity in Web Search Results Celina SantamariaJulio GonzaloJavier Artiles nlp.uned.es UNED,c/Juan del Rosal,
1 Opinion Retrieval from Blogs Wei Zhang, Clement Yu, and Weiyi Meng (2007 CIKM)
1 Using The Past To Score The Present: Extending Term Weighting Models with Revision History Analysis CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG,
Semantic v.s. Positions: Utilizing Balanced Proximity in Language Model Smoothing for Information Retrieval Rui Yan†, ♮, Han Jiang†, ♮, Mirella Lapata‡,
Probabilistic Latent Query Analysis for Combining Multiple Retrieval Sources Rong Yan Alexander G. Hauptmann School of Computer Science Carnegie Mellon.
Mining Dependency Relations for Query Expansion in Passage Retrieval Renxu Sun, Chai-Huat Ong, Tat-Seng Chua National University of Singapore SIGIR2006.
1 Evaluating High Accuracy Retrieval Techniques Chirag Shah,W. Bruce Croft Center for Intelligent Information Retrieval Department of Computer Science.
Post-Ranking query suggestion by diversifying search Chao Wang.
More Than Relevance: High Utility Query Recommendation By Mining Users' Search Behaviors Xiaofei Zhu, Jiafeng Guo, Xueqi Cheng, Yanyan Lan Institute of.
Date: 2012/11/29 Author: Chen Wang, Keping Bi, Yunhua Hu, Hang Li, Guihong Cao Source: WSDM’12 Advisor: Jia-ling, Koh Speaker: Shun-Chen, Cheng.
1 Adaptive Subjective Triggers for Opinionated Document Retrieval (WSDM 09’) Kazuhiro Seki, Kuniaki Uehara Date: 11/02/09 Speaker: Hsu, Yu-Wen Advisor:
1 What Makes a Query Difficult? David Carmel, Elad YomTov, Adam Darlow, Dan Pelleg IBM Haifa Research Labs SIGIR 2006.
Michael Bendersky, W. Bruce Croft Dept. of Computer Science Univ. of Massachusetts Amherst Amherst, MA SIGIR
A Framework to Predict the Quality of Answers with Non-Textual Features Jiwoon Jeon, W. Bruce Croft(University of Massachusetts-Amherst) Joon Ho Lee (Soongsil.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
Autumn Web Information retrieval (Web IR) Handout #14: Ranking Based on Click Through data Ali Mohammad Zareh Bidoki ECE Department, Yazd University.
Meta-Path-Based Ranking with Pseudo Relevance Feedback on Heterogeneous Graph for Citation Recommendation By: Xiaozhong Liu, Yingying Yu, Chun Guo, Yizhou.
INFORMATION RETRIEVAL MEASUREMENT OF RELEVANCE EFFECTIVENESS 1Adrienn Skrop.
Information Retrieval and Extraction 2009 Term Project – Modern Web Search Advisor: 陳信希 TA: 蔡銘峰&許名宏.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
LEARNING IN A PAIRWISE TERM-TERM PROXIMITY FRAMEWORK FOR INFORMATION RETRIEVAL Ronan Cummins, Colm O’Riordan (SIGIR’09) Speaker : Yi-Ling Tai Date : 2010/03/15.
Exploiting Wikipedia as External Knowledge for Document Clustering
and Knowledge Graphs for Query Expansion Saeid Balaneshinkordan
Presentation 王睿.
Relevance and Reinforcement in Interactive Browsing
Presentation transcript:

Integrating Multiple Resources for Diversified Query Expansion Arbi Bouchoucha, Xiaohua Liu, and Jian-Yun Nie Dept. of Computer Science and Operations Research University of Montreal Montreal (Quebec), Canada { E uropean C onference on I nformation R etrieval Presenter: SHIH KAI WUN

Outline Introduction Proposed Framework Experiments Conclusions and Future Work 2

Introduction Queries in Web search are usually short and ambiguous (e.g., ”Java”). To address this issue, various s earch r esult d iversification (SRD) technologies have been proposed. Traditional SRD approaches are based on q uery e xpansion (QE) and p seudo- r elevance f eedback (PRF). One weakness of such approaches is that their performance much depends on the initial retrieval results. D iversified q uery e xpansion (DQE) represents the most recent approach to SRD. One distinguished feature of DQE is the utilization of external resources, e.g., ConceptNet, Wikipedia, or query logs, to generate a set of diversified queries, whose retrieval results are then combined into a new list. 3

Introduction One representative work of DQE is conducted by Bouchoucha et al., which expands queries using ConceptNet and uses the M aximal M arginal R elevance (MMR) strategy to select diversified terms. Their approach outperforms the state-of-the-art existing SRD methods on TREC data. Following this work, we propose to combine multiple resources to diversify queries. Our approach is largely motivated by the following observation: there are a large number of queries for which ConceptNet cannot yield good performance but some other resources can suggest good terms. “defender”, the #20 query from the TREC 2009 Web track. It has six different subtopics. In our experiments, traditional IR models for this query return no relevant documents. Therefore PRF does not work. 4

Introduction ConceptNet returns results covering subtopic 2, 3 and 6, while Wikipedia and query logs provide documents covering subtopic 1, 2, 3, 4 and 1, 2, 4, 5, respectively. By integrating all these sources, we obtain a list of documents covering all the subtopics. We further propose a unified framework to integrate multiple resources for DQE. For a given resource (e.g., ConceptNet), our framework first generates expansion candidates, from which a set of diversified terms are selected following the MMR principle. Then the retrieved documents for any expansion query of any resource are combined and again the MMR principle is used to output diversified results. In this work, we integrate four typical resources, i.e., ConceptNet, Wikipedia, query logs, and initial search results. 5

Introduction It is worth noting that the idea of combining multiple resources has been successfully exploited for other IR tasks. For example, Deveaud et al. combine several external resources to generalize the topical context, and suggest that increasing the number of resources tends to improve the topical representation of the user information need. Bendersky et al. integrate multiple information sources to compute the importance features associated with the explicit query concepts, and to perform PRF. Compared with these studies, our work has two significant differences: 1)These resources are used to directly generate diversified queries. 2)MMR is used to cover as many aspects as possible of the query. 6

Introduction We evaluate our approach using TREC 2009, 2010 and 2011 Web tracks. The experimental results show that multiple resources do complete each other, and integrating multiple resources can often yield substantial improvements compared to using one single resource. Our contributions are twofold: 1)We propose the integration of multiple resources for DQE, and a general framework based on MMR for the implementation. 2)We show the effectiveness of our method on several public datasets. 7

Proposed Framework Our proposed framework consists of two layers. The first layer integrates a set of resources, denoted by R, to generate diversified queries. Given a query Q, it iteratively generates a good expansion term c ∗ for each resource r ∈ R, which is both similar to the initial query Q and dissimilar to the expansion terms already selected: Here, C r,Q and S r,Q represent the set of candidate terms and the set of selected terms for r, respectively; the parameter λ r (in [0,1]) controls the trade-off between relevance and redundancy of the selected term; sim r (c, c i ) returns the similarity score of two terms for resource r; 8

Proposed Framework sim r (c, Q) is the similarity score between term c and the query Q, which is computed using Formula 2, where q is a subset of Q and |q| denotes the number of words of q. Once c ∗ is selected, it is removed from C r,Q and appended to S r,Q. With the parameter λ r, initial term candidates C r,Q, and the term pair similarity function sim r (c, c i ), which depend on the particular resource, Formula 1 becomes a generalized version of M aximal M arginal R elevance- based E xpansion (MMRE) proposed by Bouchoucha et al. And by instantiating λ r, C r,Q and sim r (c, c i ), our framework can integrate any resource. 9

Proposed Framework We investigate four typical resources in this work: ConceptNet, Wikipedia, query logs, and initial search results, hereafter denoted by C, W, QL and D, respectively. For ConceptNet, we use the same approach introduced in [3] to compute sim C. where N ci (resp. N cj ) is the set of nodes from the graph of ConceptNet that are related to the node of the concept ci(resp. cj).The more common node ci and node cj share, the more they are considered to be similar. 10

Proposed Framework For Wikipedia, we define C W,Q as the outlinks, categories, and the set of terms that co-occur with Q or a part of Q; sim W (c, c i ) is defined by Formula 3, where W(W i ) is the set of vectors containing term c(c i ) obtained by ESA, and sim(w, w i ) is simply the cosine similarity of vector w and w i. In cases where no Wikipedia pages match Q or a part of Q, we use E xplicit S emantic A nalysis (ESA) to get semantically related Wikipedia pages, from which to extract the outlinks, categories and representative terms to obtain C W,Q. For query logs, C QL,Q includes the queries that share the same click- through data with Q, as well as the reformulated queries of Q that appear in a user session within a 30 minutes-time window; 11

Proposed Framework sim QL (c, c i ) is defined by Formula 4 : For initial search results, we consider top K returned results as relevant documents, and use PRF to generate C D,Q ; sim D (c, c i ) is computed using Formula 5, where freq(c, c i ) refers to the co-occurrence of term c and c i within a fixed window of size 15. The second layer of our framework generates diversified search results in three steps. First, for each resource r, it generates a set of ranked documents D r,Q using the expansion terms S r,Q, which are then combined into one unique set D Q. 12

Proposed Framework Finally, it uses again the MMR principal to iteratively select d ∗ from the current document candidates. Formula 6 defines this process, where DC Q denotes the document candidates, which is initialized as D Q ; DS Q denotes the set of selected documents, which is empty at the very beginning; λ is the parameter that controls the trade-off between relevance and diversity; rel(d, Q) measures the similarity between document d and query Q; sim(d, d i ) denotes the similarity between two documents. The selected document d ∗ is then removed from DC Q to DS Q. 13

Proposed Framework One core element of the second layer is rel(d, Q), which is defined using Formula 7, where rel(D r,Q, d) and rank(D r,Q, d) are the normalized relevance score and the rank of document d in D r,Q, respectively. This formula captures our intuition that the more a document is ranked on top and with high relevance score, the more relevant it is to the query. 14

Experiments We conduct experiments on the ClueWeb09 dataset, which contains 50,220,423 documents, and use the test queries from TREC 2009, 2010 and 2011 Web tracks. Indri is used as the basic retrieval system. Our baseline is a query generative language model with Dirichlet smoothing ( μ =2000), Krovetz stemmer, and stopword removal using the standard INQUERY stopword list. We consider four typical resources: the last version of ConceptNet, the English Wikipedia dumps of July 8th, 2013, the log data of Microsoft Live Search 2006, which spans over one month consisting of almost 14.9M queries shared between around 5.4M user sessions, and the top 50 results returned for the original query. 15

Experiments The evaluation results in the diversity task of the TREC 2009, 2010 and 2011 Web tracks are reported based on five official measures: MAP and nDCG for adhoc performance, α -nDCG ( α = 0.5) and ERR-IA for diversity measure. We also use S-recall to measure the ratio of covered subtopics for a given query. Using greedy search on each resource, we empirically set λ C = 0.6, λ W = 0.5, λ QL = 0.4, λ D = 0.6, and λ = 0.3. For each test query, we generate 10 expansion terms using MMRE, with respect to each resource. Table 1 reports our evaluation results, from which we make four main observations. Firstly, when each resource is considered separately, using query logs often yields significantly better adhoc retrieval performance and diversity. 16

Experiments 17

Experiments Maybe, this is because the candidate expansion terms generated from query logs are those suggested by users (through their query reformulations), which reflect well the user intent. This suggests the important role of query logs for the diversity task. Secondly, Wikipedia outperforms ConceptNet for TREC 2009 and TREC 2010, but not significantly in general. However, ConceptNet significantly outperforms Wikipedia for TREC 2011 in all the measures. To understand the reason, we manually assess the different queries to see whether they have an exact matching page from Wikipedia. We find that 36/50, 34/48 and 19/50 queries from TREC 2009, TREC 2010 and TREC 2011 respectively, have exact matching pages from Wikipedia,and that only when the query corresponds to a known concept ( i.e. page) from Wikipedia, the candidate expansion terms suggested by Wikipedia tend to be relevant. 18

Experiments This means that Wikipedia helps promoting the diversity of the query results, if the query corresponds to a known concept. Thirdly, the set of feedback documents has the poorest performance among all resources under consideration. Its performance drastically decreases from TREC 2009 to TREC 2010 to TREC 2011 in terms of adhoc retrieval and diversity. This may be due to the fact that the topics of TREC 2011 are harder than the topics of TREC 2010, and the topics of the latter are harder than those of TREC The more the collection contains difficult queries, the more likely the set of top returned documents are irrelevant. Hence, the candidate expansion terms generated from these documents tend to include a lot of noise. 19

Experiments Finally, combing all these resources gives better performance, and in most cases the improvement is significant for almost all the measures. In particular, the diversity scores obtained (for α and are the highest scores. This means that the considered resources are complementary in term of coverage of query subtopics: the subtopics missed by some resources can be recovered by other ones. Moreover, our combination strategy promotes the selection of the most relevant documents in the final results set, which explains why higher scores for MAP and are obtained. 20

Conclusions and Future Work This paper presents a unified framework to integrate multiple resources for DQE. By implementing two functions, one to generate expansion term candidates and the other to compute the similarity of two terms, any resource can be plugged into this framework. Experimental results on TREC 2009, 2010 and 2011 Web tracks show that combining several complementary resources performs better than using one single resource. We have observed that the degree of the contribution of a resource to SRD depends on the query. In future, we are interested in other approaches to resource integration for DQE, e.g., assigning different resources with different weights that are sensitive to the query. 21