DQR : A Probabilistic Approach to Diversified Query recommendation Date: 2013/05/20 Author: Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Source:

Slides:



Advertisements
Similar presentations
Date: 2013/1/17 Author: Yang Liu, Ruihua Song, Yu Chen, Jian-Yun Nie and Ji-Rong Wen Source: SIGIR12 Advisor: Jia-ling Koh Speaker: Chen-Yu Huang Adaptive.
Advertisements

A Framework for Result Diversification
Psychological Advertising: Exploring User Psychology for Click Prediction in Sponsored Search Date: 2014/03/25 Author: Taifeng Wang, Jiang Bian, Shusen.
Term Level Search Result Diversification DATE : 2013/09/11 SOURCE : SIGIR’13 AUTHORS : VAN DANG, W. BRUCE CROFT ADVISOR : DR.JIA-LING, KOH SPEAKER : SHUN-CHEN,
Diversity Maximization Under Matroid Constraints Date : 2013/11/06 Source : KDD’13 Authors : Zeinab Abbassi, Vahab S. Mirrokni, Mayur Thakur Advisor :
Date: 2012/8/13 Source: Luca Maria Aiello. al(CIKM’11) Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou Behavior-driven Clustering of Queries into Topics.
Date : 2013/05/27 Author : Anish Das Sarma, Lujun Fang, Nitin Gupta, Alon Halevy, Hongrae Lee, Fei Wu, Reynold Xin, Gong Yu Source : SIGMOD’12 Speaker.
CUBELSI : AN EFFECTIVE AND EFFICIENT METHOD FOR SEARCHING RESOURCES IN SOCIAL TAGGING SYSTEMS Bin Bi, Sau Dan Lee, Ben Kao, Reynold Cheng The University.
Time-sensitive Personalized Query Auto-Completion
Toward Whole-Session Relevance: Exploring Intrinsic Diversity in Web Search Date: 2014/5/20 Author: Karthik Raman, Paul N. Bennett, Kevyn Collins-Thompson.
Query Dependent Pseudo-Relevance Feedback based on Wikipedia SIGIR ‘09 Advisor: Dr. Koh Jia-Ling Speaker: Lin, Yi-Jhen Date: 2010/01/24 1.
Searchable Web sites Recommendation Date : 2012/2/20 Source : WSDM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh Jia-ling 1.
Context-aware Query Suggestion by Mining Click-through and Session Data Authors: H. Cao et.al KDD 08 Presented by Shize Su 1.
Evaluating Search Engine
Creating Concept Hierarchies in a Customer Self-Help System Bob Wall CS /29/05.
J. Chen, O. R. Zaiane and R. Goebel An Unsupervised Approach to Cluster Web Search Results based on Word Sense Communities.
SIGIR’09 Boston 1 Entropy-biased Models for Query Representation on the Click Graph Hongbo Deng, Irwin King and Michael R. Lyu Department of Computer Science.
1 Hierarchical Tag visualization and application for tag recommendations CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Personalization in Local Search Personalization of Content Ranking in the Context of Local Search Philip O’Brien, Xiao Luo, Tony Abou-Assaleh, Weizheng.
Tag Clouds Revisited Date : 2011/12/12 Source : CIKM’11 Speaker : I- Chih Chiu Advisor : Dr. Koh. Jia-ling 1.
SEEKING STATEMENT-SUPPORTING TOP-K WITNESSES Date: 2012/03/12 Source: Steffen Metzger (CIKM’11) Speaker: Er-gang Liu Advisor: Dr. Jia-ling Koh 1.
1 Context-Aware Search Personalization with Concept Preference CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Date: 2012/10/18 Author: Makoto P. Kato, Tetsuya Sakai, Katsumi Tanaka Source: World Wide Web conference (WWW "12) Advisor: Jia-ling, Koh Speaker: Jiun.
A Comparative Study of Search Result Diversification Methods Wei Zheng and Hui Fang University of Delaware, Newark DE 19716, USA
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
1 Applying Collaborative Filtering Techniques to Movie Search for Better Ranking and Browsing Seung-Taek Park and David M. Pennock (ACM SIGKDD 2007)
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Speaker: Ruirui Li 1 The University of Hong Kong.
Personalized Search Cheng Cheng (cc2999) Department of Computer Science Columbia University A Large Scale Evaluation and Analysis of Personalized Search.
Exploring Online Social Activities for Adaptive Search Personalization CIKM’10 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
A Probabilistic Graphical Model for Joint Answer Ranking in Question Answering Jeongwoo Ko, Luo Si, Eric Nyberg (SIGIR ’ 07) Speaker: Cho, Chin Wei Advisor:
Similar Document Search and Recommendation Vidhya Govindaraju, Krishnan Ramanathan HP Labs, Bangalore, India JOURNAL OF EMERGING TECHNOLOGIES IN WEB INTELLIGENCE.
Clustering Personalized Web Search Results Xuehua Shen and Hong Cheng.
Quantitative Comparisons of Search Engine Results Mike Thlwall School of Computing and Information Technology, University of Wolverhampton ( 伍爾弗漢普頓 UK)
1 Statistical source expansion for question answering CIKM’11 Advisor : Jia Ling, Koh Speaker : SHENG HONG, CHUNG.
Detecting Dominant Locations from Search Queries Lee Wang, Chuang Wang, Xing Xie, Josh Forman, Yansheng Lu, Wei-Ying Ma, Ying Li SIGIR 2005.
Contextual Ranking of Keywords Using Click Data Utku Irmak, Vadim von Brzeski, Reiner Kraft Yahoo! Inc ICDE 09’ Datamining session Summarized.
Probabilistic Models of Novel Document Rankings for Faceted Topic Retrieval Ben Cartrette and Praveen Chandar Dept. of Computer and Information Science.
Query Suggestion Naama Kraus Slides are based on the papers: Baeza-Yates, Hurtado, Mendoza, Improving search engines by query clustering Boldi, Bonchi,
Diversifying Search Result WSDM 2009 Intelligent Database Systems Lab. School of Computer Science & Engineering Seoul National University Center for E-Business.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
1 Date: 2012/9/13 Source: Yang Song, Dengyong Zhou, Li-wei Heal(WSDM’12) Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou Query Suggestion by Constructing.
Templated Search over Relational Databases Date: 2015/01/15 Author: Anastasios Zouzias, Michail Vlachos, Vagelis Hristidis Source: ACM CIKM’14 Advisor:
Date : 2013/03/18 Author : Jeffrey Pound, Alexander K. Hudek, Ihab F. Ilyas, Grant Weddell Source : CIKM’12 Speaker : Er-Gang Liu Advisor : Prof. Jia-Ling.
Jiafeng Guo(ICT) Xueqi Cheng(ICT) Hua-Wei Shen(ICT) Gu Xu (MSRA) Speaker: Rui-Rui Li Supervisor: Prof. Ben Kao.
Institute of Computing Technology, Chinese Academy of Sciences 1 A Unified Framework of Recommending Diverse and Relevant Queries Speaker: Xiaofei Zhu.
Diversifying Search Results Rakesh Agrawal, Sreenivas Gollapudi, Alan Halverson, Samuel Ieong Search Labs, Microsoft Research WSDM, February 10, 2009 TexPoint.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Date: 2012/08/21 Source: Zhong Zeng, Zhifeng Bao, Tok Wang Ling, Mong Li Lee (KEYS’12) Speaker: Er-Gang Liu Advisor: Dr. Jia-ling Koh 1.
Post-Ranking query suggestion by diversifying search Chao Wang.
Date: 2012/11/29 Author: Chen Wang, Keping Bi, Yunhua Hu, Hang Li, Guihong Cao Source: WSDM’12 Advisor: Jia-ling, Koh Speaker: Shun-Chen, Cheng.
ASSOCIATIVE BROWSING Evaluating 1 Jinyoung Kim / W. Bruce Croft / David Smith for Personal Information.
Advisor: Koh Jia-Ling Nonhlanhla Shongwe EFFICIENT QUERY EXPANSION FOR ADVERTISEMENT SEARCH WANG.H, LIANG.Y, FU.L, XUE.G, YU.Y SIGIR’09.
Multi-Aspect Query Summarization by Composite Query Date: 2013/03/11 Author: Wei Song, Qing Yu, Zhiheng Xu, Ting Liu, Sheng Li, Ji-Rong Wen Source: SIGIR.
Divided Pretreatment to Targets and Intentions for Query Recommendation Reporter: Yangyang Kang /23.
PERSONALIZED DIVERSIFICATION OF SEARCH RESULTS Date: 2013/04/15 Author: David Vallet, Pablo Castells Source: SIGIR’12 Advisor: Dr.Jia-ling, Koh Speaker:
Predicting User Interests from Contextual Information R. W. White, P. Bailey, L. Chen Microsoft (SIGIR 2009) Presenter : Jae-won Lee.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Predicting Short-Term Interests Using Activity-Based Search Context CIKM’10 Advisor: Jia Ling, Koh Speaker: Yu Cheng, Hsieh.
CS791 - Technologies of Google Spring A Web­based Kernel Function for Measuring the Similarity of Short Text Snippets By Mehran Sahami, Timothy.
Improving Search Relevance for Short Queries in Community Question Answering Date: 2014/09/25 Author : Haocheng Wu, Wei Wu, Ming Zhou, Enhong Chen, Lei.
Where Did You Go: Personalized Annotation of Mobility Records
Learning to Personalize Query Auto-Completion
Inferring People’s Site Preference in Web Search
Mining Query Subtopics from Search Log Data
Intent-Aware Semantic Query Annotation
Date : 2013/1/10 Author : Lanbo Zhang, Yi Zhang, Yunfei Chen
Learning to Rank with Ties
Connecting the Dots Between News Article
Preference Based Evaluation Measures for Novelty and Diversity
Presentation transcript:

DQR : A Probabilistic Approach to Diversified Query recommendation Date: 2013/05/20 Author: Ruirui Li, Ben Kao, Bin Bi, Reynold Cheng, Eric Lo Source: CIKM "12 Advisor: Jia-ling, Koh Speaker: Jiun Jia, Chiou 1

2 Outline ▪ Introduction ▪ T▪ The DQR Approach ▪C▪Concept Mining ▪Q▪Query Recommendation ▪E▪Experiment ▪C▪Conclusion

Introduction 3 The effectiveness of keyword-based search engines, depends largely on the ability of a user to formulate proper queries that are both expressive and selective. Jaguar Disney Theme Parks Stores Cartoon characters on the film-making studios ambiguous Short & not specific

Introduction 4 Web search queries issued by casual users are often short and with limited expressiveness. Traditional similarity-based methods often result in redundant and monotonic recommendations Query recommendation Help users refine their queries Diversified Redundancy free

5 The DQR framework

Sample AOL search log records (1) It clusters search log queries to extract query concepts, based on which recommended queries are selected. (2) It employs a probabilistic model and a algorithm to achieve recommendation diversification. Relevance Redundancy Free Diversity Ranking Efficiency 6

7 Outline ▪ I▪ Introduction ▪ T▪ The DQR Approach ▪C▪Concept Mining ▪Q▪Query Recommendation ▪E▪Experiment ▪C▪Conclusion

8 DQR Approach-Concept Mining Using a similarity measure to cluster queries, and use the set of URLs as the features of queries. : the number of unique users who have issued query q i and subsequently clicked URL d j : the number of queries that led to the clicking of URL d j user-frequency-inverse-query-frequency (UF-IQF)

9 User IDQueryClicked URL 001q1d1 002q1d2 003q1d1 004q2d2 005q2d1 006q2d3 007q3d3 |Q|=3, |D|=3 in search log w(q1,d1) = 2*log(3/2) = 3.16 w(q1,d2) = 1*log(3/2) = 1.58 w(q1,d3) = 0 Feature vector : w(q2,d1) = 1*log(3/2) = 1.58 w(q2,d2) = 1*log(3/2) = 1.58 w(q2,d3) = 1*log(3/2) = 1.58 Feature vector : Normalized Feature vector :

10 q1 Feature vector : q2 Feature vector : d( q1,q2 )= = 0.467

11

12

13 Outline ▪ I▪ Introduction ▪ T▪ The DQR Approach ▪C▪Concept Mining ▪Q▪Query Recommendation ▪E▪Experiment ▪C▪Conclusion

14 DQR Approach-Query Recommendation A user issues a query q to a search engine, he/she may click a set of URLs, s, returned by the search engine. They use I: q → s to denote such an interaction, which is captured by a set of records in the search log. Query(Q) : q1: mapsq2: map searchq3: driving directionsq4: rand mcnally Click sets(CS) : s1: maps.yahoo.com, maps.google.com s2: Interactions(I): q1→ s1q2→ s1q2→ s2q3→ s2q2→ s3q4→ s3

15 Input query [ Concept Selection ] Diversified Concept Recommendation

16 q1 q2 q3 q4 q5 q6 q7 C1C2C3 QueryClicked URL q1s1 q1s2 q2s1 q2s3 q3s1 q3s2 q3s2 q3s3 QueryClicked URL q4s1 q4s2 q4s2 q4s3 q5s2 q5s3 QueryClicked URL q6s1 q6s1 q6s2 q6s3 q6s3 q7s2 q7s3 q7s3 q7s3 Assume: Query: q4, CS={s1,s2,s3}

17 q1 q2 q3 q4 q5 q6 q7 C1C2C3 QueryClicked URL q1s1 q1s2 q2s1 q2s3 q3s1 q3s2 q3s2 q3s3 QueryClicked URL q4s1 q4s2 q4s2 q4s3 q5s2 q5s3 QueryClicked URL q6s1 q6s1 q6s2 q6s3 q6s3 q7s2 q7s3 q7s3 q7s3 Assume: Query: q4, CS={s1,s2,s3} C2 Yc’

18 q1 q2 q3 q4 q5 q6 q7 C1C2C3 QueryClicked URL q1s1 q1s2 q2s1 q2s3 q3s1 q3s2 q3s2 q3s3 QueryClicked URL q4s1 q4s2 q4s2 q4s3 q5s2 q5s3 QueryClicked URL q6s1 q6s1 q6s2 q6s3 q6s3 q7s2 q7s3 q7s3 q7s3 Assume: Query: q4, CS={s1,s2,s3} C2 Yc’

19 q1 q2 q3 q4 q5 q6 q7 C1C2C3 QueryClicked URL q1s1 q1s2 q2s1 q2s3 q3s1 q3s2 q3s2 q3s3 QueryClicked URL q4s1 q4s2 q4s2 q4s3 q5s2 q5s3 QueryClicked URL q6s1 q6s2 q6s2 q6s3 q6s2 q7s2 q7s3 q7s3 q7s3 Assume: Query: q4, CS={s1,s2,s3} C2 Yc’

C2 Yc’ C2 C1 Yc’ C2 Yc’ C2 C3 Yc’ = =

21 [ Pick one representative query from each concept ] UserID QueryClicked URL 001q4s1 002q5s2 003q5s2 004q6s3 005q4s1 006q5s3 q4 : 1 q5 : 3 q6 : 2 C4C4 C2C2 C6C6 C9C9

22 Outline ▪ I▪ Introduction ▪ T▪ The DQR Approach ▪C▪Concept Mining ▪Q▪Query Recommendation ▪E▪Experiment ▪C▪Conclusion

23 Experiment Data Sets : AOL ─ search log SOGOU ─ search log of Chinese search engine 1)Removing non-alpha-numerical characters from query strings. 2)Removing redundant space characters. 3)Removing queries that appear only once in the search log.

24 Experiment compactness requirement: “ pairwise distance of queries in a cluster ≤ 0.5”.

25 Experiment Typo queries Part 1 :Comparing different recommendation methods – recommendation lists for the query “yahoo” using the AOL dataset. some Typo queries Similarity-based recommender(SR) : Relevant but Redundant Maximal Marginal Relevance(MMR) : Diverse but Redundant DQR-ND(No Diversity) and DQR : Redundancy-free Context-Aware Concept-Based(CACB) : not match the user’s search intent (“yahoo”)

26 Experiment Part 2 : Evaluate the performances of the recommendation methods 12 people to judge the recommendations given by the six methods Select 60 test queries from the AOL search log Evaluation form:

27 N 1,2 : the average number of partially relevant or relevant recommended queries in an evaluation. S 1,2 : the average score of non-irrelevant recommended queries. S 0,1,2 : the average score of all the recommended queries. Relevancy Performance

28 Diversity Performance  Define : ( ) number of search intents so identified among the top k recommended queries on an evaluation form.

29 Ranking Performance Two Ranking Measure: 1)Normalized Discounted Cumulative Gain (NDCG) 2) Mean Reciprocal Rank (MRR) Recommendation query rq1 rq2 rq3 rq4 rq5 rq6 rq7 rq8 rel irrel rel partial_rel irrel rel irrel

30 Experiment

31 Outline ▪ I▪ Introduction ▪ T▪ The DQR Approach ▪C▪Concept Mining ▪Q▪Query Recommendation ▪E▪Experiment ▪C▪Conclusion

Conclusion  In this paper, they investigated the problem of recommending queries to better capture the search intents of search engine users.  Five important requirements of a query recommendation a) Relevancy b) Redundancy-free c) Diversity d) Ranking-performance e) Efficiency  Through a comprehensive user study, they showed that DQR performed very well. 32