Relevance feedback using query-logs Gaurav Pandey Supervisors: Prof. Gerhard Weikum Julia Luxenburger.

Slides:



Advertisements
Similar presentations
Relevance Feedback User tells system whether returned/disseminated documents are relevant to query/information need or not Feedback: usually positive sometimes.
Advertisements

Evaluating the Robustness of Learning from Implicit Feedback Filip Radlinski Thorsten Joachims Presentation by Dinesh Bhirud
Improvements and extras Paul Thomas CSIRO. Overview of the lectures 1.Introduction to information retrieval (IR) 2.Ranked retrieval 3.Probabilistic retrieval.
Chapter 5: Introduction to Information Retrieval
Introduction to Information Retrieval
Optimizing search engines using clickthrough data
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
1 Learning User Interaction Models for Predicting Web Search Result Preferences Eugene Agichtein Eric Brill Susan Dumais Robert Ragno Microsoft Research.
Information Retrieval Visualization CPSC 533c Class Presentation Qixing Zheng March 22, 2004.
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Query Refinement: Relevance Feedback Information Filtering.
Chapter 5: Query Operations Baeza-Yates, 1999 Modern Information Retrieval.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
Ch 4: Information Retrieval and Text Mining
Modern Information Retrieval Chapter 5 Query Operations.
Recall: Query Reformulation Approaches 1. Relevance feedback based vector model (Rocchio …) probabilistic model (Robertson & Sparck Jones, Croft…) 2. Cluster.
Query Reformulation: User Relevance Feedback. Introduction Difficulty of formulating user queries –Users have insufficient knowledge of the collection.
The Vector Space Model …and applications in Information Retrieval.
1 CS 430 / INFO 430 Information Retrieval Lecture 10 Probabilistic Information Retrieval.
An Overview of Relevance Feedback, by Priyesh Sudra 1 An Overview of Relevance Feedback PRIYESH SUDRA.
CS344: Introduction to Artificial Intelligence Vishal Vachhani M.Tech, CSE Lecture 34-35: CLIR and Ranking in IR.
Faculty of Informatics and Information Technologies Slovak University of Technology Personalized Navigation in the Semantic Web Michal Tvarožek Mentor:
1 Cross-Lingual Query Suggestion Using Query Logs of Different Languages SIGIR 07.
Improving Web Search Ranking by Incorporating User Behavior Information Eugene Agichtein Eric Brill Susan Dumais Microsoft Research.
PERSONALIZED SEARCH Ram Nithin Baalay. Personalized Search? Search Engine: A Vital Need Next level of Intelligent Information Retrieval. Retrieval of.
Query Routing in Peer-to-Peer Web Search Engine Speaker: Pavel Serdyukov Supervisors: Gerhard Weikum Christian Zimmer Matthias Bender International Max.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Personalized Search Xiao Liu
CSE3201/CSE4500 Term Weighting.
LATENT SEMANTIC INDEXING Hande Zırtıloğlu Levent Altunyurt.
IR Theory: Relevance Feedback. Relevance Feedback: Example  Initial Results Search Engine2.
LANGUAGE MODELS FOR RELEVANCE FEEDBACK Lee Won Hee.
Faculty of Informatics and Information Technologies Slovak University of Technology Personalized Navigation in the Semantic Web Michal Tvarožek Mentor:
A Statistical Comparison of Tag and Query Logs Mark J. Carman, Robert Gwadera, Fabio Crestani, and Mark Baillie SIGIR 2009 June 4, 2010 Hyunwoo Kim.
Query Suggestion. n A variety of automatic or semi-automatic query suggestion techniques have been developed  Goal is to improve effectiveness by matching.
C.Watterscsci64031 Probabilistic Retrieval Model.
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
More Than Relevance: High Utility Query Recommendation By Mining Users' Search Behaviors Xiaofei Zhu, Jiafeng Guo, Xueqi Cheng, Yanyan Lan Institute of.
Intelligent Database Systems Lab Presenter: CHANG, SHIH-JIE Authors: Longzhuang Li, Yi Shang, Wei Zhang 2002.ACM. Improvement of HITS-based Algorithms.
Chaoyang University of Technology Clustering web transactions using rough approximation Source : Fuzzy Sets and Systems 148 (2004) 131–138 Author : Supriya.
Divided Pretreatment to Targets and Intentions for Query Recommendation Reporter: Yangyang Kang /23.
Adaptive Faceted Browsing in Job Offers Danielle H. Lee
Artificial Intelligence Techniques Internet Applications 4.
The Development of a search engine & Comparison according to algorithms Sung-soo Kim The final report.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
CS307P-SYSTEM PRACTICUM CPYNOT. B13107 – Amit Kumar B13141 – Vinod Kumar B13218 – Paawan Mukker.
Contextual Text Cube Model and Aggregation Operator for Text OLAP
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Query Refinement and Relevance Feedback.
1 CS 430: Information Discovery Lecture 21 Interactive Retrieval.
Generate Reports. Generate Result Slips Assessment Report End of Year Report Year 7 & Year 8 SBA Report Year 8 Final Report Year 7 and Year 8 Reporting.
IR 6 Scoring, term weighting and the vector space model.
Plan for Today’s Lecture(s)
Evaluation Anisio Lacerda.
Information Retrieval on the World Wide Web
INFORMATION RETRIEVAL TECHNIQUES BY DR. ADNAN ABID
Relevance Feedback Hongning Wang
Author: Kazunari Sugiyama, etc. (WWW2004)
Personalized Search and Visualization for YourNews
موضوع پروژه : بازیابی اطلاعات Information Retrieval
Web Information retrieval (Web IR)
Agenda What is SEO ? How Do Search Engines Work? Measuring SEO success ? On Page SEO – Basic Practices? Technical SEO - Source Code. Off Page SEO – Social.
CS246: Leveraging User Feedback
Content Based Image Retrieval
Cumulated Gain-Based Evaluation of IR Techniques
Learning to Rank with Ties
MAPO: Mining and Recommending API Usage Patterns
VECTOR SPACE MODEL Its Applications and implementations
Presentation transcript:

Relevance feedback using query-logs Gaurav Pandey Supervisors: Prof. Gerhard Weikum Julia Luxenburger

Motivation Results Search Engine Query “ One size fits all ”

Motivation Results Search Engine User info Query

Motivation Python

Motivation Python CGI code Debugging programming

Usage of Query Logs Clickthrough data Past queries Documents clicked

Usage of Query Logs Query Clicked Documents History Instance

Query Reformulation Result Query: “python information CGI code examples program code debugging bug removal programming” But, p(python/query)=? p(CGI)/query)=? p(code)/query)=? …………………………..

Language Model Normally(without using history), w:term d: document q:query Importance of term w in current query Considers only the current query But, not history instances

Language Model Normally(without using history), w:term d: document q:query Importance of term w in current query Now, using history: Importance of term w in current query + history instances ?

Language Model+History Importance of the term w at one instance in the history Importance of term w in history instances History query: “ CGI code ” Documents: “CGI examples”, “program code ” History query: “ CGI code ” Documents: “CGI examples”, “program code ”

Equal Weighting Works,but can be improved

Discriminative Weighting Choose different for every history instance.. How?

Overlap if a history query has common terms with the current query then λ i = 1, Else if there is no common term λ i =0 Example: Current query “python information” History query:”python code” λ i = 1 History query:”world cup” λ i = 0

Soft overlap if a history query has common terms with the current query then λ i = a, Else if there is no common term λ i =b (a>b) Example: Current query “python information” History query:”python code” λ i = 8 History query:”world cup” λ i = 2

Decrease with time Use uniformly decreasing values If there are n history instances, 1 =n 2 =n-1 3 =n-2 …… n-1 =2 n =1

Decrease with time Use geometrically decreasing values If there are n history instances, 1 =n 2 =n/2 3 =n/3 …… n-1 =n/(n-1) n =1

Experiment Comparison of the  4 techniques  Equal weighting  Basic model (without history) Use similar techniques for:  Probabilistic model  Vector space model

Thanks