Evaluating the Performance of IR Sytems

Slides:



Advertisements
Similar presentations
Introduction to Information Retrieval Introduction to Information Retrieval Lecture 7: Scoring and results assembly.
Advertisements

Chapter 5: Introduction to Information Retrieval
Introduction to Information Retrieval
Multimedia Database Systems
Lecture 11 Search, Corpora Characteristics, & Lucene Introduction.
Introduction to Information Retrieval (Part 2) By Evren Ermis.
Intelligent Information Retrieval 1 Vector Space Model for IR: Implementation Notes CSC 575 Intelligent Information Retrieval These notes are based, in.
Evaluating Search Engine
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
Information Retrieval Review
Database Management Systems, R. Ramakrishnan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides.
Modern Information Retrieval
1 CS 430 / INFO 430 Information Retrieval Lecture 12 Probabilistic Information Retrieval.
Modern Information Retrieval Chapter 5 Query Operations.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
Information Retrieval Ch Information retrieval Goal: Finding documents Search engines on the world wide web IR system characters Document collection.
Computer comunication B Information retrieval. Information retrieval: introduction 1 This topic addresses the question on how it is possible to find relevant.
1 CS 430 / INFO 430 Information Retrieval Lecture 3 Vector Methods 1.
9/21/2000Information Organization and Retrieval Ranking and Relevance Feedback Ray Larson & Marti Hearst University of California, Berkeley School of Information.
WXGB6106 INFORMATION RETRIEVAL Week 3 RETRIEVAL EVALUATION.
Information Retrieval
HYPERGEO 1 st technical verification ARISTOTLE UNIVERSITY OF THESSALONIKI Baseline Document Retrieval Component N. Bassiou, C. Kotropoulos, I. Pitas 20/07/2000,
Recuperação de Informação. IR: representation, storage, organization of, and access to information items Emphasis is on the retrieval of information (not.
1 CS 502: Computing Methods for Digital Libraries Lecture 11 Information Retrieval I.
CS246 Basic Information Retrieval. Today’s Topic  Basic Information Retrieval (IR)  Bag of words assumption  Boolean Model  Inverted index  Vector-space.
Chapter 5: Information Retrieval and Web Search
CSCI 5417 Information Retrieval Systems Jim Martin Lecture 6 9/8/2011.
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
1 Vector Space Model Rong Jin. 2 Basic Issues in A Retrieval Model How to represent text objects What similarity function should be used? How to refine.
Search Engines and Information Retrieval Chapter 1.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
1 CS 430 / INFO 430 Information Retrieval Lecture 2 Text Based Information Retrieval.
CSE 6331 © Leonidas Fegaras Information Retrieval 1 Information Retrieval and Web Search Engines Leonidas Fegaras.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Chapter 6: Information Retrieval and Web Search
1 Computing Relevance, Similarity: The Vector Space Model.
CSE3201/CSE4500 Term Weighting.
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
CPSC 404 Laks V.S. Lakshmanan1 Computing Relevance, Similarity: The Vector Space Model Chapter 27, Part B Based on Larson and Hearst’s slides at UC-Berkeley.
IR Homework #1 By J. H. Wang Mar. 21, Programming Exercise #1: Vector Space Retrieval Goal: to build an inverted index for a text collection, and.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Basic Implementation and Evaluations Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Search Engines WS 2009 / 2010 Prof. Dr. Hannah Bast Chair of Algorithms and Data Structures Department of Computer Science University of Freiburg Lecture.
Performance Measurement. 2 Testing Environment.
Information Retrieval
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
Evaluation of Information Retrieval Systems Xiangming Mu.
Natural Language Processing Topics in Information Retrieval August, 2002.
Evaluation. The major goal of IR is to search document relevant to a user query. The evaluation of the performance of IR systems relies on the notion.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
Relevance Feedback Prof. Marti Hearst SIMS 202, Lecture 24.
Information Retrieval and Web Search IR models: Vector Space Model Term Weighting Approaches Instructor: Rada Mihalcea.
3: Search & retrieval: Structures. The dog stopped attacking the cat, that lived in U.S.A. collection corpus database web d1…..d n docs processed term-doc.
IR 6 Scoring, term weighting and the vector space model.
Sampath Jayarathna Cal Poly Pomona
Text Based Information Retrieval
Information Retrieval and Web Search
IST 516 Fall 2011 Dongwon Lee, Ph.D.
Multimedia Information Retrieval
Evaluation.
Modern Information Retrieval
IR Theory: Evaluation Methods
Basic Information Retrieval
موضوع پروژه : بازیابی اطلاعات Information Retrieval
CS 430: Information Discovery
Chapter 5: Information Retrieval and Web Search
INF 141: Information Retrieval
Information Retrieval and Web Design
Presentation transcript:

Evaluating the Performance of IR Sytems

Outline Summary of IR system. Evaluating IR indexing: stop list, stemming, term weights file organisation for term indexes. retrieval model: vector space Evaluating IR Recall and Precision

Indexing Process Objective: Refinements: creating general and specific descriptions of a document. Refinements: Stop list: removing common words/terms. Stemming: reducing the various formed of words into a single token. Term weights: tf * idf tf: number of occurrence of a term in a document idf: the inverse on the number of documents in the collection that contains a term.

Inverted File Doc-id Word-id weight word-id doc-id weight 1 45 10 75 5 200 3 12 6 2 16 38 212 4 word-id doc-id weight 12 1 6 16 2 38 45 10 75 5 200 3 212 4 Inverted file Sorted on word-id Normal indexed file Sorted on doc-id

A query after indexing: 16, 212 Inverted File - Search word-id doc-id weight 12 1 6 16 2 38 45 10 75 5 200 3 212 4 documents Doc.1 Doc.2 Doc.3 A query after indexing: 16, 212

The Vector Space Model Each document and query is represented by a vector. A vector is obtained for each document and query from sets of index terms with associated weights. The document and query representatives are considered as vectors in n dimensional space where n is the number of unique terms in the dictionary/document collection. Measuring vectors similarity: inner product value of cosine of the angle between the two vectors.

Inner Product

Cosine Using the weight of the term as the components of D and Q:

Simple Example The vectors for the query and documents: Doc# wt1 wt2 1 0.5 0.3 2 0.6 3 0.4 Doc-1= (0.5,0.3) Doc-2= (0.6,0) Doc-3= (0,0.4) Query = ( 0, 0.5)

Performance Measurement

Performance Measures A perfect retrieval system will show exactly the documents in which a user is interested, and show them in the appropriate order. The effectiveness of a text retrieval system can be measures in terms of recall and precision.

Performance Measure Precision: Recall: proportion of retrieved items actually relevant Recall: proportion of relevant information actually retrieved

Measures Recall = No. of relevant documents retrieved / no. of relevant documents Precision = No. of relevant documents retrieved / No of retrieved documents

Measures Notice that the more documents that are retrieved the easier it is to obtain better recall, and the fewer documents retrieved the easier it is obtain better precision. the diagram illustrates this. The ideal is to obtain 1.0 for both! Relevant Document Space Retrieved Retrieved Retrieved Improved Recall Improved precision Relevant Relevant

Example A document database consists of 20 documents. A query submitted to a system resulted in the ranked output: D1,D2,D3,…,D20. Assume the relevant documents are D1,D3,D7,D10. No of documents viewed Recall Precision 1 ¼ 2 ½ 3 2/4 2/3 … 10 4/4 4/10

Recall-Precision Sometime it is easier to see the relation between recall and precision when precision is measured at a given recall level, eg precision at 10%, 20% recall.

Recall and Precision Relation Inverse relation. Increase in recall usually leads to decrease in precision.

Recall and Precision The designers of an IR system need to be aware of the user requirements. Is precision the most important consideration? Eg. Search engine Is recall the most important consideration? Eg. Patent office, Law Firm searches for similar cases.

Is 100% Recall and Precision Possible? An ideal IR system should produce 100% recall and precision. Impossible: Recall and precision relations. Imprecise formulation of information need. Ambiguity of language.

Which System Is Better? Suppose two systems produce different ranking of documents. How do we decide which one is better? We need to prepare the test environment and run some experiments. Test environment: a test database, a set of queries, and the relevance judgments for these queries.

Test Environment

Test Collection(1) Traditional collections: ADI, MEDLINE, CACM.

Test Collection(2) Text Retrieval Conference (TREC) - http://trec.nist.gov/ Large collection – up to 5 Gb Relevant judgement is created using polling methods. Each participants sends a ranked output to a given query. A relevant judgment is created on the merging of the output.

Measure the Test We then run both systems to produce, say, two different rankings. A simple way of comparing is to provide average precision at various levels of recall for both systems and examine the numbers. The system with the highest number wins!

Other Measures Other important measures include: time taken to index documents speed with which documents are matched against the query and retrieved storage overhead required for the indexing system.

Relevance Feedback

IR Model - Revisited

Search Process Most of IR systems allow user to refine their query. Why? Some new knowledge may be gained by the user after reading some of the suggested documents. A clearer understanding of the information need may be gained by reading the suggested documents. The refining process is called Relevance Feedback.

Relevance Feedback Relevance Feedback can be achieved by means of: Re-weight the terms in the query. Use the documents as an example, eg find documents which discussed similar topic to D2,D6,D8 but NOT like D1,D5.