Www.monash.edu.au Performance Measurement. www.monash.edu.au 2 Testing Environment.

Slides:



Advertisements
Similar presentations
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
Advertisements

Chapter 5: Introduction to Information Retrieval
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
1 Retrieval Performance Evaluation Modern Information Retrieval by R. Baeza-Yates and B. Ribeiro-Neto Addison-Wesley, (Chapter 3)
Ranked Retrieval INST 734 Module 3 Doug Oard. Agenda  Ranked retrieval Similarity-based ranking Probability-based ranking.
Exercising these ideas  You have a description of each item in a small collection. (30 web sites)  Assume we are looking for information about boxers,
Introduction to Information Retrieval (Part 2) By Evren Ermis.
Information Retrieval IR 7. Recap of the last lecture Vector space scoring Efficiency considerations Nearest neighbors and approximations.
Evaluating Search Engine
Search Engines and Information Retrieval
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.

Modern Information Retrieval
Information Retrieval in Practice
INFO 624 Week 3 Retrieval System Evaluation
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
CS 430 / INFO 430 Information Retrieval
Evaluating the Performance of IR Sytems
1 CS 430 / INFO 430 Information Retrieval Lecture 3 Vector Methods 1.
SLIDE 1IS 240 – Spring 2007 Prof. Ray Larson University of California, Berkeley School of Information Tuesday and Thursday 10:30 am - 12:00.
Web Search – Summer Term 2006 II. Information Retrieval (Basics Cont.) (c) Wolfgang Hürst, Albert-Ludwigs-University.
WXGB6106 INFORMATION RETRIEVAL Week 3 RETRIEVAL EVALUATION.
Information Retrieval
1 CS 502: Computing Methods for Digital Libraries Lecture 11 Information Retrieval I.
Evaluation David Kauchak cs458 Fall 2012 adapted from:
Evaluation David Kauchak cs160 Fall 2009 adapted from:
Search Engines and Information Retrieval Chapter 1.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
1 CS 430 / INFO 430 Information Retrieval Lecture 2 Text Based Information Retrieval.
Modern Information Retrieval Computer engineering department Fall 2005.
Evaluating Search Engines in chapter 8 of the book Search Engines Information Retrieval in Practice Hongfei Yan.
Thanks to Bill Arms, Marti Hearst Documents. Last time Size of information –Continues to grow IR an old field, goes back to the ‘40s IR iterative process.
Querying Structured Text in an XML Database By Xuemei Luo.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
Lecture 1: Overview of IR Maya Ramanath. Who hasn’t used Google? Why did Google return these results first ? Can we improve on it? Is this a good result.
WIRED Week 3 Syllabus Update (next week) Readings Overview - Quick Review of Last Week’s IR Models (if time) - Evaluating IR Systems - Understanding Queries.
Structure of IR Systems INST 734 Module 1 Doug Oard.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Basic Implementation and Evaluations Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
© 2004 Chris Staff CSAW’04 University of Malta of 15 Expanding Query Terms in Context Chris Staff and Robert Muscat Department of.
Measuring How Good Your Search Engine Is. *. Information System Evaluation l Before 1993 evaluations were done using a few small, well-known corpora of.
1 Information Retrieval LECTURE 1 : Introduction.
Information Retrieval CSE 8337 Spring 2007 Introduction/Overview Some Material for these slides obtained from: Modern Information Retrieval by Ricardo.
Information Retrieval
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
1 CS 430 / INFO 430 Information Retrieval Lecture 8 Evaluation of Retrieval Effectiveness 1.
Web Search and Text Mining Lecture 5. Outline Review of VSM More on LSI through SVD Term relatedness Probabilistic LSI.
What Does the User Really Want ? Relevance, Precision and Recall.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Evaluation of Information Retrieval Systems Xiangming Mu.
Evaluation. The major goal of IR is to search document relevant to a user query. The evaluation of the performance of IR systems relies on the notion.
Information Retrieval Quality of a Search Engine.
Xiaoying Gao Computer Science Victoria University of Wellington COMP307 NLP 4 Information Retrieval.
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
1 CS 430 / INFO 430 Information Retrieval Lecture 10 Evaluation of Retrieval Effectiveness 1.
Information Retrieval in Practice
Text Based Information Retrieval
Multimedia Information Retrieval
Evaluation.
Information Retrieval
Modern Information Retrieval
IR Theory: Evaluation Methods
موضوع پروژه : بازیابی اطلاعات Information Retrieval
Evaluation of IR Performance
INF 141: Information Retrieval
Information Retrieval and Web Design
Presentation transcript:

Performance Measurement

2 Testing Environment

3 Test Collection(1) Traditional collections: ADI, MEDLINE, CACM.

4 Test Collection(2) Text Retrieval Conference (TREC) - Large collection – up to 5 Gb Relevant judgement is created using polling methods. –Each participants sends a ranked output to a given query. –A relevant judgment is created on the merging of the output.

5 Measure the Test We then run both systems to produce, say, two different rankings. A simple way of comparing is to provide average precision at various levels of recall for both systems and examine the numbers. The system with the highest number wins!

6 Performance Measures A perfect retrieval system will show exactly the documents in which a user is interested, and show them in the appropriate order. The effectiveness of a text retrieval system can be measures in terms of recall and precision.

7 Performance Measure Precision: –proportion of retrieved items actually relevant Recall: –proportion of relevant information actually retrieved

8 Measures Recall = No. of relevant documents retrieved / no. of relevant documents Precision = No. of relevant documents retrieved / No of retrieved documents

9 Measures Notice that the more documents that are retrieved the easier it is to obtain better recall, and the fewer documents retrieved the easier it is obtain better precision. the diagram illustrates this. The ideal is to obtain 1.0 for both! Document Space Retrieved Relevant Retrieved Relevant Improved precision Improved Recall

10 Example A document database consists of 20 documents. A query submitted to a system resulted in the ranked output: D1,D2,D3,…,D20. Assume the relevant documents are D1,D3,D7,D10. No of documents viewedRecallPrecision 1¼1 2¼½ 32/42/3 ……… 104/44/10

11 Recall-Precision Sometime it is easier to see the relation between recall and precision when precision is measured at a given recall level, eg precision at 10%, 20% recall.

12 Recall and Precision Relation Inverse relation. –Increase in recall usually leads to decrease in precision.

13 Recall and Precision The designers of an IR system need to be aware of the user requirements. –Is precision the most important consideration? >Eg. Search engine –Is recall the most important consideration? >Eg. Patent office, Law Firm searches for similar cases.

14 Is 100% Recall and Precision Possible? An ideal IR system should produce 100% recall and precision. Impossible: –Recall and precision relations. –Imprecise formulation of information need. –Ambiguity of language.

15 Other Measures Other important measures include: – time taken to index documents – speed with which documents are matched against the query and retrieved – storage overhead required for the indexing system.