1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.

Slides:



Advertisements
Similar presentations
TWO STEP EQUATIONS 1. SOLVE FOR X 2. DO THE ADDITION STEP FIRST
Advertisements

A B C D E F G H I J K L M N O P Q R S T U V W X Y Z
You have been given a mission and a code. Use the code to complete the mission and you will save the world from obliteration…
Advanced Piloting Cruise Plot.
Kapitel 21 Astronomie Autor: Bennett et al. Galaxienentwicklung Kapitel 21 Galaxienentwicklung © Pearson Studium 2010 Folie: 1.
Chapter 1 The Study of Body Function Image PowerPoint
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 5 Author: Julia Richards and R. Scott Hawley.
1 Copyright © 2010, Elsevier Inc. All rights Reserved Fig 2.1 Chapter 2.
By D. Fisher Geometric Transformations. Reflection, Rotation, or Translation 1.
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination.
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Jeopardy Q 1 Q 6 Q 11 Q 16 Q 21 Q 2 Q 7 Q 12 Q 17 Q 22 Q 3 Q 8 Q 13
Title Subtitle.
Exit a Customer Chapter 8. Exit a Customer 8-2 Objectives Perform exit summary process consisting of the following steps: Review service records Close.
Determine Eligibility Chapter 4. Determine Eligibility 4-2 Objectives Search for Customer on database Enter application signed date and eligibility determination.
My Alphabet Book abcdefghijklm nopqrstuvwxyz.
Multiplying binomials You will have 20 seconds to answer each of the following multiplication problems. If you get hung up, go to the next problem when.
0 - 0.
ALGEBRAIC EXPRESSIONS
DIVIDING INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
SUBTRACTING INTEGERS 1. CHANGE THE SUBTRACTION SIGN TO ADDITION
MULT. INTEGERS 1. IF THE SIGNS ARE THE SAME THE ANSWER IS POSITIVE 2. IF THE SIGNS ARE DIFFERENT THE ANSWER IS NEGATIVE.
FACTORING ax2 + bx + c Think “unfoil” Work down, Show all steps.
Addition Facts
Year 6 mental test 5 second questions
Around the World AdditionSubtraction MultiplicationDivision AdditionSubtraction MultiplicationDivision.
ZMQS ZMQS
Introduction Lesson 1 Microsoft Office 2010 and the Internet
BT Wholesale October Creating your own telephone network WHOLESALE CALLS LINE ASSOCIATED.
1 of Audience Survey Results Larry D. Gustke, Ph.D. – October 5, 2013.
ABC Technology Project
1 Undirected Breadth First Search F A BCG DE H 2 F A BCG DE H Queue: A get Undiscovered Fringe Finished Active 0 distance from A visit(A)
Dr. Lorayne Robertson, UOIT
© Charles van Marrewijk, An Introduction to Geographical Economics Brakman, Garretsen, and Van Marrewijk.
© Charles van Marrewijk, An Introduction to Geographical Economics Brakman, Garretsen, and Van Marrewijk.
© Charles van Marrewijk, An Introduction to Geographical Economics Brakman, Garretsen, and Van Marrewijk.
VOORBLAD.
1 Breadth First Search s s Undiscovered Discovered Finished Queue: s Top of queue 2 1 Shortest path from s.
Factor P 16 8(8-5ab) 4(d² + 4) 3rs(2r – s) 15cd(1 + 2cd) 8(4a² + 3b²)
Squares and Square Root WALK. Solve each problem REVIEW:
Traditional IR models Jian-Yun Nie.
© 2012 National Heart Foundation of Australia. Slide 2.
Lets play bingo!!. Calculate: MEAN Calculate: MEDIAN
Past Tense Probe. Past Tense Probe Past Tense Probe – Practice 1.
Understanding Generalist Practice, 5e, Kirst-Ashman/Hull
Chapter 5 Test Review Sections 5-1 through 5-4.
SIMOCODE-DP Software.
GG Consulting, LLC I-SUITE. Source: TEA SHARS Frequently asked questions 2.
Addition 1’s to 20.
Model and Relationships 6 M 1 M M M M M M M M M M M M M M M M
25 seconds left…...
Test B, 100 Subtraction Facts
1 Atlantic Annual Viewing Trends Adults 35-54, Total TV, By Daypart Average Minute Audience (000) Average Weekly Reach (%) Average Weekly Hours Viewed.
Januar MDMDFSSMDMDFSSS
Week 1.
We will resume in: 25 Minutes.
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
Figure Essential Cell Biology (© Garland Science 2010)
A SMALL TRUTH TO MAKE LIFE 100%
1 Unit 1 Kinematics Chapter 1 Day
PSSA Preparation.
1 PART 1 ILLUSTRATION OF DOCUMENTS  Brief introduction to the documents contained in the envelope  Detailed clarification of the documents content.
How Cells Obtain Energy from Food
Immunobiology: The Immune System in Health & Disease Sixth Edition
1 Chap 14 Ranking Algorithm 指導教授 : 黃三益 博士 學生 : 吳金山 鄭菲菲.
CpSc 3220 Designing a Database
Traktor- og motorlære Kapitel 1 1 Kopiering forbudt.
Presentation transcript:

1 Evaluations in information retrieval

2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied to assess the quality of »information retrieval systems, and more concretely of search systems »the resulting set of records obtained after performing a query in an information retrieval system Note: This should not be confused with assessing the quality and value of the content of an information source.

3 Evaluations in information retrieval: introduction The quality of the results, the outcome of any search using any retrieval system depends on many components / factors. These components can be evaluated and modified to increase the quality of the results more or less independently.

4 Evaluations in information retrieval: important factors The information retrieval system ( = contents + system) The user of the retrieval system and the search strategy applied to the system Result of a search

5 Evaluations in information retrieval: why? (Part 1) To study the differences in outcome/results when a component of a retrieval system is changed, such as »the user interface »the retrieval algorithm »addition by the database of uncontrolled, natural language keywords versus keywords selected from a more rigid, controlled vocabulary

6 Evaluations in information retrieval: why? (Part 2) To study the differences in outcome/results when a search strategy is changed. To study the differences in outcome/results when searches are performed by different groups of users, such as »children versus adults »inexperienced users versus more experienced, professional information intermediaries/professionals

7 Evaluations in information retrieval: the simple Boolean model Boolean model: # items in database = # items selected + # items not selected # Items selected = # relevant items + # irrelevant items Relevant Yes 1 In Irrelevant No 0 Out

8 Relevant items in a database: scheme Dependent on the aims, independent of the search strategy Relevant items! (In most cases the small subset) Irrelevant / NOT relevant items (In most cases the large subset)

9 Selecting relevant items by searching a database: scheme Dependent on the aims, independent of the search strategy Selected and relevant! Selected but not relevant Not selected but relevant Not selected and not relevant Dependent on the aims and dependent on the search strategy

10 Recall: definition and meaning Definition: # Of selected relevant items “Recall” = * 100% Total # of relevant items in database Aim: high recall Problem: in most practical cases, the total # of relevant items in a database cannot be measured.

11 Selecting relevant items: recall Selected and relevant! Selected but not relevant Not selected but relevant Not selected and not relevant

12 !? Question !? Task !? Problem !? How to use of the concept “recall”, when you do not know the total number of relevant items in the database ?

13 Recall: how to use the concept of recall Using the same database, variations in recall express the effect of search variations »Variations in search terms »Use of a classification scheme »Use of a thesaurus »...

14 !? Question !? Task !? Problem !? How can you change your search strategy to increase the recall?

15 Precision: definition and meaning Definition: # Of selected relevant items “Precision” = * 100% Total # of selected items Aim: high precision

16 Selecting relevant items: precision Selected and relevant! Selected but not relevant Not selected but relevant Not selected and not relevant

17 !? Question !? Task !? Problem !? How can you change your search strategy to increase the precision?

18 !? Question !? Task !? Problem !? When you change your search strategy to increase the precision, which consequence do you expect for the recall, in most cases?

19 Relation between recall and precision of searches 100% Recall 0 0 Precision 100% Ideal = Impossible to reach in most systems Ideal = Impossible to reach in most systems Search (results)

20 !? Question !? Task !? Problem !? Indicate on the figure that a user improves a search.

21 !? Question !? Task !? Problem !? Indicate on the figure that a database producer and / or the retrieval system improves the retrieval quality.

22 !? Question !? Task !? Problem !? Indicate the relation between the recall and precision in a classical information retrieval system in the form of a figure. Indicate in that figure a good and a bad search.

23 Recall and precision should be considered together Examples: Increase in retrieved number of relevant items may be accompanied by an impractical decrease in precision. Precision of a search close to 100% may NOT be ideal, because the recall of the search may be too low. Make search / query broader to increase recall ! Poor (low) precision is more noticeable than bad (low) recall.

24 Evaluation in the case of systems offering relevance ranking Many modern information retrieval systems offer output with relevance ranking. This is more complicated than simple Boolean retrieval, and the simple concepts of recall and precision cannot be applied. To compare retrieval systems or search strategies, decide to consider for comparison a particular number of items ranked highest in each output. This brings us to for instance: “first-20 precision”.

25 !? Question !? Task !? Problem !? Give examples of retrieval systems that offer relevance ranking.