Knowledge and Information Retrieval Dr Nicholas Gibbins 32/4037.

Slides:



Advertisements
Similar presentations
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
Advertisements

Chapter 5: Introduction to Information Retrieval
Modern information retrieval Modelling. Introduction IR systems usually adopt index terms to process queries IR systems usually adopt index terms to process.
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Crawling, Ranking and Indexing. Organizing the Web The Web is big. Really big. –Over 3 billion pages, just in the indexable Web The Web is dynamic Problems:
Learning Algorithm Evaluation
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Probabilistic Language Processing Chapter 23. Probabilistic Language Models Goal -- define probability distribution over set of strings Unigram, bigram,
Evaluating Search Engine
Information Retrieval Ling573 NLP Systems and Applications April 26, 2011.
Information Retrieval Review
Evaluation.  Allan, Ballesteros, Croft, and/or Turtle Types of Evaluation Might evaluate several aspects Evaluation generally comparative –System A vs.
T.Sharon - A.Frank 1 Internet Resources Discovery (IRD) IR Queries.
Retrieval Evaluation: Precision and Recall. Introduction Evaluation of implementations in computer science often is in terms of time and space complexity.
ITCS 6010 Natural Language Understanding. Natural Language Processing What is it? Studies the problems inherent in the processing and manipulation of.
Computer comunication B Information retrieval Repetition Retrieval models Wildcards Web information retrieval Digital libraries.
1 CS 430: Information Discovery Lecture 2 Introduction to Text Based Information Retrieval.
Information Retrieval
CS246 Basic Information Retrieval. Today’s Topic  Basic Information Retrieval (IR)  Bag of words assumption  Boolean Model  Inverted index  Vector-space.
Evaluation of Image Retrieval Results Relevant: images which meet user’s information need Irrelevant: images which don’t meet user’s information need Query:
Chapter 5: Information Retrieval and Web Search
CSCI 347 / CS 4206: Data Mining Module 06: Evaluation Topic 07: Cost-Sensitive Measures.
Evaluating Classifiers
Modeling (Chap. 2) Modern Information Retrieval Spring 2000.
Query Relevance Feedback and Ontologies How to Make Queries Better.
Processing of large document collections Part 3 (Evaluation of text classifiers, applications of text categorization) Helena Ahonen-Myka Spring 2005.
Evaluation – next steps
1 CS 430 / INFO 430 Information Retrieval Lecture 2 Text Based Information Retrieval.
Web Searching Basics Dr. Dania Bilal IS 530 Fall 2009.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
1 Information Retrieval Acknowledgements: Dr Mounia Lalmas (QMW) Dr Joemon Jose (Glasgow)
Classification Performance Evaluation. How do you know that you have a good classifier? Is a feature contributing to overall performance? Is classifier.
Weighting and Matching against Indices. Zipf’s Law In any corpus, such as the AIT, we can count how often each word occurs in the corpus as a whole =
Chapter 6: Information Retrieval and Web Search
Introduction to Digital Libraries hussein suleman uct cs honours 2003.
Lecture 1: Overview of IR Maya Ramanath. Who hasn’t used Google? Why did Google return these results first ? Can we improve on it? Is this a good result.
Evaluation of (Search) Results How do we know if our results are any good? Evaluating a search engine  Benchmarks  Precision and recall Results summaries:
Chapter 23: Probabilistic Language Models April 13, 2004.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Computational Intelligence: Methods and Applications Lecture 16 Model evaluation and ROC Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Basic Implementation and Evaluations Aj. Khuanlux MitsophonsiriCS.426 INFORMATION RETRIEVAL.
Web- and Multimedia-based Information Systems Lecture 2.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Information Retrieval COMP3211 Advanced Databases Dr Nicholas Gibbins –
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Information Retrieval
What Does the User Really Want ? Relevance, Precision and Recall.
Ranking of Database Query Results Nitesh Maan, Arujn Saraswat, Nishant Kapoor.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Chapter 5: Credibility. Introduction Performance on the training set is not a good indicator of performance on an independent set. We need to predict.
CS798: Information Retrieval Charlie Clarke Information retrieval is concerned with representing, searching, and manipulating.
Information Retrieval Lecture 3 Introduction to Information Retrieval (Manning et al. 2007) Chapter 8 For the MSc Computer Science Programme Dell Zhang.
Information Retrieval COMP3017 Advanced Databases Dr Nicholas Gibbins –
Sampath Jayarathna Cal Poly Pomona
7. Performance Measurement
Designing Cross-Language Information Retrieval System using various Techniques of Query Expansion and Indexing for Improved Performance  Hello everyone,
Text Based Information Retrieval
7CCSMWAL Algorithmic Issues in the WWW
Evaluation of IR Systems
Multimedia Information Retrieval
Basic Information Retrieval
CS 430: Information Discovery
Chapter 5: Information Retrieval and Web Search
CS246: Information Retrieval
Dr. Sampath Jayarathna Cal Poly Pomona
INF 141: Information Retrieval
Information Retrieval and Web Design
Information Retrieval and Web Design
Dr. Sampath Jayarathna Cal Poly Pomona
Presentation transcript:

Knowledge and Information Retrieval Dr Nicholas Gibbins 32/4037

Introduction Information retrieval is the task of finding documents that are relevant to a user’s need for information Best known examples are search engines on the Web User types a query into a search engine and sees a list of relevant pages

Characterising IR Systems Document collection –Each system must decide what it wants to treat as a document: paragraph, page or multi-page text Query posed in query language –The query specifies what the user wants to know. Can be list of words [AI book], a phrase [“AI book”], contain boolean operators [AI AND book] and non-boolean operators [AI NEAR book] Result Set –Subset of documents that the system judges to be relevant (of use) to the query. Presentation of result set Can be a simple ranked list, 3-D map, etc.

Document Models Information retrieval systems are distinguished by how they represent documents, and how they match documents to queries Several common models: –Set-theoretic (boolean keywords) –Algebraic (vector spaces) –Probabilistic

Evaluating IR Systems Precision measures proportion of documents in the result set that are actually relevant Recall measures the population of all relevant documents in the collection that are in the result set. Very difficult to compute in WWW because no easy way to examine every page of the web Estimate recall by sampling or ignore completely

Relevance Relevance is (generally) a subjective notion Defined in terms of a user’s information need Users may differ in their assessment of relevance of a document to a particular query

Logical Relevance Logical relevance is a special case: –Relevance defined in terms of logical consequence –Queries represented as a set of sentences (the component sentences) –Contents of KB (stored sentences) may form premise for component sentences (if component sentence logically follows from stored sentence) –A stored sentence is logically relevant to an information need iff it is part of a premise for a component sentence of that need

Relevance The set of all documents The set of retrieved documents The set of relevant documents

Precision and Recall Precision = |A ∩ B| / |B| Recall = |A ∩ B| / |A| RetrievedNot Retrieved Relevant A ∩ BA ∩ ¬BA Not Relevant¬A ∩ B¬A ∩ ¬B¬A¬A B¬B¬B

Positives and Negatives True Positive Rate = Recall = TP / ( TP + FN ) True Negative Rate = Precision = TP / ( TP + FP ) False Positive Rate = FP / ( FP + TN ) = Fallout False Negative Rate = FN / ( TP + FN ) = 1 - Recall RetrievedNot Retrieved RelevantTrue PositiveFalse Negative Not RelevantFalse PositiveTrue Negative

Evaluation Precision = 30 / (30+10) = 0.75 Recall = 30 / (30+20) = 0.60 False positive rate = 10 / (10+40) = 0.20 False negative rate = 1 – 0.60 = 0.40 RetrievedNot Retrieved Relevant3020 Not Relevant1040

Precision versus Recall Can trade off precision against recall A system that returns every document in the collection as its result set guarantees recall of 100% but has low precision Returning a single document could give low recall but 100% precision

Precision versus Recall precision recall line of iso-effectiveness varying control parameter

Precision versus Recall precision recall Increasing effectiveness

Other Effectiveness Measures Specificity = TN / ( TN + FP ) –Measures how well the system rejects irrelevant documents Accuracy = TP + TN / ( TP + FP + FN + TN )

Measures ROC (receiver operating characteristic) curve summarises this –True Positive Rate on y axis (Recall or Sensitivity) –False Positive Rate on x axis (1 – Specificity) Reciprocal rank of first relevant result: if the first result is relevant it gets a score of 1; if the first two are not relevant but the third is, it gets a score of 1/3 Time to answer measures how long it takes a user to find the desired answer to a problem –Requires human subjects

Receiver Operating Characteristic Curve line of discrimination TPR FPR perfect better worse

Soundness and Completeness If a proof procedure is sound, it proves only valid formulae If a proof procedure is complete, it proves all valid formulae Parallel with precision and recall –Consider logical validity as relevance –A sound proof procedure has perfect (=1) precision –A complete proof procedure has perfect (=1) recall –Logical relevance is equivalent to logical validity

Word Significance word rank order word frequency Zipf’s law: P n ~= 1/n a PnPn n

Word Significance word rank order word frequency Resolving power of significant words significant words upper cut-off lower cut-off

Not all words are independent If query is [couch] do not exclude “COUCH”, “couches” Case folding to get couch from COUCH Stemming to root of word

Stemming Stemming to reduce “couches” to stem form “couch” Suffix stripping –“walked” to “walk” –“walking” to “walk” –“quickly” to “quick” Lemmatisation –Determine part of speech (noun, verb, etc) and normalise to root Yields 2% increase in recall for English More important in other languages Can harm precision: stock and stocking Newer algorithms use dictionaries to avoid this

Other Techniques Synonyms –Example: sofa for couch –Small gains in recall –Danger of decreasing precision if applied too aggressively: Tim Couch and sofa Spelling correction Metadata!!!

Presentation of Results Probability ranking principle –List ordered by probability of relevance –Good for speed –Doesn’t consider utility –If two copies of most relevant document, second has equal relevance but zero utility once you’ve seen the first. Many IR systems eliminate results that are too similar

Document Classification Results classified into pre-existing taxonomy of topics –Example: News as world news, local news, business news, sports news Document clustering creates categories from scratch for each result set Classification is good for small number of topics in collection, clustering good for broader collections such as WWW