User-centered Information System Evaluation Tutorial

Slides:



Advertisements
Similar presentations
DEVELOPING A METHODOLOGY FOR MS3305 CW2 Some guidance.
Advertisements

Fatma Y. ELDRESI Fatma Y. ELDRESI ( MPhil ) Systems Analysis / Programming Specialist, AGOCO Part time lecturer in University of Garyounis,
Developing Satisfaction Surveys: Integrating Qualitative and Quantitative Information David Cantor, Sarah Dipko, Stephanie Fry, Pamela Giambo and Vasudha.
CWS: A Comparative Web Search System Jian-Tao Sun, Xuanhui Wang, § Dou Shen Hua-Jun Zeng, Zheng Chen Microsoft Research Asia University of Illinois at.
EIONET European Environment Information and Observation Network Version * * * Quality assurance of Eurowaternet.
1 Evaluations in information retrieval. 2 Evaluations in information retrieval: summary The following gives an overview of approaches that are applied.
Efficient Top-k Search across Heterogeneous XML Data Sources Jianxin Li 1 Chengfei Liu 1 Jeffrey Xu Yu 2 Rui Zhou 1 1 Swinburne University of Technology.
Chapter 5: Introduction to Information Retrieval
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
How to Create an MLA citation for a web document....
Query Chains: Learning to Rank from Implicit Feedback Paper Authors: Filip Radlinski Thorsten Joachims Presented By: Steven Carr.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Evaluating Search Engine
 How many pages does it search?  How does it access all those pages?  How does it give us an answer so quickly?  How does it give us such accurate.
Project Title: Deepin Search Member: Wenxu Li & Ziming Zhai CSCI 572 Project.
Part 4: Evaluation Days 25, 27, 29, 31 Chapter 20: Why evaluate? Chapter 21: Deciding on what to evaluate: the strategy Chapter 22: Planning who, what,
A machine learning approach to improve precision for navigational queries in a Web information retrieval system Reiner Kraft
Measuring the quality of academic library electronic services and resources Jillian R Griffiths Research Associate CERLIM – Centre for Research in Library.
INFO 624 Week 3 Retrieval System Evaluation
Retrieval Evaluation. Brief Review Evaluation of implementations in computer science often is in terms of time and space complexity. With large document.
Part 4: Evaluation Chapter 20: Why evaluate? Chapter 21: Deciding on what to evaluate: the strategy Chapter 22: Planning who, what, where, and when Chapter.
Senior Project Database: Design and Usability Evaluation Stephanie Cheng Rachelle Hom Ronald Mg Hoang Bao CSC 484 – Winter 2005.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
Topic-Sensitive PageRank Taher H. Haveliwala. PageRank Importance is propagated A global ranking vector is pre-computed.
Web Archive Information Retrieval Miguel Costa, Daniel Gomes (speaker) Portuguese Web Archive.
How to Assess Student Learning in Arts Partnerships Part II: Survey Research Revised April 2, 2012 Mary Campbell-Zopf, Ohio Arts Council
Search is not only about the Web An Overview on Printed Documents Search and Patent Search Walid Magdy Centre for Next Generation Localisation School of.
Quality-aware Collaborative Question Answering: Methods and Evaluation Maggy Anastasia Suryanto, Ee-Peng Lim Singapore Management University Aixin Sun.
Databases & Data Warehouses Chapter 3 Database Processing.
1 SKODA-AUTO.CZ prototype evaluation Poznań, 23th of March 2015.
Evaluation IMD07101: Introduction to Human Computer Interaction Brian Davison 2010/11.
Some Internet Videos May be More Appropriate For Teaching Than Others Jeffrey Bell 1,3 and Jim Bidlack 2,3 Abstract IntroductionMaterials & Methods Results.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
An Analysis of Assessor Behavior in Crowdsourced Preference Judgments Dongqing Zhu and Ben Carterette University of Delaware.
1 Can People Collaborate to Improve the relevance of Search Results? Florian Eiteljörge June 11, 2013Florian Eiteljörge.
Nielsen’s Ten Usability Heuristics
Usability Evaluation/LP Usability: how to judge it.
UOS 1 Ontology Based Personalized Search Zhang Tao The University of Seoul.
CIKM’09 Date:2010/8/24 Advisor: Dr. Koh, Jia-Ling Speaker: Lin, Yi-Jhen 1.
Evaluation INST 734 Module 5 Doug Oard. Agenda  Evaluation fundamentals Test collections: evaluating sets Test collections: evaluating rankings Interleaving.
Quantitative Comparisons of Search Engine Results Mike Thlwall School of Computing and Information Technology, University of Wolverhampton ( 伍爾弗漢普頓 UK)
Giorgos Giannopoulos (IMIS/”Athena” R.C and NTU Athens, Greece) Theodore Dalamagas (IMIS/”Athena” R.C., Greece) Timos Sellis (IMIS/”Athena” R.C and NTU.
Crawling and Aligning Scholarly Presentations and Documents from the Web By SARAVANAN.S 09/09/2011 Under the guidance of A/P Min-Yen Kan 10/23/
The Anatomy of a Large-Scale Hyper textual Web Search Engine S. Brin, L. Page Presenter :- Abhishek Taneja.
Best practice in design on NREN websites A heuristic evaluation made for TERENA TF-PR By Julia Gardner & Gitte Kudsk, UNIC
Search Pages and Results LIS 385E: Information Architecture and Design By: Alex Chung
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Retroactive Answering of Search Queries Beverly Yang Glen Jeh.
Measuring How Good Your Search Engine Is. *. Information System Evaluation l Before 1993 evaluations were done using a few small, well-known corpora of.
Performance Measures. Why to Conduct Performance Evaluation? 2 n Evaluation is the key to building effective & efficient IR (information retrieval) systems.
Heuristic Evaluation Short tutorial to heuristic evaluation
A Novel Visualization Model for Web Search Results Nguyen T, and Zhang J IEEE Transactions on Visualization and Computer Graphics PAWS Meeting Presented.
Advantages of Query Biased Summaries in Information Retrieval by A. Tombros and M. Sanderson Presenters: Omer Erdil Albayrak Bilge Koroglu.
1 IFLA FRBR & MIC Metadata Evaluation Ying Zhang Yuelin Li October 14, 2003.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
Chapter. 3: Retrieval Evaluation 1/2/2016Dr. Almetwally Mostafa 1.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Market Research Market Research – What is it? Gathering information on (for example)… …finding new markets …evaluating public opinion …gaining insight.
1 Personalizing Search via Automated Analysis of Interests and Activities Jaime Teevan, MIT Susan T. Dumais, Microsoft Eric Horvitz, Microsoft SIGIR 2005.
Ten Usability Heuristics These are ten general principles for user interface design. They are called "heuristics" because they are more in the nature of.
Opinion spam and Analysis 소프트웨어공학 연구실 G 최효린 1 / 35.
Search Engine Optimization
Evaluation Anisio Lacerda.
How to use By Zainab Muman
Heuristic Evaluation Jon Kolko Professor, Austin Center for Design.
Q4 Measuring Effectiveness
10 Design Principles.
Combining Keyword and Semantic Search for Best Effort Information Retrieval  Andrew Zitzelberger 1.
ภาควิชาวิทยาการคอมพิวเตอร์ คณะวิทยาศาสตร์
Nilesen 10 hueristics.
Presentation transcript:

User-centered Information System Evaluation Tutorial

Goal Understand how to do user-centered information system evaluation Using query performance of different queries generated by users to measure the efficiency of different information retrieval systems. Evaluating systems using precision and recall based measures

Example Scenario: Finding expert is important, especially when you want to find potential collaborators, ask for consultancy, look for potential reviewers, etc. Aim: comparing two expert finding systems Pair 1: http://vivo.cornell.edu vs. http://experts.umich.edu Pair 2: http://vivosearch.org vs. http://www.arnetminer.org Pair 3: Google vs. Yahoo Goal: which one is a better expert finding system based on the following criteria easy to use, user-friendly quality of search

Quantitative data analysis

How Approach (2 persons form a group) Demographic questionnaire About your user, whether he/she is familiar with this system Pre-evaluation questionnaire whether you are familiar with the topic Evaluation Easy to use, user-friendly Develop a survey/questionnaire with 3-5 5-scale questions and 1-2 open-end questions E.g., easy to find help file, easy to know where you are, focusing on the design of the system Quality of search (2 queries) Pair 1: Q1: find experts working on “stem cell”, Q2: create another query Pair 2: Q1: find experts good at “data mining”, Q2: create another query Pair 3: Q1: create your own query, Q2: create another query Post-evaluation questionnaire Usability Whether the system is easy to search, does it take too long to get the result, whether the results are ranked, whether each ranked URL has short summary Exit questionnaire preference

Measuring quality of search Query System Rank Page title Relevance Judgment (0/1) Q1: ??????? 1 2 3 4 5 Same for Query 2

User judgment of relevance Fill in the following table with the title and short summary of top 5 results Show the results to your partner and let your partner decide the relevance (0=irrelevant, 1=relevant) Rank Page title + short summary (random order) Relevance Judgment (0/1) 1 2 3 4 5

Calculate Precision@topN Query System ID Evaluation Metric Q1 System 1 Precision@3 = Precision@5 = System 2 Q2 System ID Evaluation Metric System 1 Precision@3 = Precision@5 = System 2 Precision@Top3=(1+0+1)/3=0.67 Precision@Top5=(1+0+1+0+1)/5=0.6 Rank Relevance Judgment (0/1) 1 2 3 4 5 Precision@Top3_S1= (Precision@Top3_Q1+ Precision@Top3_Q2)/2

Compare your results with your partner’s Will your results be consistent with your partner’s? Compare the precision@top3 and precision@top5 If your results are consistent, you can draw the conclusion, which system performs better If your results are not consistent, what can you draw, why? Limitation and potential reasons.

Evaluation Report Write a short report (1-2 pages) about this evaluation Including the evaluation outputs Quantitative analysis Qualitative analysis Final conclusions Limitation Lessons-Learned If you want to have this as a real evaluation, what can you improve.