© Anselm Spoerri Lecture 13 Housekeeping –Term Projects Evaluations –Morse, E., Lewis, M., and Olsen, K. (2002) Testing Visual Information Retrieval Methodologies.

Slides:



Advertisements
Similar presentations
Recuperação de Informação B Cap. 10: User Interfaces and Visualization 10.1,10.2,10.3 November 17, 1999.
Advertisements

1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Search Engines Information Retrieval in Practice All slides ©Addison Wesley, 2008.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
GENERATING AUTOMATIC SEMANTIC ANNOTATIONS FOR RESEARCH DATASETS AYUSH SINGHAL AND JAIDEEP SRIVASTAVA CS DEPT., UNIVERSITY OF MINNESOTA, MN, USA.
To See, or Not to See—Is That the Query? Robert R. Korfhage Dept. of Information Science University of Pittsburgh 1991 Reviewed by Yi-Bu Chen LIS 551 Information.
1 Entity Ranking Using Wikipedia as a Pivot (CIKM 10’) Rianne Kaptein, Pavel Serdyukov, Arjen de Vries, Jaap Kamps 2010/12/14 Yu-wen,Hsu.
Explorations in Tag Suggestion and Query Expansion Jian Wang and Brian D. Davison Lehigh University, USA SSM 2008 (Workshop on Search in Social Media)
Information Retrieval Visualization CPSC 533c Class Presentation Qixing Zheng March 22, 2004.
Evaluating Search Engine
Search Engines and Information Retrieval
© Anselm Spoerri Lecture 10 Visual Tools for Text Retrieval (cont.)
Mobile Web Search Personalization Kapil Goenka. Outline Introduction & Background Methodology Evaluation Future Work Conclusion.
Learning to Advertise. Introduction Advertising on the Internet = $$$ –Especially search advertising and web page advertising Problem: –Selecting ads.
1 CS 430 / INFO 430 Information Retrieval Lecture 24 Usability 2.
Affinity Rank Yi Liu, Benyu Zhang, Zheng Chen MSRA.
Scalable Text Mining with Sparse Generative Models
Web Search – Summer Term 2006 II. Information Retrieval (Basics Cont.) (c) Wolfgang Hürst, Albert-Ludwigs-University.
Xiaomeng Su & Jon Atle Gulla Dept. of Computer and Information Science Norwegian University of Science and Technology Trondheim Norway June 2004 Semantic.
Databases & Data Warehouses Chapter 3 Database Processing.
IT Introduction to Website Development Welcome!
Search Engines and Information Retrieval Chapter 1.
©2008 Srikanth Kallurkar, Quantum Leap Innovations, Inc. All rights reserved. Apollo – Automated Content Management System Srikanth Kallurkar Quantum Leap.
1 Formal Models for Expert Finding on DBLP Bibliography Data Presented by: Hongbo Deng Co-worked with: Irwin King and Michael R. Lyu Department of Computer.
Redeeming Relevance for Subject Search in Citation Indexes Shannon Bradshaw The University of Iowa
Xiaoying Gao Computer Science Victoria University of Wellington Intelligent Agents COMP 423.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
Proposal for Term Project J. H. Wang Mar. 2, 2015.
Probabilistic Query Expansion Using Query Logs Hang Cui Tianjin University, China Ji-Rong Wen Microsoft Research Asia, China Jian-Yun Nie University of.
Distributed Information Retrieval Server Ranking for Distributed Text Retrieval Systems on the Internet B. Yuwono and D. Lee Siemens TREC-4 Report: Further.
Web Image Retrieval Re-Ranking with Relevance Model Wei-Hao Lin, Rong Jin, Alexander Hauptmann Language Technologies Institute School of Computer Science.
University of Malta CSA3080: Lecture 6 © Chris Staff 1 of 20 CSA3080: Adaptive Hypertext Systems I Dr. Christopher Staff Department.
The Anatomy of a Large-Scale Hyper textual Web Search Engine S. Brin, L. Page Presenter :- Abhishek Taneja.
Enhancing Cluster Labeling Using Wikipedia David Carmel, Haggai Roitman, Naama Zwerdling IBM Research Lab (SIGIR’09) Date: 11/09/2009 Speaker: Cho, Chin.
21/11/20151Gianluca Demartini Ranking Clusters for Web Search Gianluca Demartini Paul–Alexandru Chirita Ingo Brunkhorst Wolfgang Nejdl L3S Info Lunch Hannover,
1 CS 430: Information Discovery Lecture 19 User Interfaces.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
Intelligent Database Systems Lab N.Y.U.S.T. I. M. Externally growing self-organizing maps and its application to database visualization and exploration.
Adish Singla, Microsoft Bing Ryen W. White, Microsoft Research Jeff Huang, University of Washington.
Automatic Video Tagging using Content Redundancy Stefan Siersdorfer 1, Jose San Pedro 2, Mark Sanderson 2 1 L3S Research Center, Germany 2 University of.
A Novel Visualization Model for Web Search Results Nguyen T, and Zhang J IEEE Transactions on Visualization and Computer Graphics PAWS Meeting Presented.
Relevance-Based Language Models Victor Lavrenko and W.Bruce Croft Department of Computer Science University of Massachusetts, Amherst, MA SIGIR 2001.
L&I SCI 110: Information science and information theory Instructor: Xiangming(Simon) Mu Sept. 9, 2004.
TREC-2003 (CDVP TRECVID 2003 Team)- 1 - Center for Digital Video Processing C e n t e r f o r D I g I t a l V I d e o P r o c e s s I n g CDVP & TRECVID-2003.
© Anselm Spoerri Lecture 10 Search Visualization –“Search User Interfaces” by Hearst –Visual Query Formulation –Search Result Visualization –InfoCrystal.
ApproxHadoop Bringing Approximations to MapReduce Frameworks
1 What Makes a Query Difficult? David Carmel, Elad YomTov, Adam Darlow, Dan Pelleg IBM Haifa Research Labs SIGIR 2006.
1 CS 430: Information Discovery Lecture 14 Usability I.
The Loquacious ( 愛說話 ) User: A Document-Independent Source of Terms for Query Expansion Diane Kelly et al. University of North Carolina at Chapel Hill.
A code-centric cluster-based approach for searching online support forums for programmers Christopher Scaffidi, Christopher Chambers, Sheela Surisetty.
Citation-Based Retrieval for Scholarly Publications 指導教授:郭建明 學生:蘇文正 M
DISTRIBUTED INFORMATION RETRIEVAL Lee Won Hee.
1 CS 430: Information Discovery Lecture 5 Ranking.
CS798: Information Retrieval Charlie Clarke Information retrieval is concerned with representing, searching, and manipulating.
User-Friendly Systems Instead of User-Friendly Front-Ends Present user interfaces are not accepted because the underlying systems are too difficult to.
Learning to Rank: From Pairwise Approach to Listwise Approach Authors: Zhe Cao, Tao Qin, Tie-Yan Liu, Ming-Feng Tsai, and Hang Li Presenter: Davidson Date:
Toward Entity Retrieval over Structured and Text Data Mayssam Sayyadian, Azadeh Shakery, AnHai Doan, ChengXiang Zhai Department of Computer Science University.
Navigation Aided Retrieval Shashank Pandit & Christopher Olston Carnegie Mellon & Yahoo.
PSY 325 AID Education Expert/psy325aid.com FOR MORE CLASSES VISIT
Text Similarity: an Alternative Way to Search MEDLINE James Lewis, Stephan Ossowski, Justin Hicks, Mounir Errami and Harold R. Garner Translational Research.
Using Blog Properties to Improve Retrieval Gilad Mishne (ICWSM 2007)
1 CS 430 / INFO 430: Information Retrieval Lecture 20 Web Search 2.
University Of Seoul Ubiquitous Sensor Network Lab Query Dependent Pseudo-Relevance Feedback based on Wikipedia 전자전기컴퓨터공학 부 USN 연구실 G
Proposal for Term Project
Evaluation of IR Systems
HI 5354 – Cognitive Engineering
Data Mining: Concepts and Techniques Course Outline
CSc4730/6730 Scientific Visualization
Disambiguation Algorithm for People Search on the Web
Lab 2: Information Retrieval
Presentation transcript:

© Anselm Spoerri Lecture 13 Housekeeping –Term Projects Evaluations –Morse, E., Lewis, M., and Olsen, K. (2002) Testing Visual Information Retrieval Methodologies Case Study: Comparative Analysis of Textual, Icon Graphical and 'Spring' Displays Journal of the American Society for Information Science and Technology (JASIST) PDFTesting Visual Information Retrieval Methodologies Case Study: Comparative Analysis of Textual, Icon Graphical and 'Spring' DisplaysPDF –Reiterer H., Mußler G., Mann T.: Visual Information Retrieval for the WWW, in: Smith M.J. et al. (eds.), Usability Evaluation and Interface Design, Lawrence Erlbaum, 2001 PDFPDF –searchCrystal Studies

© Anselm Spoerri Prototype Project –Motivate domain choice. –Perform task and need analysis. –Describe design approach and information visualization principles used. –Develop prototype. –Have an "domain expert" use the prototype and provide feedback. Class Presentation You have 15 min. to describe task analysis and your design approach. Demonstrate your prototype. Report on the "domain expert" feedback. Create Report 20 to 25 pages, written as a standard paper  10pt, double-spaced Provide screenshots of prototype and explain design approach. Include URL of prototype. Hand-in Hardcopy of report. Post report online and send instructor an with the URL.

© Anselm Spoerri Text Retrieval Visualizations – Evaluations : Morse et al. Many Tools Proposed Few Tested and Often Inconclusive / Fare Poorly Simplify Evaluation  Focus on Method (instead of implementation)  Only Static Aspects POI = Point of Interest Visualizations –Position Coding Glyph = Graphical Entity –Conveys data values via attributes such as shape, size, color

© Anselm Spoerri Glyph = Graphical Entity

© Anselm Spoerri Evaluation – Morse et al.

© Anselm Spoerri Evaluation – Morse et al. : Two-Term Boolean Test

© Anselm Spoerri Evaluation – Morse et al. : Two-Term Boolean Test

© Anselm Spoerri Evaluation – Morse et al. : Three-Term Boolean Test

© Anselm Spoerri Evaluation – Morse et al. : Vector Studies – Text List

© Anselm Spoerri Evaluation – Morse et al. : Vector Studies – Table

© Anselm Spoerri Evaluation – Morse et al. : Vector Studies – Icons

© Anselm Spoerri Evaluation – Morse et al. : Vector Studies – VIBE

© Anselm Spoerri Evaluation – Morse et al. : Vector Studies Time

© Anselm Spoerri Evaluation – Reiterer et al.

© Anselm Spoerri Evaluation – Reiterer et al.

© Anselm Spoerri Evaluation – Reiterer et al.

© Anselm Spoerri Evaluation – Reiterer et al.

© Anselm Spoerri Evaluation – Reiterer et al.

© Anselm Spoerri searchCrystal – Studies Validate Design Approach How does Overlap between Results Actually Correlate with Relevance? User Study

© Anselm Spoerri Overlap between Search Results Correlated with Relevance? Method –Use Ad-hoc track data for TREC 3, 6, 7, 8 –Systems search the SAME Database –Automatic Short Runs –50 Topics and 1,000 Documents per topic  50,000 documents –Retrieval systems can submit multiple runs  Select Best Run based Mean Average Precision TREC 319 systems 928,709 documents found TREC 624 systems 1,192,557 documents found TREC 728 systems 1,327,166 documents found TREC 835 systems 1,723,929 documents found –Compute Average by summing over all 50 topics and divide by 50

© Anselm Spoerri How does Overlap Correlate with Relevance?  Authority Effect Percentage of Documents that are Relevant Systems

© Anselm Spoerri TREC 8 – Impact of Average Rank Position?  Ranking Effect Systems Percentage of Documents that are Relevant Compute overlap structure between top 50 search results of 35 random groupings of 5 retrieval systems for 50 topics.

© Anselm Spoerri searchCrystal – Studies How does Overlap between Search Results Correlate with Relevance? Authority Effect – the more systems that find a document, the greater the probability that it is relevant Ranking Effect – the higher up a document in a ranked list and the more systems that find it, the greater the probability of its relevance  Validates searchCrystal’s Design Approach  searchCrystal Visualizes Authority & Ranking Effects  searchCrystal can Guide User’s Exploration Toward Relevant Documents

© Anselm Spoerri searchCrystal – Studies Validate Design Approach How does Overlap between Results Actually Correlate with Relevance? User Study

© Anselm Spoerri User Study – Cluster Bulls-Eye

© Anselm Spoerri User Study – RankSpiral

© Anselm Spoerri User Study – Compare Cluster Bull’s Eye and RankSpiral Nine undergraduates. Short Introduction and No Training. Randomized presentation order of data sets and display type. Subject selects ten document; Visual feedback about correct top 10 Test for Cluster Bull’s Eye and RankSpiral displays: 1) How well can novices use visual cues to find the documents that are most likely to be relevant? 2) Performance difference in terms of effectiveness and/or efficiency? 3) How much document’s distance from the display center will interfere with the size coding used to encode its probability of being relevant

© Anselm Spoerri User Study – Results Hypothesis 1: “Novices can perform the task.” Error is minimal for the top 7 documents and increases rapidly after the top 7 documents for both displays. Novice users can use the Cluster Bulls-Eye and RankSpiral displays to select highly relevant documents, especially the top 7 documents. Hypothesis 2: “RankSpiral outperforms Cluster Bulls-Eye.” 8 of the 9 subjects performed the task faster using the RankSpiral. Average time difference was 7.89 seconds. The one-sided T-test value is 0.033, which is significant at the 0.05 level. 7 out of 9 subjects performed the task more effectively using the RankSpiral. Average “relevance score” difference is The one-sided T-test value is 0.037, which is significant at the 0.05 level. Hypothesis 3: “Distance from center dominant cue.”

© Anselm Spoerri Discussion Relax searchCrystal’s design principles? –Mapping documents found by the same number of engines into the same concentric ring. Option: Distance and Size encode likelihood that a document is relevant. Internet search results: –Concentric rings are of value, because it is much harder to estimate a document’s probability of being relevant.

© Anselm Spoerri Cluster Bulls-Eye  Size = Distance from Center

© Anselm Spoerri Cluster Bulls-Eye  Size = Distance from Center

© Anselm Spoerri searchCrystal - Studies Authority & Ranking Effects Comparing Results of All Retrieval Systems at once Comparing Results of Random Subsets of Five Systems  Validating searchCrystal’s Design Principles User Study Identify Top 10 Docs in Cluster Bull’s Eye and RankSpiral  Novice Users can use the two searchCrystal displays  Statistical Difference between two displays  Distance from center is dominant visual feature

© Anselm Spoerri What is Popular on Wikipedia? Why? Please read the two papers published by me in First Monday: Approach 1 Visualize Popular Wikipedia Pages Overlap between 100 Most Visited Pages on Wikipedia for September 2006 to January 2007 Information Visualization helps to gain quick insights 2 Categorize Popular Wikipedia Pages 3 Examine Popular Search Queries 4 Determine Search Result Position of Popular Wikipedia pages 5 Implications