Text REtrieval Conference (TREC) The TREC Conferences Ellen Voorhees.

Slides:



Advertisements
Similar presentations
Metadata in Carrot II Current metadata –TF.IDF for both documents and collections –Full-text index –Metadata are transferred between different nodes Potential.
Advertisements

Text Analysis Conference Knowledge Base Population 2013 Hoa Trang Dang National Institute of Standards and Technology Sponsored by:
Search Results Need to be Diverse Mark Sanderson University of Sheffield.
Overview of Collaborative Information Retrieval (CIR) at FIRE 2012 Debasis Ganguly, Johannes Leveling, Gareth Jones School of Computing, CNGL, Dublin City.
Presenters: Başak Çakar Şadiye Kaptanoğlu.  Typical output of an IR system – static predefined summary ◦ Title ◦ First few sentences  Not a clear view.
Search Engines and Information Retrieval
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Intelligent Information Retrieval CS 336 Lisa Ballesteros Spring 2006.
Amanda Spink : Analysis of Web Searching and Retrieval Larry Reeve INFO861 - Topics in Information Science Dr. McCain - Winter 2004.
Modern Information Retrieval
Information Retrieval in Practice
INFO 624 Week 3 Retrieval System Evaluation
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Reference Collections: Task Characteristics. TREC Collection Text REtrieval Conference (TREC) –sponsored by NIST and DARPA (1992-?) Comparing approaches.
1 CS 430: Information Discovery Lecture 20 The User in the Loop.
An Overview of Relevance Feedback, by Priyesh Sudra 1 An Overview of Relevance Feedback PRIYESH SUDRA.
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
Enhance legal retrieval applications with an automatically induced knowledge base Ka Kan Lo.
Evaluation Information retrieval Web. Purposes of Evaluation System Performance Evaluation efficiency of data structures and methods operational profile.
 Official Site: facility.org/research/evaluation/clef-ip-10http:// facility.org/research/evaluation/clef-ip-10.
Automatic Term Mismatch Diagnosis for Selective Query Expansion Le Zhao and Jamie Callan Language Technologies Institute School of Computer Science Carnegie.
Search and Retrieval: Relevance and Evaluation Prof. Marti Hearst SIMS 202, Lecture 20.
Modern Retrieval Evaluations Hongning Wang
CS598CXZ Course Summary ChengXiang Zhai Department of Computer Science University of Illinois, Urbana-Champaign.
Challenges in Information Retrieval and Language Modeling Michael Shepherd Dalhousie University Halifax, NS Canada.
The Evolution of Shared-Task Evaluation Douglas W. Oard College of Information Studies and UMIACS University of Maryland, College Park, USA December 4,
Search Engines and Information Retrieval Chapter 1.
TREC 2009 Review Lanbo Zhang. 7 tracks Web track Relevance Feedback track (RF) Entity track Blog track Legal track Million Query track (MQ) Chemical IR.
ATLAS Outreach & Education News & Collaboration. News Reporting ATLAS progress and results to the world ATLAS Week - 11 Oct 2011S. Goldfarb - ATLAS Outreach.
Minimal Test Collections for Retrieval Evaluation B. Carterette, J. Allan, R. Sitaraman University of Massachusetts Amherst SIGIR2006.
TERM IMPACT- BASED WEB PAGE RAKING School of Electrical Engineering and Computer Science Falah Al-akashi and Diana Inkpen
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Philosophy of IR Evaluation Ellen Voorhees. NIST Evaluation: How well does system meet information need? System evaluation: how good are document rankings?
IR Evaluation Evaluate what? –user satisfaction on specific task –speed –presentation (interface) issue –etc. My focus today: –comparative performance.
THOMSON SCIENTIFIC Web of Science Using the specialized search and analyze features Jackie Stapleton, librarian Fall 2006.
Modern Information Retrieval: A Brief Overview By Amit Singhal Ranjan Dash.
2008 © ChengXiang Zhai Dragon Star Lecture at Beijing University, June 21-30, Frame an IR Research Problem and Form Hypotheses ChengXiang Zhai Department.
TRECVID Evaluations Mei-Chen Yeh 05/25/2010. Introduction Text REtrieval Conference (TREC) – Organized by National Institute of Standards (NIST) – Support.
Cross-Language Evaluation Forum (CLEF) IST Expected Kick-off Date: August 2001 Carol Peters IEI-CNR, Pisa, Italy Carol Peters: blabla Carol.
Ben Carterette Paul Clough Evangelos Kanoulas Mark Sanderson.
Sogang University A. I. Lab. Effective site finding using link anchor information Effective site finding using link anchor information Sung Hae, Jun Artificial.
1 01/10/09 1 INFILE CEA LIST ELDA Univ. Lille 3 - Geriico Overview of the INFILE track at CLEF 2009 multilingual INformation FILtering Evaluation.
Text REtrieval Conference (TREC) Implementing a Question-Answering Evaluation for AQUAINT Ellen M. Voorhees Donna Harman.
Introduction to Information Retrieval Example of information need in the context of the world wide web: “Find all documents containing information on computer.
Measuring How Good Your Search Engine Is. *. Information System Evaluation l Before 1993 evaluations were done using a few small, well-known corpora of.
 Used MapReduce algorithms to process a corpus of web pages and develop required index files  Inverted Index evaluated using TREC measures  Used Hadoop.
Comparing Document Segmentation for Passage Retrieval in Question Answering Jorg Tiedemann University of Groningen presented by: Moy’awiah Al-Shannaq
The Cross Language Image Retrieval Track: ImageCLEF Breakout session discussion.
1 13/05/07 1/20 LIST – DTSI – Interfaces, Cognitics and Virtual Reality Unit The INFILE project: a crosslingual filtering systems evaluation campaign Romaric.
Relevance Models and Answer Granularity for Question Answering W. Bruce Croft and James Allan CIIR University of Massachusetts, Amherst.
1 CS 430 / INFO 430 Information Retrieval Lecture 9 Evaluation of Retrieval Effectiveness 2.
Carnegie Mellon School of Computer Science Language Technologies Institute CMU Team-1 in TDT 2004 Workshop 1 CMU TEAM-A in TDT 2004 Topic Tracking Yiming.
Evaluation of Information Retrieval Systems Xiangming Mu.
Indri at TREC 2004: UMass Terabyte Track Overview Don Metzler University of Massachusetts, Amherst.
November 8, 2005NSF Expedition Workshop Supporting E-Discovery with Search Technology Douglas W. Oard College of Information Studies and Institute for.
UIC at TREC 2006: Blog Track Wei Zhang Clement Yu Department of Computer Science University of Illinois at Chicago.
Usefulness of Quality Click- through Data for Training Craig Macdonald, ladh Ounis Department of Computing Science University of Glasgow, Scotland, UK.
1 INFILE - INformation FILtering Evaluation Evaluation of adaptive filtering systems for business intelligence and technology watch Towards real use conditions.
Information Retrieval in Practice
Developments in Evaluation of Search Engines
Information Storage and Retrieval Fall Lecture 1: Introduction and History.
Evaluation Anisio Lacerda.
Information Retrieval (in Practice)
An Empirical Study of Learning to Rank for Entity Search
Implementation Issues & IR Systems
Text REtrieval Conference (TREC)
CSE 635 Multimedia Information Retrieval
Search Engine Architecture
Retrieval Evaluation - Reference Collections
Information Retrieval and Web Design
Presentation transcript:

Text REtrieval Conference (TREC) The TREC Conferences Ellen Voorhees

Text REtrieval Conference (TREC) TREC Philosophy TREC is a modern example of the Cranfield tradition –system evaluation based on test collections Emphasis on advancing the state of the art from evaluation results –TREC’s primary purpose is not competitive benchmarking –experimental workshop: sometimes experiments fail!

Text REtrieval Conference (TREC) Cranfield at Fifty Evaluation methodology is still valuable… carefully calibrated level of abstraction –has sufficient fidelity to real user tasks to be informative –general enough to be broadly applicable, feasible, relatively inexpensive …but is showing some signs of age size is overwhelming our ability to evaluate new abstractions need to carefully accommodate variability to maintain power

Text REtrieval Conference (TREC) Evaluation Difficulties Variability despite stark abstraction, user effect still dominates Cranfield results Size matters effective pooling has corpus size dependency test collection construction costs depend on number of judgments Model coarseness even slightly different tasks may not be good fit –e.g., legal discovery, video features

Text REtrieval Conference (TREC) TREC 2009 All tracks used some new, large document set Different trade-offs in adapting evaluation strategy tension between evaluating current participants’ ability to do the task and building reusable test collections variety of tasks that are not simple ranked-list retrieval

Text REtrieval Conference (TREC) ClueWeb09 Document Set Snapshot of the WWW in early 2009 crawled by CMU with support from NSF distributed through CMU used in four TREC 2009 tracks: Web, Relevance Feedback, Million Query, and Entity Full corpus about one billion pages and 25 terabytes of text about half is in English Category B English-only subset of about 50 million pages (including Wikipedia) to permit wider participation

Text REtrieval Conference (TREC) TREC 2009 Participants Applied DiscoveryLogik Systems, Inc.University of Applied Science Geneva Beijing Institute of TechnologyMicrosoft Research AsiaUniversity of Arkansas, Little Rock Beijing U. of Posts and Telecommunications Microsoft Research CambridgeUniversity of California, Santa Cruz Cairo Microsoft Innovation CenterMilwaukee School of EngineeringUniversity of Delaware (2) Carnegie Mellon UniversityMugla UniversityUniversity of Glasgow Chinese Academy of Sciences (2)National Institute of Information and Communications Technology University of Illinois, Urbana-Champaign Clearwell Systems, Inc.Northeastern UniversityUniversity of Iowa Clearly Gottlieb Steen & Hamilton, with Backstop LLC Open Text CorporationUniversity of Lugano Dalian University of TechnologyPeking UniversityUniversity of Maryland, College Park Delft University of TechnologyPohang U. of Science & TechnologyUniversity of Massachusetts, Amherst EMC - CMA - R&DPurdue UniversityThe University of Melbourne EquivioQueensland University of TechnologyUniversity of Padova Fondazione Ugo BordoniRMIT UniversityUniversity of Paris Fraunhofer SCAISabir ResearchUniversity of Pittsburgh Fudan UniversitySouth China University of TechnologyUniversity of Twente H5SUNY BuffaloUniversity of Waterloo (2) Heilongjiang Inst. of TechnologyTsinghua UniversityUrsinus College IntegreonUniversidade do PortoYahoo! Research International Inst. of Information Technology, Hyderabad University College DublinYork University (2) Know-CenterUniversity of Alaska, FairbanksZL Technologies, Inc. Lehigh UniversityUniversity of Amsterdam (2)

Text REtrieval Conference (TREC) The TREC Tracks Blog Spam Chemical IR Genomics Novelty QA, Entity Legal Enterprise Terabyte, Million Query Web VLC Video Speech OCR Cross-language Chinese Spanish Interactive, HARD, Feedback Filtering Routing Ad Hoc, Robust Personal documents Retrieval in a domain Answers, not documents Searching corporate repositories Size, efficiency, & web search Beyond text Beyond just English Human-in-the-loop Streamed text Static text

Text REtrieval Conference (TREC) TREC 2010 Blog, Chemical IR, Entity, Legal, Relevance Feedback, Web continuing Million Query merged with Web New “Sessions” track: investigate search behavior over a series of queries (series of length 2 for first running in 2010)

Text REtrieval Conference (TREC) TREC 2011 Track proposals due Monday (Sept 27) New track on searching free text fields of medical records likely