We are Gathered Here Together: Why? Introduction to the NII Workshop on Whole-Session Evaluation of Interactive Information Retrieval Systems Whole-Session.

Slides:



Advertisements
Similar presentations
SUGGESTIONS ON A DRAFT FRAMEWORK OUTLINE Presented to the Ad Hoc Inter-governmental Group on Earth Observations Gregory W. Withee U.S. Delegation August.
Advertisements

Orientation for New Site Visitors CIDA’s Mission, Value, and the Guiding Principles of Peer Review.
Copyright © The McGraw-Hill Companies, Inc. Permission required for reproduction or display. Alternative Assessments FOUN 3100 Fall 2003 Sondra M. Parmer.
© 2005 Prentice Hall6-1 Stumpf and Teague Object-Oriented Systems Analysis and Design with UML.
February 9, 2012 Session 1: Observing Lessons NYSED Principal Evaluation Training Program.
Canadian Geomatics Community Round Table Know Edge Limited Location Information Innovation Overview of Workshop Presented by: Ed Kennedy Hickling Arthurs.
Understanding community education and engagement FACTSHEET 4 Food and Garden Organics Best Practice Collection Manual.
Knowledge Acquisitioning. Definition The transfer and transformation of potential problem solving expertise from some knowledge source to a program.
User Requirements Gathering for the Humanities: How do we establish best practice for the community? Ruth Kirkham – Project Manager John Pybus – Technical.
COST G9 - Work group 2 Cadastral science meeting Aalborg, Dk Modeling methodology for real estate transactions Radoš Šumrada Faculty.
Content Framework for the Committee Report Using PowerPoint to prepare a formal committee report John A. Cagle.
Evaluating and Revising the Physical Education Instructional Program.
© Tefko Saracevic, Rutgers University1 digital libraries and human information behavior Tefko Saracevic, Ph.D. School of Communication, Information and.
PPA 503 – The Public Policy Making Process
Strategic Planning Process Overview Strategic Planning Process Overview PLC Strategic Plan Agenda Welcome and Introductions 3:00-3:15.
Orientation to the Social Studies K to 7 Integrated Resource Package 2006.
Learning and Development Developing leaders and managers
How to Develop the Right Research Questions for Program Evaluation
Hollis Day, MD, MS Susan Meyer, PhD.  Four domains for effective practice outlined in the Interprofessional Education Collaborative’s “Core Competencies.
Developing an Effective Evaluation to Check for Understanding Susan E. Schultz, Ph.D. Evaluation Consultant PARK Teachers.
 A set of objectives or student learning outcomes for a course or a set of courses.  Specifies the set of concepts and skills that the student must.
Teamwork Chapter 6.
Technical Report Writing
Instructional Design Eyad Hakami. Instructional Design Instructional design is a systematic process by which educational materials are created, developed,
By: Becky Guzie Chapter 5: Developing Adaptations to Promote Participation in Inclusive Environment.
Designing Real Community Partnerships That Work Maureen Rubin California State University, Northridge Innovative Educators Webinar November 4, 2009.
Definitions Learning contracts are agreements between teachers and learners. A learning contract is a written document that draws the trainer and trainee.
First 5 Kings County Evaluation Services Request for Application Bidders Teleconference – February 22, 2012 Evaluation Services Request for Application.
Evaluation Experiments and Experience from the Perspective of Interactive Information Retrieval Ross Wilkinson Mingfang Wu ICT Centre CSIRO, Australia.
Integrating Evaluation into the Design of the Minnesota Demonstration Project Paint Product Stewardship Initiative St. Paul, MN May 1, 2008 Matt Keene,
Southeast Asia Regional Senior Leadership Program Day 3 Phnom Penh, Cambodia June 8-12th, 2015.
Unit 3. demonstrate the ability to use subprograms within computer programs; use a variety of problem-solving strategies to solve different types of.
Using Business Scenarios for Active Loss Prevention Terry Blevins t
McGraw-Hill/Irwin Teaching Excellence Project funded by CELT Teaching Economics through Innovative Content and Effective Teaching Methods Necati Aydin,
INTERNATIONAL SOCIETY FOR TECHNOLOGY IN EDUCATION working together to improve education with technology Using Evidence for Educational Technology Success.
DRAFT – For Discussion Only HHSC IT Governance Executive Briefing Materials DRAFT April 2013.
Regional Seminar 2005 EVALUATING POLICY Are your policies working? How do you know? School Development Planning Initiative.
Implicit Acquisition of Context for Personalization of Information Retrieval Systems Chang Liu, Nicholas J. Belkin School of Communication and Information.
Ben Carterette Paul Clough Evangelos Kanoulas Mark Sanderson.
Recuperação de Informação B Cap. 10: User Interfaces and Visualization , , 10.9 November 29, 1999.
Making Plans for the Future April 29, 2013 Brenda M. Tanner, Ed.D.
Learning, Engagement, and Responsiveness Faculty of Business Workshop.
Designing a Training Program RATIONALE OF THE TRAINING Background or introduction of what the training is all about –Developments in the field/discipline/area.
Requirement engineering Good Practices for Requirements Engineering
October 2004 Strategy Forum Academic Quality Improvement Program The Higher Learning Commission of the North Central Association of Colleges and Schools.
AN INTEGRATED APPROACH TO ENGINEERING AND DESIGN School of Mechanical and Aeronautical Engineering (SMAE) Dr Linda Lee, Peter Lo and Lim Siew Kuan June.
Jane Reid, AMSc IRIC, QMUL, 30/10/01 1 Information seeking Information-seeking models Search strategies Search tactics.
National Public Health Performance Standards Local Assessment Instrument Essential Service:6 Enforce Laws and Regulations that Protect Health and Ensure.
This was developed as part of the Scottish Government’s Better Community Engagement Programme.
Why IR test collections are so bad Mark Sanderson University of Sheffield.
Project Milestones & project budgeting Morning session.
Interface Types and Models Dr. Dania Bilal IS 588 Spring 2008.
Leadership Skills. Team Meetings Set the agenda by defining goals and desired outcomes Set the agenda by defining goals and desired outcomes Keep the.
Developing Monitoring & Evaluation Frameworks: Process or Product? Anne Markiewicz.
Building a Framework to Support the Culture Required for Student Centered Learning DeeDee Washington, Associate Superintendent of Academics Elementary.
How to Lead When You are Not the Boss. Introduction Real leadership is never a matter of mere formal authority.
Stage 1 Integrated learning Coffee Shop. LEARNING REQUIREMENTS The learning requirements summarise the knowledge, skills, and understanding that students.
Session 2: Developing a Comprehensive M&E Work Plan.
Utilizing Research: Putting Research Evidence Into Nursing Practice Prepare by /Dr. AmiraYahia.
Technical Report Writing Dr. Shelley Thomas. Overview Selecting effective report topics Using worksheets to plan projects Developing and proposing project.
[Presentation location] [Presentation date] (Confirm ABT logo) Building Bridges and Bonds (B3): An introduction.
Training processes for extension education
Some(what) Grand Challenges for Information Retrieval
Scientific Research in Computing
Exploratory Search Beyond the Query–Response Paradigm
Families, Workplaces and Communities
Families, Workplaces and Communities
Program Design in Population, Health and Environment
Effective Meeting.
Presentation transcript:

We are Gathered Here Together: Why? Introduction to the NII Workshop on Whole-Session Evaluation of Interactive Information Retrieval Systems Whole-Session Evaluation of Interactive Information Retrieval Systems Nicholas J. Belkin School of Communication & Information Rutgers University

Who are “We”? IR researchers concerned with proper evaluation of the performance of IR systems IR researchers dissatisfied with current criteria, measures, and methods of evaluation of the performance of IR systems IR researchers who are particularly concerned with user-centered evaluation of the performance of IR systems

What’s Meant by “User-Centered” Evaluation? Understanding the goals/tasks that lead people to engage with IR systems Understanding the behaviors that people engage in during information seeking Devising criteria and measures of performance that correspond to those goals/tasks, and to those behaviors Developing methods for applying those criteria and measures

What is Meant by “Whole-Session Evaluation”? Understanding that the information-seeking activities related to motivating goals/tasks may not be limited to IR system responses to single queries, but may require support over a number of different activities during a search session, or over multiple search sessions. Developing criteria, measures and methods which are appropriate to evaluating IR system support with respect to the process of a search session, the outcome of a search session, and the outcomes of multiple search sessions

What can “We” do Here to Address Whole Session Evaluation Complain about other evaluation criteria, measures and methods – Not a starter Identify the most significant problems in accomplishing whole session evaluation Propose some means for addressing these problems Suggest a(some) framework(s) for accomplishing whole session evaluation

Possible Goals for the Workshop A document outlining: – Need for whole-session evaluation; – Problems facing whole-session evaluation – Potential approaches to those problems, with pluses, minuses and contexts of application – Proposals of research directions in implementing whole-session evaluation A set of possible frameworks in which whole-session evaluation could take place A proposal to TREC (or some other evaluation forum) for a Whole-Session Evaluation Track in 2013 Establishment of a research community to share data and experiences

Ways to Accomplish these Goals Draw on the expertise of the group: sharing and comparing your homework Identify groups of participants sharing similar concerns/solutions wrt whole-session evaluation as “break-out groups” Reports from groups as the basis for debate on proposals New groups formed as result of debate, iterate Small groups assigned to produce the products (and their components) of the workshop

Some Things We’re Likely to Argue About The value and forms of formal models of search What constitutes a “search session” How to relate system support to task outcomes, or should we even bother Simulation of search sessions, perhaps versus study of “real” goals, tasks, behaviors Comparative evaluation, a la TREC If “test collections” are thought to be desirable or necessary, what form(s) should they take

Some Things we should Agree On We’ll be (at least moderately) nice to one another We’re here to collaborate and create, not to pontificate We will devote significant time to breaks from formal work We will enjoy Shonan Village, its facilities, foods, and environment We will do our best to achieve at least drafts of our goals by Friday afternoon