LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics.

Slides:



Advertisements
Similar presentations
Ed-D 420 Inclusion of Exceptional Learners. CAT time Learner-Centered - Learner-centered techniques focus on strategies and approaches to improve learning.
Advertisements

A Vehicle to Promote Student Learning
Progress Monitoring. Progress Monitoring Steps  Monitor the intervention’s progress as directed by individual student’s RtI plan  Establish a baseline.
Evaluation Mary Rowlatt MDR Partners. Definition of project evaluation Evaluation focuses on whether the project was effective, achieved its objectives,
Chapter 10 Teaching and Learning Strategies
Social Science Research and
Consistency of Assessment
Analytical methods for Information Systems Professionals
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
USABILITY AND EVALUATION Motivations and Methods.
introduction to MSc projects
Evaluation Methodologies
SWRK 171 Qualitative Research in Social Work. What is qualitative research?
Analytical methods for Information Systems Professionals Week 13 Lecture 1 CONCLUSION.
IB Diploma Program Exams – Semester Report Cards
Research Day 2009 Assessment of Student Work on Geographically Distributed Information Technology Project Teams Charles Tappert and Allen Stix Pace University,
Colorado Learning About Science Survey for Experimental Physics Benjamin Zwickl, Heather Lewandowski & Noah Finkelstein University of Colorado Physics.
Business Consulting Services Agenda Discussion: Management Reports Discussion: Project Reports Discussion: Engagement Proposal Upcoming Events Review Project.
Formulating the research design
Proposal in Detail – Part 2
HTA as a framework for task analysis Presenter: Hilary Ince, University of Idaho.
Science Inquiry Minds-on Hands-on.
Capstone Design Project (CDP) Civil Engineering Department First Semester 1431/1432 H 10/14/20091 King Saud University, Civil Engineering Department.
Impact of Including Authentic Inquiry Experiences in Methods Courses for Pre-Service Elementary and Secondary Teachers Timothy F. Slater, Lisa Elfring,
Choosing Your Primary Research Method What do you need to find out that your literature did not provide?
Assessment Activities
Aligning Course Competencies using Text Analytics
Welcome… The attendee will understand assessment basics with a focus on creating learning activities and identifying assessment expectations. Apply the.
Looking at Student work to Improve Learning
An Introduction to Research Methodology
Improving lecture capture through usability testing and research Ilkka #mlconf13 Media & Learning 2013, Brussels
SESSION ONE PERFORMANCE MANAGEMENT & APPRAISALS.
Dr. Engr. Sami ur Rahman Assistant Professor Department of Computer Science University of Malakand Research Methods in Computer Science Lecture: Research.
What research is Noun: The systematic investigation into and study of materials and sources in order to establish facts and reach new conclusions. Verb:
Research Methods in Computer Science Lecture: Quantitative and Qualitative Data Analysis | Department of Science | Interactive Graphics System.
Meta-Cognition, Motivation, and Affect PSY504 Spring term, 2011 January 13, 2010.
Science Fair Information. The purpose of the Science Fair is to offer students the opportunity to think deeply about science as it applies to everyday.
August 2007FFP Testing and Evaluation Techniques Chapter 7 Florida State Fire College Ocala, Florida.
Semester 2: Lecture 9 Analyzing Qualitative Data: Evaluation Research Prepared by: Dr. Lloyd Waller ©
What are your interactions doing for your visualization? Remco Chang UNC Charlotte Charlotte Visualization Center.
Protocols for Mathematics Performance Tasks PD Protocol: Preparing for the Performance Task Classroom Protocol: Scaffolding Performance Tasks PD Protocol:
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 7: Focusing on Users and Their Tasks.
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
Individual Preferences for Uncertainty: An Ironically Pleasurable Stimulus Bankert, M., VanNess, K., Hord, E., Pena, S., Keith, V., Urecki, C., & Buchholz,
Jay Summet CS 1 with Robots IPRE Evaluation – Data Collection Overview.
Astronomy 114 Lab Section 211, Professor Weigel. Outline for Today About Goals for this class Attendance Syllabus Safety Star Project Apparent vs. Absolute.
ScWk 242 Course Overview and Review of ScWk 240 Concepts ScWk 242 Session 1 Slides.
Qualitative Research January 19, Selecting A Topic Trying to be original while balancing need to be realistic—so you can master a reasonable amount.
Writing Software Documentation A Task-Oriented Approach Thomas T. Barker Chapter 5: Analyzing Your Users Summary Cornelius Farrell Emily Werschay February.
Understanding User Goals in Web Search University of Seoul Computer Science Database Lab. Min Mi-young.
(c) 2007 McGraw-Hill Higher Education. All rights reserved. Accountability and Teacher Evaluation Chapter 14.
Module 4 Week 1 Dr Carol Azumah Dennis University of Hull.
LECTURE 16: (EVEN MORE) OPEN QUESTIONS IN VISUAL ANALYTICS December 9, 2015 SDS 235 Visual Analytics.
CS 4620 Intelligent Systems. What we want to do today Course introductions Make sure you know the schedule for the next three weeks.
1 SEG3120 Analysis and Design for User Interfaces LAB1: Video tape evaluation.
Identifying Assessments
Please Hand In: -Chapter 2 Guided Exercise Questions -TIDB Cover Page and Member Report.
SWRK 3150 & 4120 Mid-term Evaluation. Welcome Please take some time to review these PowerPoint slides. They contain important information for students,
Science Fair Information. The purpose of the Science Fair is to offer students the opportunity to think deeply about science as it applies to everyday.
Research Problem The role of the instructor in online courses depends on course design. Traditional instructor responsibilities include class management,
Writing a Professional Development Plan.  Step 1–Identify Indicators to be Assessed  Step 2 –Determine Average Baseline Score  Step 3 –Develop a Growth.
LECTURE 02: EVALUATING MODELS January 27, 2016 SDS 293 Machine Learning.
Monitoring and evaluation Objectives of the Session  To Define Monitoring, impact assessment and Evaluation. (commonly know as M&E)  To know why Monitoring.
Associate Professor Cathy Gunn The University of Auckland, NZ Learning analytics down under.
Systems Analysis Lecture 5 Requirements Investigation and Analysis 1 BTEC HNC Systems Support Castle College 2007/8.
Usability & Evaluation in Visualizing Biological Data Chris North, Virginia Tech VizBi.
Writing your personal project report
Lecture 16: Evaluation Techniques
CSc4730/6730 Scientific Visualization
Course Introduction Data Visualization & Exploration – COMPSCI 590
Presentation transcript:

LECTURE 14: EVALUATION TECHNIQUES November 30, 2015 SDS 235 Visual Analytics

Announcements 1/2 Friday 4:30pm: Amanda Cox invited lecture – Ford 240 (extra credit for attending and posting to Piazza) Schedule change: - Evaluation Methods today - Self-Critique on Wednesday Final Project Demonstration will be held the evening of December 14 th, time TBA (regularly scheduled class will be cancelled)

Announcements 2/2 End of the semester is nearly upon us (T-minus 2 weeks!) Some of you might be starting to stress about grades However: if you would like a personalized PDF grade sheet of your progress in the course so far, me #ProTip

Discussion How do we measure the effectiveness of a visualization system?

Tufte, 1983 “Above all else, show the data.”

Tufte, 1983

Examples

Discussion What do you think of the Data-Ink Ratio? Consider ways to maximize the ratio…

Flashback: Epistemic Action The purpose of some actions is not the effect they have on the environment but the effect they have on the humans.

ChartJunk and Recall Bateman et al. “Useful Junk? The Effects of Visual Embellishment on Comprehension and Memorability of Charts”, CHI 2010

ChartJunk and Eye Gaze

Discussion Have you seen particularly compelling examples of “visual embellishment”? Tragic ones? What’s the right balance between Tufte and ChartJunk?

Nested Model of VIS Design (Munzner, 2009) Munzner, Tamara. "A nested model for visualization design and validation." Visualization and Computer Graphics, IEEE Transactions on 15.6 (2009):

Threats

Evaluation “Threats” Mismatch: a common problem in evaluating VIS systems Examples: - The contribution of a new visual encoding cannot be tested using a quantitative timing of the algorithm - A mischaracterized task cannot be addressed in a formal lab study

Matching Methods and Metrics

Insight-Based Evaluation (North et. al, 2005) Measure the utility of a visualization system by counting the number of insights generated by the user while using it

Insight-Based Evaluation Method No “benchmark tasks” Training on data and visualization for 15 minutes Participants list questions that they would like to pursue Asked to examine the data for as long as necessary until no new insights can be gained During analysis, the users asked to comment on their observations, inferences, and conclusions

Evaluating the Results The number of insights are tallied Insights: distinct observations about the data by each participant Collect all insights generated by all participants as a baseline Various quantitative statistics collected on insight generation (time spent, time to first insight, etc.)

What does insight-based evaluation address?

Problem: defining “insight” North’s definition: “[Insight is] an individual observation about the data by the participant, a unit of discovery. It is straightforward to recognize insight occurrences in a think-aloud protocol as any data observation that the user mentions is considered an insight.”

Example 1 “Our tool allows the biologists to interactively visualize and explore the whole set of trees, providing insight into the overall distribution and possible conflicting hypothesis” Insight = knowledge about the overall distribution

Example 2 “The analyst determined the answers to these questions, but also came up with further insights that she shared with people from other administrative units. She used the discovered information to advise other administrators of certain previously unknown relationships in their data” Insight = information about previously unknown relationships

Cognitive Science Definition Something measureable in the front lobes and the temporal lobes (superior temporal gyrus). Spontaneous Insight vs. Model-building Insight

Disambiguating “Insight” Knowledge-building insight: - Discovering insight, gaining insight, and providing insight - Insight as a substance, that accumulates over time and could be measured/quantified Spontaneous insight: - Experiencing insight, having an insight, or a moment of insight - Insight as a discrete event, that occurs at a specific moment in time and could be observed

Discussion Can we measure knowledge-building insight? Can we measure spontaneous insight? Are they related?

Questions?

Multi-dimensional In-depth Long-term Case studies Hypothesis: the efficacy of tools can be assessed by documenting: - Usage (observations, interviews, surveys, logging, etc.) - How successful the users are in achieving their professional goals MILCs – Shneiderman and Plaisant (2006)

What do MILCs address?

Definition Multi-dimensional: using observations, interviews, surveys, and loggers In-Depth: intense engagement of the researchers with the expert users to the point of becoming a partner or assistant Long-term: longitudinal studies that begin with training in use of a specific tool through proficient usage that leads to strategy changes for the expert users. Case studies: detailed reporting about a small number of individuals working on their own problems, in their own environment

Motivation MILCs has been embraced by a small community of researchers interested in studying creativity support tools. Challenges: - Cannot control for the users - Cannot control for the tasks - Toy problems in laboratories are not indicative of real-world problems and environments

Execution issues with MILCs Duration is always a problem Number of participants has to be small Familiarities are difficult - Understand organization policies and work culture - Gain access and permission to observe or interview - Observe users in their workplace, and collect subjective and objective quantitative and qualitative data. - Compile data of all types in all dimensions - Interpret the results - Isolate factors - Need to repeat the process

Questions?

Grounded Evaluation Proposed by Petra Isenberg et al. (2008) Definition: A process that attempts to ensure that the evaluation of an information visualization is situated within the context of its intended use

Call For Qualitative Measurements

Questions?

Learning-Based Evaluation (Chang, 2010) Working assumption: “the goal of visualization is to gain insight and knowledge” Proposal: we should evaluate a visualization based on whether or not the user actually gains insight or knowledge after using a visualization

Similar to Learning in Education How would an instructor choose between two textbooks for a course? We could: - Ask the students which book they prefer - Issue: they might like a book because its cover is pretty - Ask colleagues what book they prefer - Issue: different students in different environments - Ask the students to find some information in the book and measure how quickly they can perform the task - Issue: this only demonstrates how well the book is organized.

Metaphor for Visualization Evaluation In a best case scenario, we would: - Ask half of the student to use book one to learn a subject - Ask other half to use another book to learn the same subject Then we give the two groups the same test, and whichever scores higher “wins”

Flow Chart: Traditional LBE

Discussion Do you see any problems with this method?

Flow Chart: Single-System LBE

Discussion How should we evaluate your final projects?

Reminders / Administrivia On Wednesday: “Self-critique and feedback” Final Project Demonstration Day: 14 December, time TBA - Feel free to invite friends, colleagues, professors, etc. - If you need additional hardware (monitor, mouse, etc.), me Final deliverable: - Short (~2-page) write up on your completed project - Guidelines will be posted to course website next week - Officially due last day of finals: 22 December by 5pm - Consider submitting to conferences and undergrad research competitions! More details on venues to come