An Evaluation of Microarray Visualization Tools for Biological Insight Presented by Tugrul Ince and Nir Peer University of Maryland Purvi Saraiya Chris.

Slides:



Advertisements
Similar presentations
Testing Relational Database
Advertisements

The Robert Gordon University School of Engineering Dr. Mohamed Amish
1 Evaluation Rong Jin. 2 Evaluation  Evaluation is key to building effective and efficient search engines usually carried out in controlled experiments.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
HOW DO PROFESSIONAL DEVELOPERS COMPREHEND TO SOFTWARE Report submitted by Tobias Roehm, Rebecca Tiarks, Rainer Koschke, Walid Maalej.
Project Proposal.
An Interactive Visualization of Super-peer P2P Networks Peiqun (Anthony) Yu.
HCI Methods for Pathway Visualization Tools Purvi Saraiya, Chris North, Karen Duca* Virginia Tech Dept. of Computer Science, Center for Human-Computer.
Introduction and Overview “the grid” – a proposed distributed computing infrastructure for advanced science and engineering. Purpose: grid concept is motivated.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Virtual Workbenches Richard Anthony The University of Greenwich
Data Mining.
UMass Lowell Computer Science Advanced Algorithms Computational Geometry Prof. Karen Daniels Spring, 2004 Project.
Knowledge Acquisition. Knowledge Aquisition Definition – The process of acquiring, organising, & studying knowledge. Identified by many researchers and.
Evaluating the Quality of Image Synthesis and Analysis Techniques Matthew O. Ward Computer Science Department Worcester Polytechnic Institute.
Interactive Exploration of Hierarchical Clustering Results HCE (Hierarchical Clustering Explorer) Jinwook Seo and Ben Shneiderman Human-Computer Interaction.
Information Visualization Design for Multidimensional Data: Integrating the Rank-by-Feature Framework with Hierarchical Clustering Dissertation Defense.
(c) 2007 Mauro Pezzè & Michal Young Ch 1, slide 1 Software Test and Analysis in a Nutshell.
1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.
Assessing Student Learning
© 2011 Pearson Prentice Hall, Salkind. Nonexperimental Research: Qualitative Methods.
Info Vis: Multi-Dimensional Data Chris North cs3724: HCI.
HTA as a framework for task analysis Presenter: Hilary Ince, University of Idaho.
Microarray Gene Expression Data Analysis A.Venkatesh CBBL Functional Genomics Chapter: 07.
Multimodality and Activity Theory Methodological issues in their combination Dr. Mohammed Alhuthali Taif University, Saudi Arabia
Usability Methods: Cognitive Walkthrough & Heuristic Evaluation Dr. Dania Bilal IS 588 Spring 2008 Dr. D. Bilal.
Presenter: Shant Mandossian EFFECTIVE TESTING OF HEALTHCARE SIMULATION SOFTWARE.
CS 235: User Interface Design November 12 Class Meeting Department of Computer Science San Jose State University Fall 2014 Instructor: Ron Mak
Charts and Graphs V
Information Design and Visualization
Graphical Analysis. Why Graph Data? Graphical methods Require very little training Easy to use Massive amounts of data can be presented more readily Can.
Using Bayesian Networks to Analyze Expression Data N. Friedman, M. Linial, I. Nachman, D. Hebrew University.
Presented by Abirami Poonkundran.  Introduction  Current Work  Current Tools  Solution  Tesseract  Tesseract Usage Scenarios  Information Flow.
Requirements Engineering Requirements Elicitation Process Lecture-8.
Exploratory Research Design Week 02
Human Computer Interaction
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 7: Focusing on Users and Their Tasks.
Chapter 12 Evaluating Products, Processes, and Resources.
Course Organization & Format Visualization II MSIM 842, CS 795/895 Instructor: Jessica Crouch.
Visualization and analysis of microarray and gene ontology data with treemaps Eric H Baehrecke, Niem Dang, Ketan Babaria and Ben Shneiderman Presenter:
An Internet of Things: People, Processes, and Products in the Spotfire Cloud Library Dr. Brand Niemann Director and Senior Data Scientist/Data Journalist.
Usability Testing Chapter 6. Reliability Can you repeat the test?
1 STAT 500 – Statistics for Managers STAT 500 Statistics for Managers.
Document Clustering for Forensic Analysis: An Approach for Improving Computer Inspection.
Building Simulation Model In this lecture, we are interested in whether a simulation model is accurate representation of the real system. We are interested.
Process Quality in ONS Rachel Skentelbery, Rachael Viles & Sarah Green
Presented By: Solutions Delivery Managing Reports in CRMnext.
ASSESSING LEARNING ALGORITHMS Yılmaz KILIÇASLAN. Assessing the performance of the learning algorithm A learning algorithm is good if it produces hypotheses.
Theme 2: Data & Models One of the central processes of science is the interplay between models and data Data informs model generation and selection Models.
Knowledge-based Support in a Group Decision Making Context: An Expert-Novice Comparison Fiona Fui-Hoon Nah, University of Nebraska-Lincoln Izak Benbasat,
Just Another Pretty Visualization? The What, Where, When, Why and How of Evaluating Visualizations Jim Foley College of Computing Georgia Institute of.
ESIP Semantic Web Products and Services ‘triples’ “tutorial” aka sausage making ESIP SW Cluster, Jan ed.
Computational Approaches for Biomarker Discovery SubbaLakshmiswetha Patchamatla.
Architecture View Models A model is a complete, simplified description of a system from a particular perspective or viewpoint. There is no single view.
Identifying Needs and Establishing Requirements Presenters: Veronica Gasca Jennifer Rhough.
1 Mining Images of Material Nanostructure Data Aparna S. Varde, Jianyu Liang, Elke A. Rundensteiner and Richard D. Sisson Jr. ICDCIT December 2006 Bhubaneswar,
Library Online Resource Analysis (LORA) System Introduction Electronic information resources and databases have become an essential part of library collections.
Bell Ringer You will need a new bell ringer sheet – write your answers in the Monday box. 3. Airport administrators take a sample of airline baggage and.
1 An Efficient Optimal Leaf Ordering for Hierarchical Clustering in Microarray Gene Expression Data Analysis Jianting Zhang Le Gruenwald School of Computer.
Supporting the design of interactive systems a perspective on supporting people’s work Hans de Graaff 27 april 2000.
Usability Engineering Dr. Dania Bilal IS 587 Fall 2007.
Evaluation / Usability. ImplementDesignAnalysisEvaluateDevelop ADDIE.
What is Research Design? RD is the general plan of how you will answer your research question(s) The plan should state clearly the following issues: The.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Micro array Data Analysis. Differential Gene Expression Analysis The Experiment Micro-array experiment measures gene expression in Rats (>5000 genes).
Usability & Evaluation in Visualizing Biological Data Chris North, Virginia Tech VizBi.
SIE 515 Design Evaluation Lecture 7.
Title of your science project
Pre-Activity: 1. Recap? 2. Research Says?
Paul Craig, Jessie Kennedy
Presentation transcript:

An Evaluation of Microarray Visualization Tools for Biological Insight Presented by Tugrul Ince and Nir Peer University of Maryland Purvi Saraiya Chris North Dept. of Computer Science Virginia Polytechnic Institute and State University Karen Duca Virginia Bioinformatics Institute Virginia Polytechnic Institute and State University

2 Goals Evaluate five popular visualization tools Evaluate five popular visualization tools Cluster/Treeview Cluster/Treeview TimeSearcher TimeSearcher Hierarchical Clustering Explorer (HCE) Hierarchical Clustering Explorer (HCE) Spotfire Spotfire GeneSpring GeneSpring Do so in the context of bioinformatics data exploration Do so in the context of bioinformatics data exploration

3 Goals Research Questions Research Questions How successful are these tools in stimulating insight? How successful are these tools in stimulating insight? How do various visualization techniques affect the users’ perception of data? How do various visualization techniques affect the users’ perception of data? How does users’ background affect the tool usage? How does users’ background affect the tool usage? How do these tools support hypothesis generation? How do these tools support hypothesis generation? Can insight be measured in a controlled experiment? Can insight be measured in a controlled experiment?

4 Visualization Evaluations Typically evaluations consist of Typically evaluations consist of controlled measurements of user performance and accuracy on predetermined tasks controlled measurements of user performance and accuracy on predetermined tasks We are looking for an evaluation that better simulates a bioinformatics data analysis scenario We are looking for an evaluation that better simulates a bioinformatics data analysis scenario We use a protocol the focuses on We use a protocol the focuses on recognition and quantification of insights gained from actual exploratory use of visualizations recognition and quantification of insights gained from actual exploratory use of visualizations

5 Insights Hard to define what is an “insight” Hard to define what is an “insight” We need this term to be quantifiable and reproducible We need this term to be quantifiable and reproducible Solution Solution Encourage users to think aloud Encourage users to think aloud and report any findings they have about the dataset and report any findings they have about the dataset Videotape a session to capture and characterize individual insights as they occur Videotape a session to capture and characterize individual insights as they occur generally provides more information than subjective measures from post-experiment surveys generally provides more information than subjective measures from post-experiment surveys

6 Insights Define insight as Define insight as an individual observation about the data by the participant an individual observation about the data by the participant a unit of discovery a unit of discovery Essentially, any data observation made during the think aloud protocol Essentially, any data observation made during the think aloud protocol Now we can quantify some characteristics of each insight Now we can quantify some characteristics of each insight

7 Insight Characteristics Observation Observation The actual finding about the data The actual finding about the data Time Time The amount of time taken to reach the insight The amount of time taken to reach the insight Domain Value Domain Value The significance of the insight. Coded by a domain expert. The significance of the insight. Coded by a domain expert. Hypotheses Hypotheses Hypothesis and direction of research Hypothesis and direction of research Directed vs. Unexpected Directed vs. Unexpected Recall: participants are asked to identify questions they want to explore Recall: participants are asked to identify questions they want to explore Correctness Correctness Breadth vs. Depth Breadth vs. Depth

8 Insight Characteristics Category Category Overview – overall distributions of gene expression Overview – overall distributions of gene expression Patterns – identification or comparison across data attributes Patterns – identification or comparison across data attributes Groups – identification or comparison of groups of genes Groups – identification or comparison of groups of genes Details – focused information about specific genes Details – focused information about specific genes

9 Experiment Design A 3  5 between-subjects design A 3  5 between-subjects design between-subjects  different subjects for each pair between-subjects  different subjects for each pair Dataset: 3 treatments Dataset: 3 treatments Visualization tool: 5 treatments Visualization tool: 5 treatments

10 Experiment Design Participants Participants 2 participants per dataset per tool 2 participants per dataset per tool Have at least a Bachelor’s degree in a biological field Have at least a Bachelor’s degree in a biological field Assigned to tools they had never worked with before Assigned to tools they had never worked with before to prevent advantage to prevent advantage measure learning time measure learning time Categories Categories 10 Domain Experts 10 Domain Experts Senior researchers with extensive experience in microarray experiments and microarray data analysis Senior researchers with extensive experience in microarray experiments and microarray data analysis 11 Domain Novices 11 Domain Novices Lab technicians or graduate student research assistants Lab technicians or graduate student research assistants 9 Software Developers 9 Software Developers Professionals who implement microarray software tools Professionals who implement microarray software tools

11 Protocol and Measures Chose new users with only minimal tool training Chose new users with only minimal tool training Success in the initial usage period is critical for the tool’s adoption by biologists Success in the initial usage period is critical for the tool’s adoption by biologists Participants received an initial training Participants received an initial training Background description about the dataset Background description about the dataset 15-minute tool tutorial 15-minute tool tutorial Participants listed some analysis questions Participants listed some analysis questions Instructed to examine the data with the tool as long as needed Instructed to examine the data with the tool as long as needed They were allowed to ask for help about the tool They were allowed to ask for help about the tool Simulates training by colleagues Simulates training by colleagues

12 Protocol and Measures Every 15 minutes, participants estimated percent of total potential insight they obtained so far Every 15 minutes, participants estimated percent of total potential insight they obtained so far Finally, assessed overall experience with the tools during session Finally, assessed overall experience with the tools during session Entire session was videotaped for later analysis Entire session was videotaped for later analysis Later, all individual occurrences of insights were identified and codified Later, all individual occurrences of insights were identified and codified

13 Show me pictures Here are the tools!!!

14 Cluster/TreeView = ClusterView Cluster Cluster to cluster data to cluster data TreeView TreeView Visualize the clusters Visualize the clusters Uses heat-maps Uses heat-maps

15 TimeSearcher 1 Parallel Coordinate Visualization Parallel Coordinate Visualization Interactive Filtering Interactive Filtering Line Graphs for each data entity Line Graphs for each data entity

16 HCE Clusters data Clusters data Several Visualizations Several Visualizations Heat-Maps Heat-Maps Parallel Coordinates Parallel Coordinates Scatter Plots Scatter Plots Histograms Histograms Brushing and Linking Brushing and Linking

17 Spotfire General Purpose Visualization Tool General Purpose Visualization Tool Several Displays Several Displays Scatter Plots Scatter Plots Bar Graphs Bar Graphs Histograms Histograms Pie/Line Charts Pie/Line Charts Others… Others… Dynamic Query Sliders Dynamic Query Sliders Brushing and Linking Brushing and Linking

18 GeneSpring Suitable for Microarray data analysis Suitable for Microarray data analysis Shows physical positions on genomes Shows physical positions on genomes Array layouts Array layouts Pathways Pathways Gene-to-gene comparison Gene-to-gene comparison Brushing and Linking Brushing and Linking Clustering capability Clustering capability

19 Enough about Tools, Tell me the Results!!!

20 ClusterViewTimeSearcher 1HCESpotfireGeneSpring Number of Insights Spotfire: Highest number of insights Spotfire: Highest number of insights HCE: poorest HCE: poorest

21 Total Domain Value Spotfire: Highest insight value Spotfire: Highest insight value HCE, GeneSpring: poorer HCE, GeneSpring: poorer ClusterViewTimeSearcher 1HCESpotfireGeneSpring

22 Avg. Final Amount Learned Spotfire: high value in learning Spotfire: high value in learning ClusterView and HCE are poor ClusterView and HCE are poor ClusterViewTimeSearcher 1HCESpotfireGeneSpring

23 Avg. Time to First Insight ClusterView: very short time to first insight ClusterView: very short time to first insight TimeSearcher 1 and Spotfire are also quick TimeSearcher 1 and Spotfire are also quick ClusterViewTimeSearcher 1HCESpotfireGeneSpring

24 Avg. Total Time Total time users spent using the tool Total time users spent using the tool Low Values: Efficient or Not useful for insight Low Values: Efficient or Not useful for insight ClusterViewTimeSearcher 1HCESpotfireGeneSpring

25 Unexpected Insights HCE revealed several unexpected results HCE revealed several unexpected results ClusterView provided a few ClusterView provided a few TimeSearcher 1 for time series data TimeSearcher 1 for time series data Spotfire contributed to 2 unexpected insights Spotfire contributed to 2 unexpected insights Hypotheses A few insights led to hypotheses A few insights led to hypotheses Spotfire  3 Spotfire  3 ClusterView  2 ClusterView  2 TimeSearcher 1  1 TimeSearcher 1  1 HCE  1 HCE  1

26 Tools vs. Datasets

27 Insight Categories Overall Gene Expression Overall Gene Expression Overview of genes in general Overview of genes in general Expression Patterns Expression Patterns Searching patterns is critical Searching patterns is critical Clustering is useful Clustering is useful Grouping Grouping Some users wanted to group genes Some users wanted to group genes GeneSpring enables grouping GeneSpring enables grouping Detail Information Detail Information Users want detailed information about genes that are familiar to them Users want detailed information about genes that are familiar to them

28 Visual Representations and Interactions Although some tools have many visualization techniques, users tend to use only a few Although some tools have many visualization techniques, users tend to use only a few Spotfire users preferred heat-maps Spotfire users preferred heat-maps GeneSpring users preferred parallel coordinates GeneSpring users preferred parallel coordinates Lupus dataset: visualized best with heat-maps Lupus dataset: visualized best with heat-maps Most users preferred outputs of clustering algorithms Most users preferred outputs of clustering algorithms HCE not useful when a particular column arrangement is useful HCE not useful when a particular column arrangement is useful

29 Running out of time, So, wrap up Use a Visualization tool (that’s why we’re here!) Use a Visualization tool (that’s why we’re here!) Spotfire: best general performance Spotfire: best general performance GeneSpring: Hard to use GeneSpring: Hard to use Dataset dictates best tool! Dataset dictates best tool! Time Series data: TimeSearcher Time Series data: TimeSearcher Others: Spotfire, GeneSpring? Others: Spotfire, GeneSpring? Interaction is the key Interaction is the key Grouping and Clustering are necessary features Grouping and Clustering are necessary features

30 Critique In all fairness, measuring insights is really hard! Here are some possible issues In all fairness, measuring insights is really hard! Here are some possible issues Subjectivity Subjectivity Experiment relies on users always thinking aloud Experiment relies on users always thinking aloud Also, depends on a domain expert to evaluate insights Also, depends on a domain expert to evaluate insights Results may vary widely based on participants expertise (only two per tool-dataset pair) Results may vary widely based on participants expertise (only two per tool-dataset pair) Some insight characteristics are inherently subjective Some insight characteristics are inherently subjective Domain Value Domain Value Breadth vs. Depth Breadth vs. Depth

31 Critique How do one count insights? How do one count insights? Assumes honest reporting by participants Assumes honest reporting by participants Some insights may be of no great value Some insights may be of no great value What if a discovery just reaffirms a known fact? Is that an insight? What if a discovery just reaffirms a known fact? Is that an insight? Measuring time taken to reach an insight Measuring time taken to reach an insight Maybe instead of measuring from beginning of session we should measure from last insight Maybe instead of measuring from beginning of session we should measure from last insight