Testing plan outline Adam Leko Hans Sherburne HCS Research Laboratory University of Florida.

Slides:



Advertisements
Similar presentations
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
Advertisements

By Eva Freund, The IV&V Group, Inc.
1 Steve Chenoweth Tuesday, 10/04/11 Week 5, Day 2 Right – Typical tool for reading out error codes logged by your car’s computer, to help analyze its problems.
Verification “Auditability” Can Be “Free” James H. Jones Optants Documented Decision Support Company (ODDSCO)
System Design and Analysis
1 Software Testing and Quality Assurance Lecture 15 - Planning for Testing (Chapter 3, A Practical Guide to Testing Object- Oriented Software)
Chapter 3 Applications Software: Getting the Work Done.
Non-functional requirements
Software Documentation Written By: Ian Sommerville Presentation By: Stephen Lopez-Couto.
Welcome to CMPE003 Personal Computer Concepts: Hardware and Software Winter 2003 UC Santa Cruz Instructor: Guy Cox.
Innovations in Justice Information Sharing Strategies and Best Practices February 2007 Melissa R. Johnson, CCA Communications Director, International Association.
Software Development, Programming, Testing & Implementation.
Database Design IST 7-10 Presented by Miss Egan and Miss Richards.
Introduction to Computer Technology
The Project AH Computing. Functional Requirements  What the product must do!  Examples attractive welcome screen all options available as clickable.
United Nations Economic Commission for Europe Statistical Division Applying the GSBPM to Business Register Management Steven Vale UNECE
Introduction to Systems Analysis and Design Trisha Cummings.
PAPI Tool Evaluation Bryan Golden 1/4/2004 HCS Research Laboratory University of Florida.
Semester 1, 2003 Week 7 CSE9020 / 1 Software Testing and Quality Assurance With thanks to Shonali Krishnaswamy and Sylvia Tucker.
Chapter 8: Systems analysis and design
UPC/SHMEM PAT High-level Design v.1.1 Hung-Hsun Su UPC Group, HCS lab 6/21/2005.
MpiP Evaluation Report Hans Sherburne, Adam Leko UPC Group HCS Research Laboratory University of Florida.
Bottlenecks: Automated Design Configuration Evaluation and Tune.
Robert Sopko Stephen Miller Amy Gandhi Jazimar Bailey.
Parallel Programming Models Jihad El-Sana These slides are based on the book: Introduction to Parallel Computing, Blaise Barney, Lawrence Livermore National.
Software System Engineering: A tutorial
1 Technical & Business Writing (ENG-315) Muhammad Bilal Bashir UIIT, Rawalpindi.
11 July 2005 Tool Evaluation Scoring Criteria Professor Alan D. George, Principal Investigator Mr. Hung-Hsun Su, Sr. Research Assistant Mr. Adam Leko,
Software Project Planning Defining the Project Writing the Software Specification Planning the Development Stages Testing the Software.
User-Centered System Design. - a philosophy of user interface design introduced by Don Norman & Steve Draper in 1986.
An Internet of Things: People, Processes, and Products in the Spotfire Cloud Library Dr. Brand Niemann Director and Senior Data Scientist/Data Journalist.
SvPablo Evaluation Report Hans Sherburne, Adam Leko UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red:
Martin Schulz Center for Applied Scientific Computing Lawrence Livermore National Laboratory Lawrence Livermore National Laboratory, P. O. Box 808, Livermore,
Software Development Process.  You should already know that any computer system is made up of hardware and software.  The term hardware is fairly easy.
ICOM 6115: COMPUTER SYSTEMS PERFORMANCE MEASUREMENT AND EVALUATION Nayda G. Santiago August 16, 2006.
Automatically Repairing Broken Workflows for Evolving GUI Applications Sai Zhang University of Washington Joint work with: Hao Lü, Michael D. Ernst.
COMP 208/214/215/216 – Lecture 8 Demonstrations and Portfolios.
1 Duplicate Analyzer Exercises. 2 Installation and Initial Configuration: Exercises Exercises 1.Install Duplicate Analyzer on your local PC. 2.Configure.
FIX Eye FIX Eye Getting started: The guide EPAM Systems B2BITS.
Lesson 1 Operating Systems, Part 1. Objectives Describe and list different operating systems Understand file extensions Manage files and folders.
Test Plan Objectives of Test Plan Product –May need to sell test plan Military mil spec FDA specs Phone company (need to support after your company.
Dynaprof Evaluation Report Adam Leko, Hans Sherburne UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red:
August 2005 TMCOps TMC Operator Requirements and Position Descriptions Phase 2 Interactive Tool Project Presentation.
HPCToolkit Evaluation Report Hans Sherburne, Adam Leko UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red:
1 EndNote X2 Your Bibliographic Management Tool 29 September 2009 Humanities and Social Sciences Resource Teams.
Tool Visualizations, Metrics, and Profiled Entities Overview [Brief Version] Adam Leko HCS Research Laboratory University of Florida.
Dynaprof Evaluation Report Adam Leko, Hans Sherburne UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red:
CISC Machine Learning for Solving Systems Problems Presented by: Suman Chander B Dept of Computer & Information Sciences University of Delaware Automatic.
Overview of dtrace Adam Leko UPC Group HCS Research Laboratory University of Florida Color encoding key: Blue: Information Red: Negative note Green: Positive.
Thomas Kern | The system documentation as binding agent for and in between internal and external customers April 24th, 2009 | Page 1 The system documentation.
NSF DUE ; Wen M. Andrews J. Sargeant Reynolds Community College Richmond, Virginia.
21 Sep UPC Performance Analysis Tool: Status and Plans Professor Alan D. George, Principal Investigator Mr. Hung-Hsun Su, Sr. Research Assistant.
Spring ’05 Independent Study Midterm Review Hans Sherburne HCS Research Laboratory University of Florida.
UPC Status Report - 10/12/04 Adam Leko UPC Project, HCS Lab University of Florida Oct 12, 2004.
Project Planning Defining the project Software specification Development stages Software testing.
Profiling/Tracing Method and Tool Evaluation Strategy Summary Slides Hung-Hsun Su UPC Group, HCS lab 1/25/2005.
Unit 3 Computer Systems. What is software? unlike hardware it can’t be physically touched it’s the missing link between the computer hardware and the.
1 April 14, Starting New Open Source Software Projects William Cohen NCSU CSC 591W April 14, 2008.
ICAICT201A USE COMPUTER OPERATING SYSTEM. USING THE CONTROL PANEL The Control Panel contains many options for configuring your computer, including: adding.
SOFTWARE TESTING TRAINING TOOLS SUPPORT FOR SOFTWARE TESTING Chapter 6 immaculateres 1.
Fundamentals of Information Systems, Sixth Edition
SEVERITY & PRIORITY RELATIONSHIP
Data Analysis of EnchantedLearning.com vs. Invent.org
System Development Life Cycle (SDLC)
(Additional materials)
EIN 6133 Enterprise Engineering
❶ Start by selecting the MEPS – area where the motor will be placed
Introduction to Software Testing
Introduction to Systems Analysis and Design
Software Requirements Specification (SRS) Template.
Presentation transcript:

Testing plan outline Adam Leko Hans Sherburne HCS Research Laboratory University of Florida

2 Testing plan Tasks to accomplish  Come up with general steps to test each aspect of our tool evaluation strategy  Select 3-5 applications that exhibit particular problems (too much synchronization, load imbalance) to use for every tool Re-evaluate Paradyn, TAU? (PAPI: n/a)  Create listing of tools we wish to evaluate For each:  Determine how to obtain  Contact developers for evaluation license if necessary (make contacts w/developers)  Gauge popularity of tool based on information available and information from developers General strategy  For each tool, Describe overall architecture for tool (e.g., how instrumentation is achieved, etc) Give general comments about tool, including strong and weak points Estimate (contacting developers as necessary) how hard it would be to extend tool to support UPC and SHMEM Give scores (1-5, 5 best, 0 if n/a) for each feature listed in table 9.1 of last semester’s report  Put into “feature matrix” spreadsheet  Some scores may need to be filled in after to give relative ratings (available metrics, etc)  After installing a tool, If tool similar to other tools or simple/not useful, spend less time evaluating tool If tool is a good candidate for extending, spend more time on it In each case, feature matrix and performance factors will be recorded for each toolAfter each evaluation, put together 5-15 slide presentation and show to group for further comments

3 “Feature matrix” testing plan Available metrics ( )  Look at documentation, run tool, record what is available  Give subjective score for depth of features provided Cost (9.1.1)  Look up on web site or documentation  If free, give maximum score  If not, give score based on how much per seat $0-$500: 4 $500-$1000: 3 $1000-$2000: 2 $2000+: 1 Documentation quality (9.3.2)  Subjective score based on reading available documentation and manuals (web sites, PDF files, etc)  Give special emphasis to tools that provide documentation for other developers wishing to extend their tool Extendibility (9.3.1)  Open-source tools may be given higher scores if the code is well-written Take a quick look at existing source to determine  Best guess at how much work is required to add UPC and SHMEM support into tool  May have to talk to developers to get better information (leverage contacts here) Would existing companies be willing to work with us to extend tool? Filtering and aggregation ( )  Subjective score based on how frugal tool is with presenting information vs. usefulness of information Hardware support (9.1.4)  Obtain from documentation  One full point for each architecture, give weight to architectures that support SHMEM/UPC

4 “Feature matrix” testing plan Heterogeneity support (9.1.5)  Can determine from documentation and from running tool  (Not a necessary feature though) Installation (9.1.2)  Subjective measurement based on time and effort needed for installation  One person (Adam) should do all to give fair comparisons Leverage Adam’s SysAdmin experience Can move to two people if this becomes too demanding Interoperability ( )  Can determine from documentation and from running tool  Score based on how many formats tool can export to and how popular formats are e.g., 1 point for SvPablo’s format Question: Which formats are useful? Learning curve (9.1.6)  Subjective score based on how long it takes to become proficient with standard options of tool Manual overhead ( )  Subjective score based on how much time it takes to insert instrumentation calls in source code Less lines inserted = higher score  Tools with automatic instrumentation instrumentation will receive 5 Measurement accuracy ( )  Testing method: Run code with no instrumentation and insert timing calls Run code with instrumentation necessary to pinpoint bottlenecks and see how overall wall time was affected  Program’s correctness should not be affected Keep output from correct run and diff with output of instrumented run to make sure results are the same

5 “Feature matrix” testing plan Multiple analyses ( )  Can determine from documentation and from running tool Multiple executions (9.3.5)  Can determine from documentation and from running tool Multiple views ( )  Can determine from documentation and from running tool More views presented = higher score Special weight to hierarchical presentations  Keep record of which views tool supports Performance bottleneck identification ( )  See how well tool either: Automatically diagnoses problems Guides user to pinpoint bottlenecks in code  Use our benchmark suite for test cases  Probably give equal weight to each application in our suite the tool was successful in identifying bottleneck Profiling / tracing support ( )  Can determine from documentation and from running tool  For tracing: Keep record of size of each trace file for standard benchmark applications used during evaluations Assign trace-specific score after all tools have been evaluated based on relative trace file size

6 “Feature matrix” testing plan Response time (9.2.6)  Subjective score based on measured observations from running tool Searching (9.3.6)  Can determine from documentation and from running tool Software support (9.1.3)  Can determine from documentation and from running tool Source code correlation ( )  Can determine from documentation and from running tool System stability (9.3.3)  Subjective score based on measured observations from running tool Technical support (9.3.4)  Not really sure how to score this column  Will most likely have to be subjective

7 Benchmark suite 3-5 applications used during performance tool testing Each has a particular “bottleneck” and a decent amount of parallelism (~30-50% efficiency)  Too much unnecessary synchronization (can a tool determine this?)  Too many small-sized messages (fix w/message aggregation and communication restructuring)  Etc Each program should be “braindead” in that fixing the bottleneck will improve overall efficiency  Also include “regular” program to make sure tool doesn’t give bogus recommendations Applications should be well-known and mainly well-implemented Problem size about ~200s running time sequential

8 Task dependencies Tool evaluations  Inputs Brian  Ideas on how to evaluate usability  General rules accepted in literature (if any)  Outputs Recommendations  Good ideas from existing tools  Things to avoid that caused problems in existing tools  If any tools should be extended to support UPC/SHMEM Literature searches  Inputs None in particular  Outputs Strategies to best provide analysis capabilities in our tool Recommendations for incorporating bottleneck identification processes in our tool Some idea for Adam’s dissertation!