2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 An Under-the-Hood Review Sorin Faibish.

Slides:



Advertisements
Similar presentations
© 2003 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice Performance Measurements of a User-Space.
Advertisements

Performance Testing - Kanwalpreet Singh.
1 Sizing the Streaming Media Cluster Solution for a Given Workload Lucy Cherkasova and Wenting Tang HPLabs.
Tahir Nawaz Introduction to.NET Framework. .NET – What Is It? Software platform Language neutral In other words:.NET is not a language (Runtime and a.
Software Frame Simulator (SFS) Technion CS Computer Communications Lab (236340) in cooperation with ECI telecom Uri Ferri & Ynon Cohen January 2007.
Serverless Network File Systems. Network File Systems Allow sharing among independent file systems in a transparent manner Mounting a remote directory.
Network Redesign and Palette 2.0. The Mission of GCIS* Provide all of our users optimal access to GCC’s technology resources. *(GCC Information Services:
© 2012 IBM Corporation What’s new in OpenAdmin Tool for Informix? Erika Von Bargen May 2012.
Chapter 4 M. Keshtgary Spring 91 Type of Workloads.
 Need for a new processing platform (BigData)  Origin of Hadoop  What is Hadoop & what it is not ?  Hadoop architecture  Hadoop components (Common/HDFS/MapReduce)
An Adaptable Benchmark for MPFS Performance Testing A Master Thesis Presentation Yubing Wang Advisor: Prof. Mark Claypool.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Concurrency Control & Caching Consistency Issues and Survey Dingshan He November 18, 2002.
Module – 7 network-attached storage (NAS)
Northwestern University 2007 Winter – EECS 443 Advanced Operating Systems The Google File System S. Ghemawat, H. Gobioff and S-T. Leung, The Google File.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: February 2010.
Computer System Lifecycle Chapter 1. Introduction Computer System users, administrators, and designers are all interested in performance evaluation. Whether.
Microsoft ® Application Virtualization 4.6 Infrastructure Planning and Design Published: September 2008 Updated: November 2011.
1 An SLA-Oriented Capacity Planning Tool for Streaming Media Services Lucy Cherkasova, Wenting Tang, and Sharad Singhal HPLabs,USA.
RUP Implementation and Testing
Appendix B Planning a Virtualization Strategy for Exchange Server 2010.
1 Moshe Shadmon ScaleDB Scaling MySQL in the Cloud.
Tag line, tag line Power Management in Storage Systems Kaladhar Voruganti Technical Director CTO Office, Sunnyvale June 12, 2009.
1 Wenguang WangRichard B. Bunt Department of Computer Science University of Saskatchewan November 14, 2000 Simulating DB2 Buffer Pool Management.
Ken Cantrell / NetApp Mark Rogov / EMC July 30, 2015.
OLAP Council APB-1 OLAP Benchmark Release II
DONE-08 Sizing and Performance Tuning N-Tier Applications Mike Furgal Performance Manager Progress Software
Simulating a $2M Commercial Server on a $2K PC Alaa R. Alameldeen, Milo M.K. Martin, Carl J. Mauer, Kevin E. Moore, Min Xu, Daniel J. Sorin, Mark D. Hill.
CE Operating Systems Lecture 3 Overview of OS functions and structure.
ICOM 6115: Computer Systems Performance Measurement and Evaluation August 11, 2006.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
1 Evaluation of Cooperative Web Caching with Web Polygraph Ping Du and Jaspal Subhlok Department of Computer Science University of Houston presented at.
GFS. Google r Servers are a mix of commodity machines and machines specifically designed for Google m Not necessarily the fastest m Purchases are based.
Desktop Virtualization
11 CLUSTERING AND AVAILABILITY Chapter 11. Chapter 11: CLUSTERING AND AVAILABILITY2 OVERVIEW  Describe the clustering capabilities of Microsoft Windows.
VMware vSphere Configuration and Management v6
Network design Topic 6 Testing and documentation.
Estimating “Size” of Software There are many ways to estimate the volume or size of software. ( understanding requirements is key to this activity ) –We.
Effort Estimation In WBS,one can estimate effort (micro-level) but needed to know: –Size of the deliverable –Productivity of resource in producing that.
CSI 3125, Preliminaries, page 1 SERVLET. CSI 3125, Preliminaries, page 2 SERVLET A servlet is a server-side software program, written in Java code, that.
Module 9 Planning and Implementing Monitoring and Maintenance.
Chapter 9: Networking with Unix and Linux. Objectives: Describe the origins and history of the UNIX operating system Identify similarities and differences.
Performance Testing Test Complete. Performance testing and its sub categories Performance testing is performed, to determine how fast some aspect of a.
Ken Cantrell / NetApp Mark Rogov / EMC J Metz / Cisco October 21, 2015.
PRESENTATION TITLE GOES HERE Emerald NAS Extensions Chuck Paridon Performance Architect H-P Enterprise Data Contributed by Nick Principe – EMC, Demartek.
Jérôme Jaussaud, Senior Product Manager
1 Evaluation of Cooperative Web Caching with Web Polygraph Ping Du and Jaspal Subhlok Department of Computer Science University of Houston presented at.
7.1 Operating Systems. 7.2 A computer is a system composed of two major components: hardware and software. Computer hardware is the physical equipment.
LIOProf: Exposing Lustre File System Behavior for I/O Middleware
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
 Cloud Computing technology basics Platform Evolution Advantages  Microsoft Windows Azure technology basics Windows Azure – A Lap around the platform.
Performance Testing - LR. 6/18/20162 Contents Why Load Test Your Web Application ? Functional vs. Load Web Testing Web-Based, Multi-Tiered Architecture.
DISTRIBUTED FILE SYSTEM- ENHANCEMENT AND FURTHER DEVELOPMENT BY:- PALLAWI(10BIT0033)
PHD Virtual Technologies “Reader’s Choice” Preferred product.
Computer System Structures
Business System Development
Introduction to Distributed Platforms
Diskpool and cloud storage benchmarks used in IT-DSS
Chapter 2: System Structures
Chapter 9: The Client/Server Database Environment
Google File System CSE 454 From paper by Ghemawat, Gobioff & Leung.
Oracle Solaris Zones Study Purpose Only
FACTON Provides Businesses with a Cloud Solution That Elevates Enterprise Product Cost Management to a New Level Using the Power of Microsoft Azure MICROSOFT.
Dev Test on Windows Azure Solution in a Box
11/23/2018 3:03 PM © Microsoft Corporation. All rights reserved. MICROSOFT MAKES NO WARRANTIES, EXPRESS, IMPLIED OR STATUTORY, AS TO THE INFORMATION IN.
Chapter 2: System Structures
Cloud computing mechanisms
Mark Quirk Head of Technology Developer & Platform Group
In Today’s Class.. General Kernel Responsibilities Kernel Organization
Presentation transcript:

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 An Under-the-Hood Review Sorin Faibish Spencer Shepler

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 The Workloads and Metrics an Under-the-Hood Review  Historically, the SPEC SFS benchmark and its NFS and CIFS workloads have been the industry standard for peer reviewed, published, performance results for the NAS industry. The SPEC SFS framework has a strong history of reliable, reproducible results and the SPEC organization provides a structure to ensure that users have complete and comparable results.  The SPEC SFS 2014 benchmark generates file system workload through the use of operating system APIs instead of generating NFS or CIFS protocol messages directly as previous versions of SFS have done. The use of operating system APIs allows for measurement end-to-end and thus expands beyond the traditional NAS-only server measurement. For example the following can be measured and reported: local file systems, client-side file systems, network types (RDMA or non-RDMA), or client-side caching. This presentation will provide a review of the benchmark framework and its approach to generating workloads  This presentation will provide a review of the SPEC SFS 2014 workloads and a brief background of the development of each.  The attendee will leave with an understanding of the framework, its metrics and be ready to engage in measurement and product improvement. 2

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Today’s Outline  Intro / Contributions / Motivation  SPEC SFS 2014 Framework  SPEC SFS 2014 Reporting  Workloads / Business Metrics  VDA  VDI  SWBUILD  Database 3

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Tonight’s BOF  Drinks and Snacks  Open discussion and additional detail…  Tonight 9:00 PM – 10:00 PM  Camino Real Room 4

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC Standard Performance Evaluation Corporation  The Standard Performance Evaluation Corporation (SPEC) is a non-profit corporation formed to establish, maintain and endorse a standardized set of relevant benchmarks that can be applied to the newest generation of high-performance computers. SPEC develops benchmark suites and also reviews and publishes submitted results from member organizations and other benchmark licensees  5

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Disclaimer  The SPEC SFS 2014 benchmark, as represented in this presentation, was released SW in November 4 th

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Contributions EMCTracing code, Validation, Testing, VDI Workload HitachiValidation, Testing HuaweiValidation, Testing IBMReporting Tools, Validation, Testing, VDA Workload Iozone.orgSource Code, Development, Testing, Validation, Binaries MicrosoftNative Port to Windows, Validation, Testing, VDI Workload NetAppValidation, Testing, SWBUILD and DATABASE Workloads OracleValidation, Testing, DATABASE Workload SeagateValidation, Testing 7

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Motivation for SPEC SFS

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Motivation for SPEC SFS 2014  SPEC SFS 2014 moves to Solution Benchmarking  Realistic, Solution-based workloads  DATABASE, SWBUILD, VDA, VDI  Workloads based on traces, like previous SFS 2008  Modern scenarios based on standard solutions  Benchmark measures application-level performance  Uses file system APIs at the client  Advanced measurement – quality of service  Ops and latency don’t tell the whole story → business metrics  Ability to measure broad range of products and configurations  Clients, Servers, Local File Systems, Networking Transports  All of these now contribute to measured performance of solution  Allowing for multi-tiered storage solutions 9

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Framework 10

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Framework  Two components:  Load generator: netmist  Highly customizable, powerful, workload generator  SPEC SFS 2014 license includes full version  Wrappers: SfsManager  Provides ease of configuration  Coordinates running multiple load points (scaling)  Implements business metric logic  Framework features:  Multi-client support is fundamental  Supports many operating systems and virtual machines  Protocol/file system agnostic  Definable workloads  Full source code included with SPEC SFS 2014 benchmark 11

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Framework  Benchmark execution phases  Validation  Initialization  Warmup  Measurement (Run)  Results  This sequence of execution phases repeats for each requested load point 12

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Reporting 13

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Reporting Publication of Results  Prior to public disclosure, SPEC SFS 2014 results must be submitted for review by the SPEC SFS subcommittee  Results are peer-reviewed for consistency and compliance with the SPEC SFS 2014 Run and Reporting Rules  Disclosure must be adequate for reproducibility  Accepted results are then published to the SPEC website  Results can be released publicly without prior committee review – however, if asked, full disclosure must be provided to SPEC 14

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Reporting Run and Reporting Rules  The SPEC SFS 2014 Run and Reporting Rules bound the measurement and configuration methodology  Primary goal of rules is to support SPEC’s philosophy of fair and open benchmarking  Secondary goal is to ensure sufficient disclosure for reproducibility and comparability 15

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Reporting Run and Reporting Rules Highlights  There is no Uniform Access Rule  The WARMUP time may be set to between 5 minutes and 1 week for a publishable run  There is no requirement to reinitialize file systems before a publishable run  However, detailed documentation of actions taken since system (re)initialization is required  Single publication covers 1 workload  No requirement that all or more than one be reported at the same time 16

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Reporting No “newfs” Requirement  Re-initializing the storage under the file system may not be possible or realistic  Cloud storage, complex tiered storage  More than one file system in the storage hierarchy  Must document procedures and steps taken since last re-initialization  Must be generally available and recommended for customers – no “benchmark specials”  Documentation/review allows for reproducibility  Can be used to simulate “aged” systems  Especially in conjunction with long WARMUP 17

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Defining a Workload 18

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Workload Definition  Workloads are richly defined in SPEC SFS 2014  Separate I/O size distributions for reads/writes  Each has 16 buckets; each bucket can be a range  Min I/O size: 1 byte; Max I/O size: size_t  22 file operations available to define workload  “Data”  Read/write ops: sequential, random, whole file, memory mapped  Read-modify-write, copyfile, append  “Metadata”  POSIX file ops: mkdir, stat, rename, chmod, etc. 19

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Workload Definition  Three parameters to control write behavior  % Write commits, % O_DIRECT, % O_SYNC  The method used for direct io is OS dependent.  Other parameters to change workload and dataset behavior, such as  % Geometric – certain files will be accessed more  % Compress – compressibility of the dataset  The dataset produced by SPEC SFS 2014 is not designed to be dedupable 20

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Business Metric Definition  A business metric is a unit of workload, made of:  One or more component workloads  Execution parameters  Success criteria (thresholds)  Why business metrics?  Simulating real-world workloads  Reporting results in real-world language  Success criteria attach more meaning to results than just a load level: quality 21

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Business Metric Scaling  The definition of a single business metric is fixed  Discrete and independent units of workload  Load scaling is achieved by adding additional business metrics  As load increases, so does  Proc count  Dataset size  The oprate of each proc is constant, however! 22

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Business Metric Success Criteria  Business metric success criteria (thresholds)  Global oprate threshold monitors the average oprate of all procs  Proc oprate threshold monitors the oprate of all procs  Any single proc exceeding the threshold invalidates that load point  Achieved oprate must be >= x% of defined 23

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Business Metric Success Criteria  Business metric success criteria (thresholds)  Workload variance threshold monitors ratio of global achieved oprates between all component workloads  This ratio must be within +/- x%, as defined in the threshold  Example: DATABASE has a 5:1 oprate ratio between the DB_TABLE and DB_LOG component workloads  Ratio of achieved oprates must be within +/- 5% of 5:1 24

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Business Metric Success Criteria  Business metric success criteria (thresholds)  With these success criteria, a business metric demands a certain quality of service at all load points  If a success criteria is not met for a requested load point, that point is marked INVALID  An INVALID data point does not stop the benchmark run, but is not publishable 25

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Benchmark Results  There are two principal measures of performance in SPEC SFS 2014  Business Metrics  Overall Response Time  Achieved Oprate and Total KBps will be included in official publications as well  The sfssum file produced during a benchmark run contains all this info and more 26

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Benchmark Results  Disclosure of results must include the summary result  Maximum achieved business metrics and overall response time of the entire benchmark run  Specific format in Run and Reporting Rules  The full disclosure report is published on the SPEC website  Visual: Business Metrics vs. Response Time 27

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Benchmark Results 28

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Benchmark Results 29

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Benchmark Results 30

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Workloads and Business Metrics Overall Response Time  Overall response time is calculated differently in SPEC SFS 2014  Still the area under the curve divided by the maximum achieved business metric  Origin point (0,0) is no longer assumed  First point used in the calculation is the first achieved result  No longer seems appropriate to assume the curve will be a certain shape 31

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads 32

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Video Data Acquisition (VDA)  Simulates acquisition of data from a temporally volatile source (surveillance, big data ingest)  Metric: Concurrent STREAMS  Workload derived from IBM Redbooks  Two component workloads, 9:1 oprate ratio  VDA1, data stream  ~36 Mb/sec sequential writes (upper range of HD video)  VDA2, companion applications/user access  89% read, 2% read-modify-write, 9% metadata 33

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Video Data Acquisition (VDA)  VDA2 workload ensures that quality of data ingestion is maintained despite other activity  Starvation of reads or writes will be detected by success criteria violation  Per-proc oprate: >= 75% of requested  Overall oprate: >= 95% of requested  Component workload variance: <= 5% of defined 34

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Virtual Desktop Infrastructure (VDI)  Simulates the workload generated by a hypervisor to support a heavy steady-state knowledge worker workload  Workload derived from traces of ESXi, Hyper- V, and Xen environments  Metric: concurrent DESKTOPS  One component workload, 2 procs per desktop  Data-heavy workload: 1% metadata ops 35

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Virtual Desktop Infrastructure (VDI)  Simulates steady-state VDI workload  Does not include boot storm or login storm  All writes use Direct I/O  Dataset consists of compressible (60%) large files (500MB)  Dataset is not dedupable – simulates a VDI scenario using Full Clones 36

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Software Build (SWBUILD)  Simulates large software project compilation or build phase of an EDA workflow  Workload derived from traces taken during software build activity and ClearCase documentation  Metric: concurrent BUILDS  One component workload, 5 procs per build  Metadata-heavy: 87% metadata ops 37

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Software Build (SWBUILD)  Reads and writes are done on a whole file  Average file size is a Gaussian distribution centered at 16 KiB, ~573,000 files per build  Files are highly compressible (80%)  This workload has the most potential to be cached/modified by the load generating clients  Also most likely to introduce/measure a bottleneck on load generating clients vs. storage solution 38

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads DATABASE  Simulates an OLTP database consolidation scenario  Workload derived from data from Oracle  Metric: concurrent DATABASES  Two component workloads, 5:1 oprate ratio  DB_TABLE  Random reads and writes  DB_LOG  Mostly sequential writes 39

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads DATABASE  All DB_TABLE threads for an individual business metric share the same dataset  Multiple threads working the same tables  Workload simulates moving hot spots in the dataset  These hot spots move over time  Solution under test must provide good quality of service to both table and log I/O  Maximum component workload variance is <= 5% 40

SPEC SFS 2014 Read Size Distribution

SPEC SFS 2014 Write Size Distribution

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. SPEC SFS 2014 Workloads Summary  SPEC SFS 2014 Workloads are richly-defined, realistic, solution-based workloads  Results are measured in Business Metrics  Real-world language for real-world workloads  Quality of service is measured with success criteria  Performance is measured at the application level  Performance of whole solution is measured  Modern scenarios based on standard solutions  Workload definitions and source available to all SPEC SFS 2014 licensees  Open, transparent, and fair benchmarking 43

2015 File and Storage Technologies. ©Standard Performance Evaluation Corporation. All Rights Reserved. Future Investigations  More Workloads  Windows Homefolders (a.k.a. FSCT)  HPC  Movie Production  Video Distribution  Support More Storage APIs  Block Device  MPI-IO  HDFS  CDMI  Energy Efficiency  Work with SNIA and the EPA. Energy Star standard  Power Measurement  Source code continues to be provided for everything as SPEC maintains openness and transparency 44