CyberShake Study 2.2: Computational Review Scott Callaghan 1.

Slides:



Advertisements
Similar presentations
LIBRA: Lightweight Data Skew Mitigation in MapReduce
Advertisements

Building a Distributed Full-Text Index for the Web S. Melnik, S. Raghavan, B.Yang, H. Garcia-Molina.
SALSA HPC Group School of Informatics and Computing Indiana University.
1 High Performance Computing at SCEC Scott Callaghan Southern California Earthquake Center University of Southern California.
1 External Sorting for Query Processing Yanlei Diao UMass Amherst Feb 27, 2007 Slides Courtesy of R. Ramakrishnan and J. Gehrke.
Using TOSCA Requirements /Capabilities Monitoring Use Case (Primer Considerations) Proposal by CA Technologies, IBM, SAP, Vnomic.
CyberShake Study 14.2 Technical Readiness Review.
Overview of Broadband Platform Software as used in SWUS Project Philip Maechling BBP Modelers Meeting 12 June 2013.
Authors: Weiwei Chen, Ewa Deelman 9th International Conference on Parallel Processing and Applied Mathmatics 1.
Getting business done Information architecture ControlGovernance Operational Costs User Empowerment IT Concerns.
Descriptive Data Analysis of File Transfer Data Sudarshan Srinivasan Victor Hazlewood Gregory D. Peterson.
CYBERINFRASTRUCTURE FOR THE GEOSCIENCES High Performance Computing applications in GEON: From Design to Production Dogan Seber.
NSF Geoinformatics Project (Sept 2012 – August 2014) Geoinformatics: Community Computational Platforms for Developing Three-Dimensional Models of Earth.
Introduction to Apache OODT Yang Li Mar 9, What is OODT Object Oriented Data Technology Science data management Archiving Systems that span scientific.
Sep 21, 20101/14 LSST Simulations on OSG Sep 21, 2010 Gabriele Garzoglio for the OSG Task Force on LSST Computing Division, Fermilab Overview OSG Engagement.
Describe workflows used to maintain and provide the RDA to users – Both are 24x7 operations Transition to the NWSC with zero downtime NWSC is new environment.
SALSA HPC Group School of Informatics and Computing Indiana University.
PetaApps: Update on software engineering and performance J. Dennis M. Vertenstein N. Hearn.
CyberShake Study 15.4 Technical Readiness Review.
CyberShake Study 2.3 Technical Readiness Review. Study re-versioning SCEC software uses year.month versioning Suggest renaming this study to 13.4.
Fig. 1. A wiring diagram for the SCEC computational pathways of earthquake system science (left) and large-scale calculations exemplifying each of the.
The Million Point PI System – PI Server 3.4 The Million Point PI System PI Server 3.4 Jon Peterson Rulik Perla Denis Vacher.
CyberShake Study 15.3 Science Readiness Review. Study 15.3 Scientific Goals Calculate a 1 Hz map of Southern California Produce meaningful 2 second results.
ApproxHadoop Bringing Approximations to MapReduce Frameworks
CS 351/ IT 351 Modeling and Simulation Technologies HPC Architectures Dr. Jim Holten.
Southern California Earthquake Center CyberShake Progress Update 3 November 2014 through 4 May 2015 UGMS May 2015 Meeting Philip Maechling SCEC IT Architect.
Experiences Running Seismic Hazard Workflows Scott Callaghan Southern California Earthquake Center University of Southern California SC13 Workflow BoF.
UCERF3 Uniform California Earthquake Rupture Forecast (UCERF3) 14 Full-3D tomographic model CVM-S4.26 of S. California 2 CyberShake 14.2 seismic hazard.
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
Southern California Earthquake Center CyberShake Progress Update November 3, 2014 – 4 May 2015 UGMS May 2015 Meeting Philip Maechling SCEC IT Architect.
1 USC Information Sciences InstituteYolanda Gil AAAI-08 Tutorial July 13, 2008 Part IV Workflow Mapping and Execution in Pegasus (Thanks.
Introduction to Performance Tuning Chia-heng Tu PAS Lab Summer Workshop 2009 June 30,
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
SCEC CyberShake on TG & OSG: Options and Experiments Allan Espinosa*°, Daniel S. Katz*, Michael Wilde*, Ian Foster*°,
Compute and Storage For the Farm at Jlab
Overview of Scientific Workflows: Why Use Them?
CyberShake Study 2.3 Readiness Review
CyberShake Study 16.9 Science Readiness Review
Simplify Your Science with Workflow Tools
Pegasus WMS Extends DAGMan to the grid world
SharePoint Solutions Architect, Protiviti
Simplify Your Science with Workflow Tools
Seismic Hazard Analysis Using Distributed Workflows
Predesign.
Scott Callaghan Southern California Earthquake Center
Workflows and the Pegasus Workflow Management System
CyberShake Study 16.9 Discussion
Job workflow Pre production operations:
Philip J. Maechling (SCEC) September 13, 2015
Haiyan Meng and Douglas Thain
University of Southern California
Moodle Scalability What is Scalability?
National Center for Atmospheric Research
CyberShake Study 17.3 Science Readiness Review
CS222P: Principles of Data Management UCI, Fall 2018 Notes #09 External Sorting Instructor: Chen Li.
Pegasus Workflows on XSEDE
CS222: Principles of Data Management Lecture #10 External Sorting
CyberShake Study 2.2: Science Review Scott Callaghan 1.
Overview of Workflows: Why Use Them?
CS222P: Principles of Data Management Lecture #10 External Sorting
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
CyberShake Study 14.2 Science Readiness Review
Frieda meets Pegasus-WMS
Southern California Earthquake Center
CyberShake Study 18.8 Technical Readiness Review
CS222/CS122C: Principles of Data Management UCI, Fall 2018 Notes #09 External Sorting Instructor: Chen Li.
Production Manager Tools (New Architecture)
CyberShake Study 18.8 Planning
Best Practices in Higher Education Student Data Warehousing Forum
Presentation transcript:

CyberShake Study 2.2: Computational Review Scott Callaghan 1

Computational Goals 269 CyberShake sites on Kraken with existing SGTs 47 complete, 221 remaining, 1 lost Produce seismograms, PSA values, hazard curves Establish Kraken / Cray architecture as platform for CyberShake

Study Goal Map

Inputs 221 sets of SGTs + MD5 sums Rupture Geometries 5 on HPC 184 on Ranger disk 31 on Ranch archive Will need to be staged back to Ranger About 8.5 TB Rupture Geometries 14,000 files 1.5 GB

Outputs Files Database Access 116 M seismograms, 116 M PSA files 2.5 TB (2.1 TB new) 350,000 workflow files 1.3 TB (1.1 TB new) Small number of curves, maps Database 350 M entries (37% increase) About 40 GB Access Hazard curves, maps posted on web site PSA values in DB Seismograms on disk

Computing Environment/Resources Kraken nodes Pegasus 4.2.0 + PMC SGT extraction code Memcached In-memory rupture variation generation Seismogram/PSA code Combined CyberShake codes tagged in SVN

Computing Resources 1.2M Kraken SUs, ~5M SUs available Local disk space 3.2 TB (additional) required 4.8 TB available on scec-02 Duration Start 10/8 (with review approval) ~2 months (dependent on Kraken queue, I/O) Personnel Scott Request help from Pegasus group when needed

Reproducibility Science code tagged in SVN Metadata captured in database SGTs long-term Ranger decommissioned in Feb Either archive or throw away and regenerate

Metrics Calculate metrics previously highlighted in papers and posters, especially: Average makespan Parallel speedup Utilization Tasks/sec Delay per job SI2 metrics Number of hazard curves Compare metrics, determine improvement

Open Issues / Risk Analysis Kraken I/O Depending on file system performance, runtime is variable by a factor of 3 Kraken gridmanager Support for load? SUs Uncertain about usage of other SCEC users Statistics gathering Have had issues with pegasus-monitord in the past May have to populate DB after workflow is complete