NSF PI Meeting: The Science of the Cloud, Mar 17-18, 2011 Waterview Conference Center, Arlington, VA Cloud Computing Clusters for Scientific Research*

Slides:



Advertisements
Similar presentations
Cloud Computing Mick Watson Director of ARK-Genomics The Roslin Institute.
Advertisements

SLA-Oriented Resource Provisioning for Cloud Computing
Amazon Web Services Justin DeBrabant CIS Advanced Systems - Fall 2013.
SALSA HPC Group School of Informatics and Computing Indiana University.
Beowulf Supercomputer System Lee, Jung won CS843.
Low Cost, Scalable Proteomics Data Analysis Using Amazon's Cloud Computing Services and Open Source Search Algorithms Brian D. Halligan, Ph.D. Medical.
Parallel Computation of the 2D Laminar Axisymmetric Coflow Nonpremixed Flames Qingan Andy Zhang PhD Candidate Department of Mechanical and Industrial Engineering.
Chapter 1: Introduction
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems.
Matt Bertrand Building GIS Apps in the Cloud. Infrastructure - Provides computer infrastructure, typically a platform virtualization environment, as a.
Nikolay Tomitov Technical Trainer SoftAcad.bg.  What are Amazon Web services (AWS) ?  What’s cool when developing with AWS ?  Architecture of AWS 
1/16/2008CSCI 315 Operating Systems Design1 Introduction Notice: The slides for this lecture have been largely based on those accompanying the textbook.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 4.
Topics Problem Statement Define the problem Significance in context of the course Key Concepts Cloud Computing Spatial Cloud Computing Major Contributions.
Parallel Architectures Dr Andy Evans With additions from Dr Nick Malleson.
MATE-EC2: A Middleware for Processing Data with Amazon Web Services Tekin Bicer David Chiu* and Gagan Agrawal Department of Compute Science and Engineering.
Cloud computing Tahani aljehani.
High Performance Computing with cloud Xu Tong. About the topic Why HPC(high performance computing) used on cloud What’s the difference between cloud and.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Overview of Atmosphere.
Amazon EC2 Quick Start adapted from EC2_GetStarted.html.
U.S. Department of the Interior U.S. Geological Survey David V. Hill, Information Dynamics, Contractor to USGS/EROS 12/08/2011 Satellite Image Processing.
A Brief Overview by Aditya Dutt March 18 th ’ Aditya Inc.
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Parallel Computing The Bad News –Hardware is not getting faster fast enough –Too many architectures –Existing architectures are too specific –Programs.
STRATEGIES INVOLVED IN REMOTE COMPUTATION
Dual Stack Virtualization: Consolidating HPC and commodity workloads in the cloud Brian Kocoloski, Jiannan Ouyang, Jack Lange University of Pittsburgh.
‘Tis not folly to dream: Using Molecular Dynamics to Solve Problems in Chemistry Christopher Adam Hixson and Ralph A. Wheeler Dept. of Chemistry and Biochemistry,
Marcelo de Paiva Guimarães Bruno Barberi Gnecco Marcelo Knorich Zuffo
Accessing the Amazon Elastic Compute Cloud (EC2) Angadh Singh Jerome Braun.
J.J.Rehr1, John Vinson1, E.L.Shirley2 J.J. Kas1 and F. Vila1
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems.
Extreme scale parallel and distributed systems – High performance computing systems Current No. 1 supercomputer Tianhe-2 at petaflops Pushing toward.
Cloud Computing & Amazon Web Services – EC2 Arpita Patel Software Engineer.
 H.M.BILAL Operating System Concepts.  What is an Operating System?  Mainframe Systems  Desktop Systems  Multiprocessor Systems  Distributed Systems.
Penn State, August 2013 Cloud-WIEN2k A Scientific Cloud Computing Platform for Condensed Matter Physics K. Jorissen University of Washington, Seattle,
J. J. Rehr & R.C. Albers Rev. Mod. Phys. 72, 621 (2000) A “cluster to cloud” story: Naturally parallel Each CPU calculates a few points in the energy grid.
Infrastructure Clouds MAGIC Meeting Chris Hill, MIT September 7 th 2011 interested in modeling of Earth and planetary systems to both understand basic.
Connecting Experiment and Theory across Length and Time-scales Algorithms and Software for Materials Research C yber I nfrastructure J. J. Rehr Department.
April 26, CSE8380 Parallel and Distributed Processing Presentation Hong Yue Department of Computer Science & Engineering Southern Methodist University.
SALSA HPC Group School of Informatics and Computing Indiana University.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Enabling the use of e-Infrastructures with.
Parallelizing Spacetime Discontinuous Galerkin Methods Jonathan Booth University of Illinois at Urbana/Champaign In conjunction with: L. Kale, R. Haber,
Silberschatz and Galvin  Operating System Concepts Module 1: Introduction What is an operating system? Simple Batch Systems Multiprogramming.
CIP HPC CIP - HPC HPC = High Performance Computer It’s not a regular computer, it’s bigger, faster, more powerful, and more.
Evaluating Clouds for Smart Grid Computing: early Results using GE MARS App Ketan Maheshwari
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
© 2015 MetricStream, Inc. All Rights Reserved. AWS server provisioning © 2015 MetricStream, Inc. All Rights Reserved. By, Srikanth K & Rohit.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Multicore Applications in Physics and Biochemical Research Hristo Iliev Faculty of Physics Sofia University “St. Kliment Ohridski” 3 rd Balkan Conference.
Amazon Web Services. Amazon Web Services (AWS) - robust, scalable and affordable infrastructure for cloud computing. This session is about:
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Chapter 1: Introduction What is an Operating System? Mainframe Systems Desktop Systems Multiprocessor Systems Distributed Systems Clustered System Real.
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Elastic Cyberinfrastructure for Research Computing
Introduction to Parallel Computing: MPI, OpenMP and Hybrid Programming
ECRG High-Performance Computing Seminar
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME Outreach SESAME,
Tools and Services Workshop Overview of Atmosphere
Chapter 1: Introduction
Biology MDS and Clustering Results
Haiyan Meng and Douglas Thain
Different types of Linux installation
Presentation transcript:

NSF PI Meeting: The Science of the Cloud, Mar 17-18, 2011 Waterview Conference Center, Arlington, VA Cloud Computing Clusters for Scientific Research* J. J. Rehr, F. D. Vila, and K. Jorissen *Supported by NSF SI2-SSE Grant OCI Department of Physics, University of Washington Seattle, WA

Complete scientific computing cloud environment for materials science; robust, easy to use Challenge of NSF Grant Goals of CC for Scientific Computing* On-demand access without the need to purchase, maintain, or understand HPCs Optimized/pre-installed HPC scientific codes Low cost HPC access to a wide class of scientists *J. J. Rehr, F. Vila, J. P. Gardner, L. Svec and M. Prange CiSE May/June 2010

The Target Scientific CC User Not usually HPC “gurus”: –Little experience with parallel computing, code compilation & optimization, etc. –Limited access to HPC resources “Would rather be doing science...”

Scientific Cloud Computing Programming Philosophy 1.Develop AMIs (Amazon Machine Images) for HPC scientific applications in materials science 2.Develop shell-scripts that make the EC2 look & run like a local HPC cluster: ” virtual supercomputer on a laptop ” 3.Test serial, parallel & network performance 4.Develop GUI to control/run codes & I/O Development Strategy Preserve current HPC paradigm (MPI) on the cloud Simplify access to HPC for a broad user base

1.Start cluster 2.Configure nodes Cloud → Cluster with UW Cluster tools* EC2 Compute Instances User interacts with control workstation MPI Master startd MPI Slave startd MPI Slave startd Control Workstation ec2-clust-* *

UW EC2 Cluster Tools UW EC2 tools on local control machine NameFunctionAnalog ec2-clust-launch N Launches cluster with N instances boot ec2-clust-connect Connect to a cluster ssh ec2-clust-put Transfer data to EC2 cluster scp ec2-clust-get Transfer data from EC2 cluster scp ec2-clust-list List running clusters ec2-clust-terminate Terminate a running cluster shutdown The tools hide a lot of the “ugliness”: ec2-clust-connect ssh -i/home/fer/.ec2_clust/.ec2_clust_info.7729.r-de70cdb7/key_ pair_fdv.pem

Current EC2 AMIs New 32-bit and 64-bit AMIs Amazon’s HPC “cluster instance” EBS backed instances (faster disk access/boot time) Several Condensed Matter/Materials Science Codes: X-ray Spectroscopy: FEFF9, IFEFFIT Electronic Structure: WIEN2k, ABINIT, Quantum Espresso Excited States:AI2NBSE, OCEAN, Exc!ting, RT-SIESTA

UW CC Tools: 2 nd generation (v2) Older v1 tools : Emphasized simplicity & maximum security SLOW: 12 min for n inst =16, linear with cluster size v2 Solutions: Asynchronous tools: minimal setup, max. security Self-Assembling clusters: fast, but sacrifices security Best Solution: Hybrid approach; non-sensitive tasks on cloud keeps credentials safe

UW v2 Tools: Boot and Setup Times Setup time negligible & independent of cluster size! EBS backed clusters faster than S3

J. J. Rehr & R.C. Albers Rev. Mod. Phys. 72, 621 (2000) Sample Scientific Application FEFF: Real-space Green’s function code for electronic structure, x-ray spectra, EELS, … User Base: 1000s of users in physics, chemistry, materials science, biophysics…

FEFF Parallel Performance FEFF9 shows identical speedup on local cluster and EC2 cluster Naturally parallel:Each CPU calculates a few points in the energy grid Loosely coupled:Little communication between processes

Sample Electronic Structure Application WIEN2k Parallel Performance EC2 small312s EC2 medium116s EC2 HPC 84s UW Opteron 189s Time to solve one k-point in one core Good scaling in regular instances DFT KS equations on 128 k-point grid LOOSELY Coupled Processes

Regular instances cannot deliver network-intensive performance HPC cluster instances deliver the same speedup as local Infiniband cluster KS for large system at 1 k-point VERY DEMANDING of network performance TIGHTLY Coupled Processes WIEN2k Parallel Performance

FEFF Java GUI: XAS on the Cloud

Future Project Goals Conclusions Develop & add MS Windows support Develop SCC Java GUI for CC management, runs, I/O Optimize & benchmark other advanced scientific codes Distribute to Scientific Community: Goals partially achieved: robust SCC environment developed & benchmarked on FEFF, WIEN2k SCC-ready FEFF App with JFEFF GUI (OSX, Linux) Improved SCC AMIs and UW Cluster tools (v2): Regular EC2 AMIs OK for light network use HPC AMIs NEEDED for MPI+ScaLAPACK

NSF PI Meeting: Science of the Cloud, Arlington, VA 2011 Acknowledgments... and thank you UW:L. Svec, J. Loudermilk, D. Bitseff. M. Prange, J. Gardner, R. Coffey, E. Lazowska Amazon:D. Singh, J. Barr, T. Laxdal, P. Sirota, P. Sridharan, R. Valdez, and W. Vogels NSF:C. Bouldin UW-SCC is supported by NSF grant OCI FEFF is supported by DOE-BES grant DE-FG03-97ER45623