Penn State, August 2013 Cloud-WIEN2k A Scientific Cloud Computing Platform for Condensed Matter Physics K. Jorissen University of Washington, Seattle,

Slides:



Advertisements
Similar presentations
SLA-Oriented Resource Provisioning for Cloud Computing
Advertisements

Amazon Web Services Justin DeBrabant CIS Advanced Systems - Fall 2013.
SALSA HPC Group School of Informatics and Computing Indiana University.
Supporting Simulations on the Cloud using Workflows & Virtual Machines Gary Polhill Macaulay Land Use Research Institute Edoardo Pignotti Computing Science,
1. Topics Is Cloud Computing the way to go? ARC ABM Review Configuration Basics Setting up the ARC Cloud-Based ABM Hardware Configuration Software Configuration.
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
Implementing a menu based application in FutureGrid
Introduction CS 524 – High-Performance Computing.
S oftware T ools for A cademics and R esearchers MASSACHUSETTS INSTITUTE OF TECHNOLOGY Office of Educational Innovation and Technology.
Virtual Machines for HPC Paul Lu, Cam Macdonell Dept of Computing Science.
High Performance Computing (HPC) at Center for Information Communication and Technology in UTM.
Presenter: Joshan V John Robert Dyer, Hoan Anh Nguyen, Hridesh Rajan & Tien N. Nguyen Iowa State University, USA Instructor: Christoph Csallner 1 Joshan.
Cambodia-India Entrepreneurship Development Centre - : :.... :-:-
Installing software on personal computer
Amazon EC2 Quick Start adapted from EC2_GetStarted.html.
System Center 2012 Setup The components of system center App Controller Data Protection Manager Operations Manager Orchestrator Service.
Customized cloud platform for computing on your terms !
Accessing the Amazon Elastic Compute Cloud (EC2) Angadh Singh Jerome Braun.
J.J.Rehr1, John Vinson1, E.L.Shirley2 J.J. Kas1 and F. Vila1
Tools and Utilities for parallel and serial codes in ENEA-GRID environment CRESCO Project: Salvatore Raia SubProject I.2 C.R. ENEA-Portici. 11/12/2007.
The Pipeline Processing Framework LSST Applications Meeting IPAC Feb. 19, 2008 Raymond Plante National Center for Supercomputing Applications.
NSF PI Meeting: The Science of the Cloud, Mar 17-18, 2011 Waterview Conference Center, Arlington, VA Cloud Computing Clusters for Scientific Research*
Introduction to Cloud Computing
Chapter 2: Operating-System Structures. 2.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts – 7 th Edition, Jan 14, 2005 Operating System.
Future role of DMR in Cyber Infrastructure D. Ceperley NCSA, University of Illinois Urbana-Champaign N.B. All views expressed are my own.
Using Biological Cyberinfrastructure Scaling Science and People: Applications in Data Storage, HPC, Cloud Analysis, and Bioinformatics Training Scaling.
Cloud Computing & Amazon Web Services – EC2 Arpita Patel Software Engineer.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
BLU-ICE and the Distributed Control System Constraints for Software Development Strategies Timothy M. McPhillips Stanford Synchrotron Radiation Laboratory.
O.S.C.A.R. Cluster Installation. O.S.C.A.R O.S.C.A.R. Open Source Cluster Application Resource Latest Version: 2.2 ( March, 2003 )
J. J. Rehr & R.C. Albers Rev. Mod. Phys. 72, 621 (2000) A “cluster to cloud” story: Naturally parallel Each CPU calculates a few points in the energy grid.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Graduate Student Department Of CSE 1.
Ch 1. A Python Q&A Session Spring Why do people use Python? Software quality Developer productivity Program portability Support libraries Component.
Connecting Experiment and Theory across Length and Time-scales Algorithms and Software for Materials Research C yber I nfrastructure J. J. Rehr Department.
Cyber-Infrastructure for Materials Simulation at the Petascale Trends/opportunities Recommendations for CI support David Ceperley National Center for Supercomputing.
SALSA HPC Group School of Informatics and Computing Indiana University.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Nanco: a large HPC cluster for RBNI (Russell Berrie Nanotechnology Institute) Anne Weill – Zrahia Technion,Computer Center October 2008.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Enabling the use of e-Infrastructures with.
By Chi-Chang Chen.  Cluster computing is a technique of linking two or more computers into a network (usually through a local area network) in order.
Application Programming in Cloud via Swift Swift Tutorial, CCGrid 2013, Hour 2 Ketan Maheshwari.
OPTIMIZATION OF DIESEL INJECTION USING GRID COMPUTING Miguel Caballer Universidad Politécnica de Valencia.
Lesson 1 1 LESSON 1 l Background information l Introduction to Java Introduction and a Taste of Java.
1 Circuitscape Capstone Presentation Team Circuitscape Katie Rankin Mike Schulte Carl Reniker Sean Collins.
Selenium server By, Kartikeya Rastogi Mayur Sapre Mosheca. R
Phonon dispersion calculation
HPC need and potential of ANSYS CFD and mechanical products at CERN A. Rakai EN-CV-PJ2 5/4/2016.
INTRODUCTION TO HIGH PERFORMANCE COMPUTING AND TERMINOLOGY.
Agenda  What is Cloud Computing?  Milestone of Cloud Computing  Common Attributes of Cloud Computing  Cloud Service Layers  Cloud Implementation.
Accessing the VI-SEEM infrastructure
VisIt Project Overview
Dr.S.Sridhar, Director, RVCT, RVCE, Bangalore
Tools and Services Workshop
Virtualisation for NA49/NA61
Dag Toppe Larsen UiB/CERN CERN,
Joslynn Lee – Data Science Educator
Dag Toppe Larsen UiB/CERN CERN,
ECRG High-Performance Computing Seminar
GdX - Grid eXplorer parXXL: A Fine Grained Development Environment on Coarse Grained Architectures PARA 2006 – UMEǺ Jens Gustedt - Stéphane Vialle - Amelia.
Virtualisation for NA49/NA61
Dr.S.Sridhar, Director, RVCT, RVCE, Bangalore
CRESCO Project: Salvatore Raia
Containers in HPC By Raja.
Haiyan Meng and Douglas Thain
Development of the Nanoconfinement Science Gateway
Cornell Theory Center Cornell Theory Center (CTC) is a high-performance computing and interdisciplinary research center at Cornell.
Brandon Hixon Jonathan Moore
Using and Building Infrastructure Clouds for Science
CMPE 135: Object-Oriented Analysis and Design March 14 Class Meeting
Presentation transcript:

Penn State, August 2013 Cloud-WIEN2k A Scientific Cloud Computing Platform for Condensed Matter Physics K. Jorissen University of Washington, Seattle, U.S.A. Supported by NSF grant OCI

Materials Science Materials Science research: Theoretical models, evaluated on a computer, are usually needed for interpretation and quantification of measurements. But HPC is often not readily available. sample measurement theoretical model interpretation Hψ=Eψ E=mc 2

Anecdote (High-Performance Computing is everywhere) Computational linguistics: “We automatically identify semantically related words in the 400 million word Dutch Twente corpus to  Statistically find contextual associations and quantify association strength  Identify syntactical relations between words  Relevant to automatic translation software Multivariate analysis with dozens of variables – large computational needs.” --- an “English Lit major”

Quest How do we bring the best theory and simulations to the scientists who need it? (often applied scientists – not computational specialists) SOLUTION: Scientific Cloud Computing

FEFF-old (simple Einstein model for phonons) Are state-of-the-art calculations “work for specialists”? GUI Easy install Runs on laptop Load file & Click “Run” ~ 1 day to learn

FEFF-gold (accurate ab initio model for phonons) Are state-of-the-art calculations “work for specialists”? DFT requires cluster Configure || codes Complex workflow Command-line ~ 0.x grad students to learn Dynamical Matrix (DFT) -- ABINIT Debye Waller Factors -- DMDW X-ray Absorption -- FEFF Invented / published Clearly an improvement Nobody uses it

Hardware barrier: advanced codes need clusters Software barrier: running codes is difficult t >> 1 before improved theory reaches applied research Are state-of-the-art calculations “work for specialists”? -Buy a cluster? IT support? -Supercomputing center? -Collaborate with specialists? -Installation of || software tricky -lacking user-friendliness -multi-code workflows difficult

Scientific Cloud Computing Interface simplifies workflow (hides cloud -- app) Developer makes virtual “XAS” compute node with preinstalled WIEN2k User requests 5 node Cloud Cluster for 3 hours when needed ($20)

SCC Virtual Machine Image Contains utilities for parallel scientific computing: MPI, compilers, libraries, NFS, … Becomes compute node in SCC Cloud Cluster Developer-optimized Scientific codes for your research - WIEN2k for electronic structure calculations - latest version - optimized for performance - MPI parallellization for large calculations “My new research group was looking for a way to implement MEEP-mpi (MIT Electromagnetic Equation Propagation) to simulate EM fields in nanoscale optical devices for cavity QED experiments. We believe that Amazon EC2 is an economical and time saving solution for our finite difference time domain (FDTD) simulations. My group‘s research iterates between fabrication and simulation thus it is advantageous to buy computing power only when needed. Moreover it is a relief not to have to maintain our own small cluster within our group.” Kai-Mei Fu, University of Washington (USA)

SCC Java interface FEFF GUI For developers of GUIs Java interface library (jar) For savvy users and developers Collection of shell scripts SCC Linux interface

WIEN2k GUI (DFT) WIEN2k-cloud Starts || cluster in EC2 cloud Uploads initialized calculation Runs || calculation in EC2 cloud Downloads results to laptop Deletes EC2 cluster Other workflows / data flows can be added. Requires: -create EC2 account -install SCC program

Performance LOOSELY Coupled Processes Good scaling DFT KS equations on 128 k-point grid

Performance HPC cluster instances deliver good speedup KS for large system at 1 k-point TIGHTLY Coupled Processes VERY DEMANDING of network performance

5. WIEN2k Performance Benchmarks 1200 atom unit cell; SCALAPACK+MPI diagonalization, matrix size 50k-100k HPC cluster instances deliver similar speedup as local Infiniband cluster KS for large system at 1 k-point TIGHTLY Coupled Processes VERY DEMANDING of network performance H size 56,000 (25GB) Runtime (16x8 processors) : Local (Infiniband) 3h:48 Cloud (10Gbps) 1h:30 ($40)

Sa Sdf “Scientific Cloud Computing can bring novel theory & HPC modeling to more researchers.” We acknowledge:  FEFF: S. Story T. Ahmed B. Mattern M. Prange J. Vinson  UW: R. Coffey E. Lazowska J. Loudermilk  Amazon: D. Singh  NSF:C. Bouldin  supported by NSF OCI Comp. Phys. Comm. 183 (2012)

Backup stuff

1.Create cluster 2.Calculations 3.Stop cluster Cloud Compute Instances MPI Master FEFF9 MPI Slave FEFF9 MPI Slave FEFF9 My Laptop FEFF interface 4. Cloud-Computing on the Amazon EC2 cloud * K. Jorissen et al., Comp. Phys. Comm. 183 (2012) Create cluster 2.Calculations 3.Stop cluster 1.Create cluster 2.Calculations 3.Stop cluster

import edu.washington.scc.*; // Launch the new cluster with “cs” specifications: ClusterResult rl = clust.launch(cs); // Initialize the FEFF calculation on the cloud cluster: // Copy feff.inp: ClusterResult rp = clust.put(LocalWorkingDir+"/feff.inp", CloudWorkingDir+"/feff.inp"); // Run the FEFF9-MPI calculation: ClusterResult rf9 = clust.executeCommand(Feff9CommandLine,CloudOut); // Copy the output files back to the local computer: ClusterResult rg = clust.get(CloudWorkingDir, LocalWorkingDir); // Terminate the cloud cluster: ClusterResult rt = clust.terminate(); Developer’s view: ExecuteCloudContext.java:

End User’s view: FEFF GUI: