FutureGrid Overview July XSEDE UAB Meeting Geoffrey Fox

Slides:



Advertisements
Similar presentations
FutureGrid Overview NSF PI Science of Cloud Workshop Washington DC March Geoffrey Fox
Advertisements

Cosmic Issues and Analysis of External Comments on FutureGrid TG11 Salt Lake City July Geoffrey Fox
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
FutureGrid Image Repository: A Generic Catalog and Storage System for Heterogeneous Virtual Machine Images Javier Diaz, Gregor von Laszewski, Fugang Wang,
Overview Presented at OGF31 Salt Lake City, July 2011 Geoffrey Fox, Gregor von Laszewski, Renato Figueiredo Contact:
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid Geoffrey Fox, Andrew J. Younge, Gregor von Laszewski, Archit Kulshrestha, Fugang.
FutureGrid Summary TG’10 Pittsburgh BOF on New Compute Systems in the TeraGrid Pipeline August Geoffrey Fox
FutureGrid Summary FutureGrid User Advisory Board TG’10 Pittsburgh August Geoffrey Fox
Big Data and Clouds: Challenges and Opportunities NIST January Geoffrey Fox
Eucalyptus on FutureGrid: A case for Eucalyptus 3 Sharif Islam, Javier Diaz, Geoffrey Fox Gregor von Laszewski Indiana University.
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid Overview CTS Conference 2011 Philadelphia May Geoffrey Fox
FutureGrid SOIC Lightning Talk February Geoffrey Fox
Science of Cloud Computing Panel Cloud2011 Washington DC July Geoffrey Fox
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
FutureGrid Overview Geoffrey Fox
Future Grid FutureGrid Overview Dr. Speaker. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research on the future.
FutureGrid Overview Geoffrey Fox
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
Future Grid FutureGrid Overview Geoffrey Fox SC09 November
FutureGrid Overview Geoffrey Fox
FutureGrid Connection to Comet Testbed and On Ramp as a Service Geoffrey Fox Indiana University Infra structure.
FutureGrid Overview Geoffrey Fox
Some remarks on Use of Clouds to Support Long Tail of Science July XSEDE 2012 Chicago ILL July 2012 Geoffrey Fox.
Future Grid Future Grid All Hands Meeting Introduction Indianapolis October Geoffrey Fox
FutureGrid SOIC Lightning Talk February Geoffrey Fox
FutureGrid Cyberinfrastructure for Computational Research.
Building Effective CyberGIS: FutureGrid Marlon Pierce, Geoffrey Fox Indiana University.
RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor.
FutureGrid Computing Testbed as a Service Overview July Geoffrey Fox for FutureGrid Team
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
FutureGrid TeraGrid Science Advisory Board San Diego CA July Geoffrey Fox
FutureGrid Computing Testbed as a Service NSF Presentation NSF April Geoffrey Fox for FutureGrid Team
FutureGrid Overview Geoffrey Fox
Tutorial Presented at TG2011 Geoffrey Fox, Gregor von Laszewski, Renato Figueiredo, Kate Keahey, Andrew Younge Contact:
FutureGrid BOF Overview TG 11 Salt Lake City July Geoffrey Fox
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid PI: Geoffrey Fox*, CoPIs: Kate Keahey +, Warren Smith -, Jose Fortes #, Andrew.
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Recipes for Success with Big Data using FutureGrid Cloudmesh SDSC Exhibit Booth New Orleans Convention Center November Geoffrey Fox, Gregor von.
Science Applications on Clouds and FutureGrid June Cloud and Autonomic Computing Center Spring 2012 Workshop Cloud.
Future Grid Future Grid Overview. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research that will invent the future.
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
Grappling Cloud Infrastructure Services with a Generic Image Repository Javier Diaz Andrew J. Younge, Gregor von Laszewski, Fugang.
Big Data Workshop Summary Virtual School for Computational Science and Engineering July Geoffrey Fox
Lizhe Wang, Gregor von Laszewski, Jai Dayal, Thomas R. Furlani
Unit 3 Virtualization.
Private Public FG Network NID: Network Impairment Device
Digital Science Center II
Status and Challenges: January 2017
Introduction to Cloud Computing
Computing Testbeds as a Service or ? as an Instrument
NSF : CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science PI: Geoffrey C. Fox Software: MIDAS HPC-ABDS.
FutureGrid Computing Testbed as a Service
Digital Science Center Overview
FutureGrid Overview for VSCSE Summer School on Science Clouds
FutureGrid: a Grid Testbed
Versatile HPC: Comet Virtual Clusters for the Long Tail of Science SC17 Denver Colorado Comet Virtualization Team: Trevor Cooper, Dmitry Mishin, Christopher.
Sky Computing on FutureGrid and Grid’5000
FutureGrid Overview June HPC 2012 Cetraro, Italy Geoffrey Fox
1Indiana University, 2now Rutgers University
Clouds from FutureGrid’s Perspective
Gregor von Laszewski Indiana University
Department of Intelligent Systems Engineering
Using and Building Infrastructure Clouds for Science
Sky Computing on FutureGrid and Grid’5000
Cloud versus Cloud: How Will Cloud Computing Shape Our World?
Presentation transcript:

FutureGrid Overview July 15 2012 XSEDE UAB Meeting Geoffrey Fox gcf@indiana.edu http://www.infomall.org https://portal.futuregrid.org Director, Digital Science Center, Pervasive Technology Institute Associate Dean for Research and Graduate Studies,  School of Informatics and Computing Indiana University Bloomington

FutureGrid key Concepts I FutureGrid is an international testbed modeled on Grid5000 July 15 2012: 223 Projects, ~968 users Supporting international Computer Science and Computational Science research in cloud, grid and parallel computing (HPC) The FutureGrid testbed provides to its users: A flexible development and testing platform for middleware and application users looking at interoperability, functionality, performance or evaluation FutureGrid is user-customizable, accessed interactively and supports Grid, Cloud and HPC software with and without VM’s A rich education and teaching platform for classes See G. Fox, G. von Laszewski, J. Diaz, K. Keahey, J. Fortes, R. Figueiredo, S. Smallen, W. Smith, A. Grimshaw, FutureGrid - a reconfigurable testbed for Cloud, HPC and Grid Computing, Bookchapter – draft

FutureGrid key Concepts II Rather than loading images onto VM’s, FutureGrid supports Cloud, Grid and Parallel computing environments by provisioning software as needed onto “bare-metal” using (changing) package of open source tools Image library for MPI, OpenMP, MapReduce (Hadoop, (Dryad), Twister), gLite, Unicore, Globus, Xen, ScaleMP (distributed Shared Memory), Nimbus, Eucalyptus, OpenNebula, KVM, Windows ….. Either statically or dynamically Growth comes from users depositing novel images in library FutureGrid has ~4700 distributed cores with a dedicated network Image1 Image2 ImageN … Load Choose Run

FutureGrid Partners Indiana University (Architecture, core software, Support) San Diego Supercomputer Center at University of California San Diego (INCA, Monitoring) University of Chicago/Argonne National Labs (Nimbus) University of Florida (ViNE, Education and Outreach) University of Southern California Information Sciences (Pegasus to manage experiments) University of Tennessee Knoxville (Benchmarking) University of Texas at Austin/Texas Advanced Computing Center (Portal) University of Virginia (OGF, XSEDE Software stack) Center for Information Services and GWT-TUD from Technische Universtität Dresden. (VAMPIR) Red institutions have FutureGrid hardware

FutureGrid: a Grid/Cloud/HPC Testbed Private Public FG Network NID: Network Impairment Device 12TF Disk rich + GPU 512 cores

Secondary Storage (TB) Compute Hardware Name System type # CPUs # Cores TFLOPS Total RAM (GB) Secondary Storage (TB) Site Status india IBM iDataPlex 256 1024 11 3072 180 IU Operational alamo Dell PowerEdge 192 768 8 1152 30 TACC hotel 168 672 7 2016 120 UC sierra 2688 96 SDSC xray Cray XT5m 6 1344 foxtrot 64 2 24 UF Bravo Large Disk & memory 32 128 1.5 3072 (192GB per node) 192 (12 TB per Server) Delta Large Disk & memory With Tesla GPU’s 32 CPU 32 GPU’s 192+ 14336 GPU ? 9 1536 (192GB per node) TOTAL Cores 4384 Echo (ScaleMP) Large Disk & Memory 32 CPU 192 2 6144 IU On Order

Storage Hardware Support System Type Capacity (TB) File System Site Status Xanadu 360 180 NFS IU New System DDN 6620 120 GPFS UC SunFire x4170 96 ZFS SDSC Dell MD3000 30 TACC IBM 24 UF Substantial back up storage at IU: Data Capacitor and HPSS Support Traditional Drupal Portal with usual functions Traditional Ticket System System Admin and User facing support (small) Outreach group (small) Strong Systems Admin Collaboration with Software group

5 Use Types for FutureGrid (red latest) 223 approved projects (968 users) July 14 2012 USA, China, India, Pakistan, lots of European countries Industry, Government, Academia Training Education and Outreach (8%)(10%) Semester and short events; interesting outreach to HBCU Interoperability test-beds (3%)(2%) Grids and Clouds; Standards; from Open Grid Forum OGF Domain Science applications (31%)(26%) Life science highlighted (18%)(14%), Non Life Science (13%)(12%) Computer science (47%)(57%) Largest current category Computer Systems Evaluation (27%)(29%) XSEDE (TIS, TAS), OSG, EGI See Andrew Grimshaw’s discussion of XSEDE testing in Book Chapter

Some Training Education and Outreach Project Highlights See summary in Jerome Mitchell’s XSEDE12 paper and Renato Figueiredo’s BOF Tuesday Cloud Summer School July 30—August 3 with 10 HBCU attendees Mitchell and Younge building “Cloud Computing Handbook” loosely based on my book with Hwang and Dongarra Several classes around the world each semester Possible Interaction with (200 team) Student Competition in China organized by Beihang Univ.

FutureGrid Challenge Competition Core Computer Science FG-172 Cloud-TM from Portugal: on distributed concurrency control (software transactional memory): "When Scalability Meets Consistency: Genuine Multiversion Update Serializable Partial Data Replication,“ 32nd International Conference on Distributed Computing Systems (ICDCS'12) (top conference) used 40 nodes of FutureGrid Core Cyberinfrastructure FG-42,45 LSU/Rutgers: SAGA Pilot Job P* abstraction and applications. SAGA/BigJob use on clouds Core Cyberinfrastructure FG-130: Optimizing Scientific Workflows on Clouds. Scheduling Pegasus on distributed systems with overhead measured and reduced. Used Eucalyptus on FutureGrid Interesting application FG-133 from Univ. Arkansas: Supply Chain Network Simulator Using Cloud Computing with dynamic virtual machines supporting Monte Carlo simulation with Grid Appliance and Nimbus

https://portal.futuregrid.org/projects

New users per month Cumulative 0.5 Month XSEDE New Courses Tutorial SC11 Tutorial Cumulative Combines last two slides in one slide

Recent Projects Have Competitions Last one just finished Grand Prize Trip to SC12 Next Competition Beginning of August For our Science Cloud Summer School Recent Projects

FutureGrid Tutorials Cloud Provisioning Platforms Using Nimbus on FutureGrid [novice] Nimbus One-click Cluster Guide Using OpenStack Nova on FutureGrid Using Eucalyptus on FutureGrid [novice] Connecting private network VMs across Nimbus clusters using ViNe [novice] Using the Grid Appliance to run FutureGrid Cloud Clients [novice] Cloud Run-time Platforms Running Hadoop as a batch job using MyHadoop [novice] Running SalsaHadoop (one-click Hadoop) on HPC environment [beginner] Running Twister on HPC environment Running SalsaHadoop on Eucalyptus Running FG-Twister on Eucalyptus Running One-click Hadoop WordCount on Eucalyptus [beginner] Running One-click Twister K-means on Eucalyptus Image Management and Rain Using Image Management and Rain [novice] Storage Using HPSS from FutureGrid [novice] Educational Grid Virtual Appliances Running a Grid Appliance on your desktop Running a Grid Appliance on FutureGrid Running an OpenStack virtual appliance on FutureGrid Running Condor tasks on the Grid Appliance Running MPI tasks on the Grid Appliance Running Hadoop tasks on the Grid Appliance Deploying virtual private Grid Appliance clusters using Nimbus Building an educational appliance from Ubuntu 10.04 Customizing and registering Grid Appliance images using Eucalyptus High Performance Computing Basic High Performance Computing Running Hadoop as a batch job using MyHadoop Performance Analysis with Vampir Instrumentation and tracing with VampirTrace Experiment Management Running interactive experiments [novice] Running workflow experiments using Pegasus Pegasus 4.0 on FutureGrid Walkthrough [novice] Pegasus 4.0 on FutureGrid Tutorial [intermediary] Pegasus 4.0 on FutureGrid Virtual Cluster [advanced]

Portal Page Hits Not even Installed on FutureGrid 1/11/2019

Selected List of Services Offered FutureGrid Cloud PaaS Hadoop Twister HDFS Swift Object Store IaaS Nimbus Eucalyptus OpenStack ViNE GridaaS Genesis II Unicore SAGA Globus HPCaaS MPI OpenMP CUDA TestbedaaS FG RAIN Portal Inca Ganglia Devops Exper. Manag./Pegasus 1/11/2019

Services Offered India Sierra Hotel Alamo Xray Bravo Delta Echo Foxtrot Alamo Xray Bravo Delta Echo myHadoop ✔ Nimbus OpenStack Eucalyptus ViNe1  Genesis II Unicore MPI OpenMP ScaleMP Old Ganglia Pegasus3 Inca Portal2  PAPI Globus Services Offered ViNe can be installed on the other resources via Nimbus  Access to the resource is requested through the portal  Pegasus available via Nimbus and Eucalyptus images 1/11/2019

FutureGrid Technology and Project Requests Total Projects and Categories There is a bug in google charts that does not allow us to print this beyond June after new categories were introduced. Fugang and I try to find a way to export the data and print otherwise. took Fugang a long time to find out 

Software Components Portals including “Support” “use FutureGrid” “Outreach” Monitoring – INCA, Power (GreenIT) Experiment Manager: specify/workflow Image Generation and Repository Intercloud Networking ViNE Virtual Clusters built with virtual networks Performance library Rain or Runtime Adaptable InsertioN Service for images Security Authentication, Authorization, Note Software integrated across institutions and between middleware and systems Management (Google docs, Jira, Mediawiki) Note many software groups are also FG users “Research” Above and below Nimbus OpenStack Eucalyptus

New Developments Eucalyptus 3.0 OpenStack FG Cloud Metric Dashboard First academic cloud to have access to it and installed it Presentation at User Group meeting well received OpenStack Update to Essex Problems with our network hardware We need new switches FG Cloud Metric Dashboard Display details of the usage of IaaS frameworks: Reduced millions of Log entries to tens of thousands of instance traces 1/11/2019

May 2012 Preliminary Runtime Consumed Core Minutes 1/11/2019

Image Management and RAIN RAIN manages tools to dynamically provide custom HPC environment, Cloud environment, or virtual networks on-demand Bare Metal, Eucalyptus, OpenStack, TODO: Nimbus, OpenNebula, Amazon, Azure, Google IaaS Interoperability Collaboration between Systems and Software groups Several recent papers von Laszewski, G., J. Diaz, F. Wang, and G. Fox, Comparison of Multiple Cloud Frameworks, IEEE Cloud 2012, 06/2012 Diaz, J., G. von Laszewski, F. Wang, and G. Fox, Abstract Image Management and Universal Image Registration for Cloud and HPC Infrastructures, IEEE Cloud 2012, 06/2012 Gregor von Laszewski, Hyungro Lee, Javier Diaz, Fugang Wang, Koji Tanaka, Shubhada Karavinkoppa, Geoffrey C. Fox, Tom Furlani, Design of a Dynamic Provisioning System for a Federated Cloud and Bare-metal Environment (under preparation, draft available)

Templated(Abstract) Dynamic Provisioning OpenNebula Parallel provisioning now supported Moab/xCAT HPC – high as need reboot before use Abstract Specification of image mapped to various HPC and Cloud environments Essex replaces Cactus Current Eucalyptus 3 commercial while version 2 Open Source

Possible FutureGrid Futures Official End Date September 30 2013 FutureGrid is a Testbed – it is not (just) a Science Cloud Technology Evaluation, Education and training, Cyberinfrastructure/Computer Science more important than expected However it is a very good place to learn how to support a Science Cloud -- develop “Computational Science as a service” whether hosted commercially or academically Good commercial links Now modus operandi and core software well understood, can explore “Federating other sites in FutureGrid umbrella” US Europe China interest Need resource to explore larger scaling experiments (e.g. for MapReduce) Very little support funded in current FG but clear opportunity Experimental hosting of SaaS based environments New user mode? Join existing project to learn about its technology Open IaaS, MapReduce, MPI … projects as EOT offering

Computational Science as a Service Traditional Computer Center has a variety of capabilities supporting (scientific computing/scholarly research) users. Lets call this Computational Science as a Service IaaS, PaaS and SaaS are lower level parts of these capabilities but commercial clouds do not include Developing roles/appliances for particular users Supplying custom SaaS aimed at user communities Community Portals Integration across disparate resources for data and compute (i.e. grids) Consulting on use of particular appliances and SaaS i.e. on particular software components Debugging and other problem solving Data transfer and network link services Archival storage Administrative issues such as (local) accounting This allows us to develop a new model of a computer center where commercial companies operate base hardware/software A combination of XSEDE, Internet2 (USA) and computer center supply 1) to 9)

Using Science Clouds in a Nutshell High Throughput Computing; pleasingly parallel; grid applications Multiple users (long tail of science) and usages (parameter searches) Internet of Things (Sensor nets) as in cloud support of smart phones (Iterative) MapReduce including “most” data analysis Exploiting elasticity and platforms (HDFS, Object Stores, Queues ..) Use worker roles, services, portals (gateways) and workflow Good Strategies: Build the application as a service; Build on existing cloud deployments such as Hadoop; Use PaaS if possible; Design for failure; Use as a Service (e.g. SQLaaS) where possible; Address Challenge of Moving Data See Fox+Gannon Cloud Programming Paradigms for Technical Computing Applications