Https://portal.futuregrid.org FutureGrid Computing Testbed as a Service NSF Presentation NSF April 9 2013 Geoffrey Fox for FutureGrid Team

Slides:



Advertisements
Similar presentations
FutureGrid related presentations at TG and OGF Sun. 17th: Introduction to FutireGrid (OGF) Mon. 18th: Introducing to FutureGrid (TG) Tue. 19th –Educational.
Advertisements

FutureGrid UAB Meeting XSEDE13 San Diego July
FutureGrid Overview NSF PI Science of Cloud Workshop Washington DC March Geoffrey Fox
Cosmic Issues and Analysis of External Comments on FutureGrid TG11 Salt Lake City July Geoffrey Fox
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
FutureGrid Image Repository: A Generic Catalog and Storage System for Heterogeneous Virtual Machine Images Javier Diaz, Gregor von Laszewski, Fugang Wang,
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
FutureGrid Summary TG’10 Pittsburgh BOF on New Compute Systems in the TeraGrid Pipeline August Geoffrey Fox
FutureGrid Summary FutureGrid User Advisory Board TG’10 Pittsburgh August Geoffrey Fox
Big Data and Clouds: Challenges and Opportunities NIST January Geoffrey Fox
Eucalyptus on FutureGrid: A case for Eucalyptus 3 Sharif Islam, Javier Diaz, Geoffrey Fox Gregor von Laszewski Indiana University.
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid Overview CTS Conference 2011 Philadelphia May Geoffrey Fox
FutureGrid SOIC Lightning Talk February Geoffrey Fox
Distributed FutureGrid Clouds for Scalable Collaborative Sensor-Centric Grid Applications For AMSA TO 4 Sensor Grid Technical Interchange Meeting By Anabas,
Science of Cloud Computing Panel Cloud2011 Washington DC July Geoffrey Fox
Cloud Computing. What is Cloud Computing? Cloud computing is a model for enabling convenient, on-demand network access to a shared pool of configurable.
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
Cloud Computing 1. Outline  Introduction  Evolution  Cloud architecture  Map reduce operation  Platform 2.
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
FutureGrid Overview Geoffrey Fox
M.A.Doman Short video intro Model for enabling the delivery of computing as a SERVICE.
Presented by: Sanketh Beerabbi University of Central Florida COP Cloud Computing.
Future Grid FutureGrid Overview Dr. Speaker. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research on the future.
FutureGrid Overview Geoffrey Fox
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
Selected lessons learned from FutureGrid resulting in a toolkit for ComputingTestbedaaS: Cloudmesh HPDS 2014, Halifax, CA Gregor von Laszewski Geoffrey.
Data Science at Digital Science October Geoffrey Fox Judy Qiu
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Future Grid FutureGrid Overview Geoffrey Fox SC09 November
FutureGrid Connection to Comet Testbed and On Ramp as a Service Geoffrey Fox Indiana University Infra structure.
Image Generation and Management on FutureGrid CTS Conference 2011 Philadelphia May Geoffrey Fox
FutureGrid Overview Geoffrey Fox
Some remarks on Use of Clouds to Support Long Tail of Science July XSEDE 2012 Chicago ILL July 2012 Geoffrey Fox.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
FutureGrid Cyberinfrastructure for Computational Research.
Building Effective CyberGIS: FutureGrid Marlon Pierce, Geoffrey Fox Indiana University.
RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor.
FutureGrid Computing Testbed as a Service Overview July Geoffrey Fox for FutureGrid Team
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
FutureGrid BOF Overview TG 11 Salt Lake City July Geoffrey Fox
FutureGrid Computing Testbed as a Service for Condo_of_Condos Internet 2 panel April Jose Fortes for FutureGrid Team.
Virtual Appliances CTS Conference 2011 Philadelphia May Geoffrey Fox
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Recipes for Success with Big Data using FutureGrid Cloudmesh SDSC Exhibit Booth New Orleans Convention Center November Geoffrey Fox, Gregor von.
Future Grid Future Grid Overview. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research that will invent the future.
FutureGrid Computing Testbed as a Service EGI Technical Forum 2013 Madrid Spain September Geoffrey Fox and Gregor.
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
Grappling Cloud Infrastructure Services with a Generic Image Repository Javier Diaz Andrew J. Younge, Gregor von Laszewski, Fugang.
Indiana University Faculty Geoffrey Fox, David Crandall, Judy Qiu, Gregor von Laszewski Data Science at Digital Science Center.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Private Public FG Network NID: Network Impairment Device
Status and Challenges: January 2017
StratusLab Final Periodic Review
StratusLab Final Periodic Review
Bridges and Clouds Sergiu Sanielevici, PSC Director of User Support for Scientific Applications October 12, 2017 © 2017 Pittsburgh Supercomputing Center.
FutureGrid Computing Testbed as a Service Details
FutureGrid Computing Testbed as a Service NSF Presentation
NSF : CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science PI: Geoffrey C. Fox Software: MIDAS HPC-ABDS.
FutureGrid Computing Testbed as a Service
Digital Science Center Overview
I590 Data Science Curriculum August
Versatile HPC: Comet Virtual Clusters for the Long Tail of Science SC17 Denver Colorado Comet Virtualization Team: Trevor Cooper, Dmitry Mishin, Christopher.
Data Science Curriculum March
Cloud DIKW based on HPC-ABDS to integrate streaming and batch Big Data
Clouds from FutureGrid’s Perspective
Gregor von Laszewski Indiana University
FutureGrid Overview July XSEDE UAB Meeting Geoffrey Fox
Presentation transcript:

FutureGrid Computing Testbed as a Service NSF Presentation NSF April Geoffrey Fox for FutureGrid Team School of Informatics and Computing Digital Science Center Indiana University Bloomington

FutureGrid Testbed as a Service FutureGrid is part of XSEDE set up as a testbed with cloud focus Operational since Summer 2010 (i.e. now in third year of use) The FutureGrid testbed supports Computer Science and Computational Science research with – A flexible development and test platform for CS systems and application users looking at interoperability, functionality, performance or evaluation – User-customizable, accessed interactively environment for Grid, Cloud and HPC software with and without VM’s – A rich education and teaching platform for classes Offers OpenStack, Eucalyptus, Nimbus, OpenNebula, HPC (MPI) on same hardware moving to software defined systems Rather than loading images onto VM’s, FutureGrid supports Cloud, Grid and Parallel computing environments by provisioning software as needed onto “bare-metal” or VM’s/Hypervisors using (changing) open source tools

4 Use Types for FutureGrid TestbedaaS 292 approved projects (1734 users) April – USA(79%), Puerto Rico(3%- Students in class), India, China, lots of European countries (Italy at 2% as class) – Industry, Government, Academia Computer science and Middleware (55.6%) – Computer Science and Cyberinfrastructure; Interoperability (3.6%) for Grids and Clouds such as Open Grid Forum OGF Standards New Domain Science applications (20.4%) – Life science highlighted (10.5%), Non Life Science (9.9%) Training Education and Outreach (14.9%) – Semester and short events; focus on outreach to HBCU Computer Systems Evaluation (9.1%) – XSEDE (TIS, TAS), OSG, EGI; Campuses 3

4 NameSystem type# CPUs# CoresTFLOPS Total RAM (GB) Secondary Storage (TB) Site Status India IBM iDataPlex IU Operational Alamo Dell PowerEdge TACC Operational Hotel IBM iDataPlex UC Operational Sierra IBM iDataPlex SDSC Operational Xray Cray XT5m IU Operational Foxtrot IBM iDataPlex UF Operational Bravo Large Disk & memory (192GB per node) 192 (12 TB per Server) IU Operational Delta Large Disk & memory With Tesla GPU’s 32 CPU 32 GPU’s (192GB per node) 192 (12 TB per Server)IUOperational LimaSSD Test System (SSD) 8(SATA) SDSCOperational Echo Large memory ScaleMP IUBeta TOTAL GPU GPU Heterogeneous Systems Hardware Small systems avoid interference of specialized technologies

FutureGrid Partners Indiana University (Architecture, core software, Support) San Diego Supercomputer Center at University of California San Diego (INCA, Monitoring) University of Chicago/Argonne National Labs (Nimbus) University of Florida (ViNE, Education and Outreach) University of Southern California Information Sciences (Pegasus to manage experiments) University of Tennessee Knoxville (Benchmarking) University of Texas at Austin/Texas Advanced Computing Center (Portal, XSEDE Integration) University of Virginia (OGF, XSEDE Software stack) Red institutions have FutureGrid hardware

Sample FutureGrid Projects I FG18 Privacy preserving gene read mapping developed hybrid MapReduce. Small private secure + large public with safe data. Won 2011 PET Award for Outstanding Research in Privacy Enhancing Technologies FG132, Power Grid Sensor analytics on the cloud with distributed Hadoop. Won the IEEE Scaling challenge at CCGrid2012. FG156 Integrated System for End-to-end High Performance Networking showed that the RDMA (Remote Direct Memory Access) over Converged Ethernet Roce (InfiniBand made to work over Ethernet network frames) protocol could be used over wide-area networks FG172 Cloud-TM on distributed concurrency control (software transactional memory): "When Scalability Meets Consistency: Genuine Multiversion Update Serializable Partial Data Replication,“ 32nd International Conference on Distributed Computing Systems (ICDCS'12) (good conference) used 40 nodes of FutureGrid 6

Sample FutureGrid Projects II FG42,45 SAGA Pilot Job P* abstraction and applications. XSEDE Cyberinfrastructure used on clouds FG130 Optimizing Scientific Workflows on Clouds. Scheduling Pegasus on distributed systems with overhead measured and reduced. Used Eucalyptus on FutureGrid FG133 Supply Chain Network Simulator Using Cloud Computing with dynamic virtual machines supporting Monte Carlo simulation with Grid Appliance and Nimbus FG257 Particle Physics Data analysis for ATLAS LHC experiment used FutureGrid + Canadian Cloud resources to study data analysis on Nimbus + OpenStack with up to 600 simultaneous jobs FG254 Information Diffusion in Online Social Networks is evaluating NoSQL databases (Hbase, MongoDB, Riak) to support analysis of Twitter feeds FG323 SSD performance benchmarking for HDFS on Lima 7

Education and Training Use of FutureGrid 27 Semester long classes: 563+ students – Cloud Computing, Distributed Systems, Scientific Computing and Data Analytics 3 one week summer schools: 390+ students – Big Data, Cloudy View of Computing (for HBCU’s), Science Clouds 1 two day workshop: 28 students 5 one day tutorials: 173 students From 19 Institutions Developing 2 MOOC’s (Google Course Builder) on Cloud Computing and use of FutureGrid supported by either FutureGrid or downloadable appliances (custom images) – See FutureGrid appliances support Condor/MPI/Hadoop/Iterative MapReduce virtual clusters 8

Inca Software functionality and performance Ganglia Cluster monitoring perfSONAR Network monitoring - Iperf measurements SNAPP Network monitoring – SNMP measurements Monitoring on FutureGrid Important and even more needs to be done

Link FutureGrid and GENI Identify how to use the ORCA federation framework to integrate FutureGrid (and more of XSEDE?) into ExoGENI Allow FG(XSEDE) users to access the GENI resources and vice versa Leverage the Image generation capabilities of FG and the bare metal deployment strategies of FG within the GENI context. – Software defined networks plus cloud/bare metal dynamic provisioning gives software defined systems Enable PaaS level services (such as a distributed Hbase or Hadoop) to be deployed across FG and GENI resources Initial project on virtual distributed data systems 10

Essential and Different features of FutureGrid Unlike many clouds such as Amazon and Azure, FutureGrid allows robust reproducible (in performance and functionality) research (you can request same node with and without VM) – Open Transparent Technology Environment FutureGrid is more than a Cloud; it is a general distributed Sandbox; a cloud grid HPC testbed Supports 3 different IaaS environments (Nimbus, Eucalyptus, OpenStack) and projects involve 5 (also CloudStack, OpenNebula) Supports research on cloud tools, cloud middleware and cloud-based systems as well as use of clouds in applications FutureGrid has developed middleware and interfaces for Computing TestbedaaS e.g. Phantom (cloud user interface) Vine (virtual network) RAIN (deploy systems) and security/metric integration FutureGrid has experience in running cloud systems 11

FutureGrid as an onramp to other systems FG supports Education & Training for all systems User can do all work on FutureGrid OR User soon can use FutureGrid CloudMesh to jump to chosen production system using image tested on FutureGrid – Uses general templated image to install on various IaaS & bare metal including Amazon, Azure, Eucalyptus, Openstack, OpenNebula, Nimbus, HPC – Provisions the complete system needed by user and not just a single image; copes with resource limitations and deploys full range of software – Integrates our VM metrics package (TAS collaboration) that links to XSEDE (VM's are different from traditional Linux in metrics supported and needed) – Related to OpenStack Horizon, but aimed at multiple federated systems. – Built on current FutureGrid provisioning technology and tools like libcloud, boto with protocol (EC2) or programmatic API (python) – User can download Appliances based on templated image on to local machines (Virtual Box) 12

Lessons learnt from FutureGrid Unexpected major use from Computer Science and Middleware Rapid evolution of Technology Eucalyptus  Nimbus  OpenStack Open source IaaS maturing as in “Paypal To Drop VMware From 80,000 Servers and Replace It With OpenStack” (Forbes) eBay to switch broadly? Need interactive not batch use; nearly all jobs short Substantial TestbedaaS technology needed and FutureGrid developed (RAIN, CloudMesh, Operational model) some Lessons more positive than DoE Magellan report (aimed as an early science cloud) but goals different Still serious performance problems in clouds for networking and device (GPU) linkage; many activities outside FG addressing – One can get good Infiniband performance (MPI) on a peculiar OS + Mellanox drivers but not general yet We identified characteristics of “optimal hardware” Run system with integrated software (computer science) and systems administration team Build Computer Testbed as a Service Community 13

Future Directions for FutureGrid Poised to support more users as technology like OpenStack matures – Please encourage new users and new challenges More focus on academic Platform as a Service (PaaS) - high-level middleware (e.g. Hadoop, Hbase, MongoDB) – as IaaS gets easier to deploy Expect increased Big Data challenges Improve Education and Training with model for MOOC laboratories Finish CloudMesh (and integrate with Nimbus Phantom) to make FutureGrid as hub to jump to multiple different “production” clouds commercially, nationally and on campuses; allow cloud bursting – Several collaborations developing Build underlying software defined system model with integration with GENI and high performance virtualized devices (MIC, GPU) Improved ubiquitous monitoring at PaaS IaaS and NaaS levels Improve “Reproducible Experiment Management” environment Expand and renew hardware via federation 14

Extras 15

Selected List of Services Offered 16 Cloud PaaS Hadoop Iterative MapReduce HDFS Hbase Swift Object Store IaaS Nimbus Eucalyptus OpenStack ViNE GridaaS Genesis II Unicore SAGA Globus HPCaaS MPI OpenMP CUDA TestbedaaS FG RAIN, CloudMesh Portal Inca Ganglia Devops (Chef, Puppet, Salt) Experiment Management e.g. Pegasus

Infra structure IaaS  Software Defined Computing (virtual Clusters)  Hypervisor, Bare Metal  Operating System Platform PaaS  Cloud e.g. MapReduce  HPC e.g. PETSc, SAGA  Computer Science e.g. Compiler tools, Sensor nets, Monitors FutureGrid offers Computing Testbed as a Service Network NaaS  Software Defined Networks  OpenFlow GENI Software (Application Or Usage) SaaS  CS Research Use e.g. test new compiler or storage model  Class Usages e.g. run GPU & multicore  Applications FutureGrid Uses Testbed-aaS Tools  Provisioning  Image Management  IaaS Interoperability  NaaS, IaaS tools  Expt management  Dynamic IaaS NaaS  Devops FutureGrid Uses Testbed-aaS Tools  Provisioning  Image Management  IaaS Interoperability  NaaS, IaaS tools  Expt management  Dynamic IaaS NaaS  Devops FutureGrid RAIN uses Dynamic Provisioning and Image Management to provide custom environments that need to be created. A Rain request may involves (1) creating, (2) deploying, and (3) provisioning of one or more images in a set of machines on demand

Performance of Dynamic Provisioning 4 Phases a) Design and create image (security vet) b) Store in repository as template with components c) Register Image to VM Manager (cached ahead of time) d) Instantiate (Provision) image 18

Security issues in FutureGrid Operation Security for TestBedaaS is a good research area (and Cybersecurity research supported on FutureGrid)! Authentication and Authorization model – This is different from those in use in XSEDE and changes in different releases of VM Management systems – We need to largely isolate users from these changes for obvious reasons – Non secure deployment defaults (in case of OpenStack) – OpenStack Grizzly (just released) has reworked the role based access control mechanisms and introduced a better token format based on standard PKI (as used in AWS, Google, Azure) – Custom: We integrate with our distributed LDAP between the FutureGrid portal and VM managers. LDAP server will soon synchronize via AMIE to XSEDE Security of Dynamically Provisioned Images – Templated image generation process automatically puts security restrictions into the image; This includes the removal of root access – Images include service allowing designated users (project members) to log in – Images vetted before allowing role-dependent bare metal deployment – No SSH keys stored in images (just call to identity service) so only certified users can use 19

Related Projects Grid5000 (Europe) and OpenCirrus with managed flexible environments are closest to FutureGrid and are collaborators PlanetLab has a networking focus with less managed system Several GENI related activities including network centric EmuLab, PRObE (Parallel Reconfigurable Observational Environment), ProtoGENI, ExoGENI, InstaGENI and GENICloud BonFire (Europe) similar to Emulab Recent EGI Federated Cloud with OpenStack and OpenNebula aimed at EU Grid/Cloud federation Private Clouds: Red Cloud (XSEDE), Wispy (XSEDE), Open Science Data Cloud and the Open Cloud Consortium are typically aimed at computational science Public Clouds such as AWS do not allow reproducible experiments and bare-metal/VM comparison; do not support experiments on low level cloud technology 20