1 Cloud Systems Panel at HPDC Boston June 25 2008 Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University

Slides:



Advertisements
Similar presentations
Virtualization, Cloud Computing, and TeraGrid Kate Keahey (University of Chicago, ANL) Marlon Pierce (Indiana University)
Advertisements

© 2012 Open Grid Forum Simplifying Inter-Clouds October 10, 2012 Hyatt Regency Hotel Chicago, Illinois, USA.
The Datacenter Needs an Operating System Matei Zaharia, Benjamin Hindman, Andy Konwinski, Ali Ghodsi, Anthony Joseph, Randy Katz, Scott Shenker, Ion Stoica.
The Open Science Data Cloud Robert L. Grossman University of Chicago and Open Cloud Consortium April 4, 2012 A 501(c)(3) not-for-profit operating clouds.
1 Challenges and New Trends in Data Intensive Science Panel at Data-aware Distributed Computing (DADC) Workshop HPDC Boston June Geoffrey Fox Community.
The Great Academia/Industry Grid Debate November Geoffrey Fox Community Grids Laboratory Indiana University
Parallel Research at Illinois Parallel Everywhere
International Conference on Cloud and Green Computing (CGC2011, SCA2011, DASC2011, PICom2011, EmbeddedCom2011) University.
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
 Amazon Web Services announced the launch of Cluster Compute Instances for Amazon EC2.  Which aims to provide high-bandwidth, low- latency instances.
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
Student Visits August Geoffrey Fox
1 Multicore and Cloud Futures CCGSC September Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University
Lecture 1.
Big Data A big step towards innovation, competition and productivity.
WORKFLOWS IN CLOUD COMPUTING. CLOUD COMPUTING  Delivering applications or services in on-demand environment  Hundreds of thousands of users / applications.
Cloud computing Tahani aljehani.
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
SYSTEMS SUPPORT FOR GRAPHICAL LEARNING Ken Birman 1 CS6410 Fall /18/2014.
3DAPAS/ECMLS panel Dynamic Distributed Data Intensive Analysis Environments for Life Sciences: June San Jose Geoffrey Fox, Shantenu Jha, Dan Katz,
Chapter 1- Introduction
1 Challenges Facing Modeling and Simulation in HPC Environments Panel remarks ECMS Multiconference HPCS 2008 Nicosia Cyprus June Geoffrey Fox Community.
X-Informatics Cloud Technology (Continued) March Geoffrey Fox Associate.
1 1 Hybrid Cloud Solutions (Private with Public Burst) Accelerate and Orchestrate Enterprise Applications.
Cloud Computing Saneel Bidaye uni-slb2181. What is Cloud Computing? Cloud Computing refers to both the applications delivered as services over the Internet.
I399 1 Research Methods for Informatics and Computing D: Basic Issues Geoffrey Fox Associate Dean for Research.
SALSASALSASALSASALSA AOGS, Singapore, August 11-14, 2009 Geoffrey Fox 1,2 and Marlon Pierce 1
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
CS525: Special Topics in DBs Large-Scale Data Management Hadoop/MapReduce Computing Paradigm Spring 2013 WPI, Mohamed Eltabakh 1.
OpenQuake Infomall ACES Meeting Maui May Geoffrey Fox
Scientific Computing Environments ( Distributed Computing in an Exascale era) August Geoffrey Fox
4.2.1 Programming Models Technology drivers – Node count, scale of parallelism within the node – Heterogeneity – Complex memory hierarchies – Failure rates.
Cloud Computing John Engates CTO, Rackspace Presented: Rackspace Customer Conference, 2008 October 29, 2008.
ICETE 2012 Joint Conference on e-Business and Telecommunications Hotel Meliá Roma Aurelia Antica, Rome, Italy July
Crystal Ball Panel ORNL Heterogeneous Distributed Computing Research Al Geist ORNL March 6, 2003 SOS 7.
High Performance Computing How to use Recommended Books Spring Semester 2005 Geoffrey Fox Community Grids Laboratory Indiana University 505 N Morton Suite.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Chapter 5 McGraw-Hill/Irwin Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. Enterprise Architectures.
Scalable Systems Software for Terascale Computer Centers Coordinator: Al Geist Participating Organizations ORNL ANL LBNL.
ISERVOGrid Architecture Working Group Brisbane Australia June Geoffrey Fox Community Grids Lab Indiana University
Applications and Requirements for Scientific Workflow Introduction May NSF Geoffrey Fox Indiana University.
SALSASALSASALSASALSA Clouds Ball Aerospace March Geoffrey Fox
© 2006 Open Grid Forum Geoffrey Fox OGF 23 Barcelona GFSG Meeting June OGF eScience Function.
SALSASALSASALSASALSA Cloud Panel Session CloudCom 2009 Beijing Jiaotong University Beijing December Geoffrey Fox
1 Multicore for Science Multicore Panel at eScience 2008 December Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University.
Security: systems, clouds, models, and privacy challenges iDASH Symposium San Diego CA October Geoffrey.
SALSASALSASALSASALSA Digital Science Center February 12, 2010, Bloomington Geoffrey Fox Judy Qiu
Hadoop/MapReduce Computing Paradigm 1 CS525: Special Topics in DBs Large-Scale Data Management Presented By Kelly Technologies
HPC in the Cloud – Clearing the Mist or Lost in the Fog Panel at SC11 Seattle November Geoffrey Fox
Chapter 1- Introduction Lecture 1. Topics covered  Professional software development  What is meant by software engineering.  Software engineering.
Minimising IT costs, maximising operational efficiency IO and NIMM: Now is the time Glyn Knaresborough Director of Strategic Consulting.
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
Big Data Open Source Software and Projects ABDS in Summary II: Layer 5 I590 Data Science Curriculum August Geoffrey Fox
PARALLEL AND DISTRIBUTED PROGRAMMING MODELS U. Jhashuva 1 Asst. Prof Dept. of CSE om.
SYSTEM MODELS FOR ADVANCED COMPUTING Jhashuva. U 1 Asst. Prof CSE
© 2006 Open Grid Forum Geoffrey Fox Board Review April OGF eScience Function.
Spark on Entropy : A Reliable & Efficient Scheduler for Low-latency Parallel Jobs in Heterogeneous Cloud Huankai Chen PhD Student at University of Kent.
Panel: Beyond Exascale Computing
CompSci 280 S Introduction to Software Development
Hadoop Aakash Kag What Why How 1.
Geoffrey Fox, Shantenu Jha, Dan Katz, Judy Qiu, Jon Weissman
9/8/2018 5:55 PM © 2007 Microsoft Corporation. All rights reserved. Microsoft, Windows, Windows Vista and other product names are or may be registered.
© 2016 Global Market Insights, Inc. USA. All Rights Reserved Fuel Cell Market size worth $25.5bn by 2024Low Power Wide Area Network.
Hadoop Market
Biology MDS and Clustering Results
Cyberinfrastructure and PolarGrid
Services, Security, and Privacy in Cloud Computing
3 Questions for Cluster and Grid Use
Panel on Research Challenges in Big Data
Cloud versus Cloud: How Will Cloud Computing Shape Our World?
Presentation transcript:

1 Cloud Systems Panel at HPDC Boston June Geoffrey Fox Community Grids Laboratory, School of informatics Indiana University

HPDC Cloud Panel Members Randy Bryant, CMU Ian Foster, Chicago Greg Pfister, Consultant Dennis Quan, IBM Dan Reed, Microsoft Moderator Geoffrey Fox, Indiana University

HPDC Cloud Panel Questions What is and equally interestingly what isn't a cloud? – Especially with today’s hype – all is a cloud? What is implication of Clouds for Enterprise Data Centers? If TeraGrid evolves to PetaCloud, how would it look? What are research issues for Clouds? Can one (or who could) "trust" clouds? Will Cloud interoperability be important; if so at what interface(s) will it be provided? What is the killer app for clouds?

Some Conclusions for Scientific Applications Scientific applications and commercial applications share a set of requirements for cloud computing (customized environments, short-term and on- demand capability provisioning) But many scientific applications have additional demands for cloud computing (tightly coupled, state management) The Science community needs to take leadership to find the best ways to leverage this new technology and adapt it for scientific use. (industry won't do it for us)

Some Panel Comments I Clouds are already being used in many Web 2.0 applications which tend to involve little computing Gmail is stateful and this implies nontrivial cloud logic whereas traditional web access has no difficult issues; user restarts by reloading web page Round robin DNS critical technology Scientific data analysis natural cloud application Grids hardly mentioned Clouds likely to become dominant Enterprise data center technology Virtualization critical to cloud implementations

Some Panel Comments II MPI will only be supported if there is a business model to justify it; likely that there is no business model Mapreduce (Hadoop) has critical fault tolerance aspects; MPI needs to avoid tight synchronization that leads to inflexible non fault tolerant implementations Shared memory (between cores as needed in OpenMP) will not be supported Google does not put the fastest CPU’s in its data centers; better are more cheaper better power efficient slower CPUs

Some Panel Comments III Important emphasis on management tools allowing you better access to state of a cloud application Need to be able to track down sources of failures High level parallel (cloud) languages like Dryad or Mapreduce will be essential; again MPI not relevant as too low level Scaling is key characteristic of clouds Many QoS characteristics will be supported including guarantee not to run application in particular countries Interoperability not likely to be an early focus

Some Panel Comments IV Research issues include better high level programming models and use of clouds for science Although some critical data will not be allowed outside a corporate firewall, the panel did not see “trust” (security) as a critical issue Two interesting IBM clouds support internal IBM projects (Innoivators cloud) and IT industry in city of Wuxi IO requirements to read now petabytes and soon terabytes of data high Soon it will be considered quite routine to run queries on the fly against petabytes of data