2010 FutureGrid User Advisory Meeting Architecture Roadmap Status Now – October –Next year 11:15-12:00, Monday, August 2, 2010 Pittsburgh, PA Gregor von.

Slides:



Advertisements
Similar presentations
Using Metrics to Reduce Cost of Re-work Dwight Lamppert Senior Test Manager Franklin Templeton.
Advertisements

1 Senn, Information Technology, 3 rd Edition © 2004 Pearson Prentice Hall James A. Senns Information Technology, 3 rd Edition Chapter 7 Enterprise Databases.
Copyright © 2003 Pearson Education, Inc. Slide 1 Computer Systems Organization & Architecture Chapters 8-12 John D. Carpinelli.
1 Copyright © 2013 Elsevier Inc. All rights reserved. Chapter 1 Embedded Computing.
Copyright © 2011, Elsevier Inc. All rights reserved. Chapter 6 Author: Julia Richards and R. Scott Hawley.
Author: Julia Richards and R. Scott Hawley
1 Copyright © 2013 Elsevier Inc. All rights reserved. Chapter 3 CPUs.
Properties Use, share, or modify this drill on mathematic properties. There is too much material for a single class, so you’ll have to select for your.
Task Group Chairman and Technical Contact Responsibilities ASTM International Officers Training Workshop September 2012 Scott Orthey and Steve Mawn 1.
UNITED NATIONS Shipment Details Report – January 2006.
RXQ Customer Enrollment Using a Registration Agent (RA) Process Flow Diagram (Move-In) Customer Supplier Customer authorizes Enrollment ( )
HL7 Project Management Tool Overview for HL7 Project Facilitators
Overview of the FutureGrid Software
Business Transaction Management Software for Application Coordination 1 Business Processes and Coordination. Introduction to the Business.
1 Introducing the Specifications of the Metro Ethernet Forum MEF 19 Abstract Test Suite for UNI Type 1 February 2008.
18 Copyright © 2005, Oracle. All rights reserved. Distributing Modular Applications: Introduction to Web Services.
1 RA I Sub-Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Casablanca, Morocco, 20 – 22 December 2005 Status of observing programmes in RA I.
State of New Jersey Department of Health and Senior Services Patient Safety Reporting System Module 2 – New Event Entry.
Exit a Customer Chapter 8. Exit a Customer 8-2 Objectives Perform exit summary process consisting of the following steps: Review service records Close.
1 Click here to End Presentation Software: Installation and Updates Internet Download CD release NACIS Updates.
Auto-scaling Axis2 Web Services on Amazon EC2 By Afkham Azeez.
REVIEW: Arthropod ID. 1. Name the subphylum. 2. Name the subphylum. 3. Name the order.
Week 2 The Object-Oriented Approach to Requirements
Configuration management
Management Plans: A Roadmap to Successful Implementation
ACT User Meeting June Your entitlements window Entitlements, roles and v1 security overview Problems with v1 security Tasks, jobs and v2 security.
EU market situation for eggs and poultry Management Committee 20 October 2011.
Bright Futures Guidelines Priorities and Screening Tables
EIS Bridge Tool and Staging Tables September 1, 2009 Instructor: Way Poteat Slide: 1.
AEMCPAGE Relaunch 1 June 2009.
1 ITSS This overview deck contains two sections. Please use the links below to navigate –How to Register for ITSS Application AccessHow to Register for.
Operations to Serve You 05/17/ The Service Desk Provides an Announcement Page? The Service Desk houses a library of SOLUTIONS that are available.
1 IC GS J. Broome, Mar Introduction to the Informatics and Data Aspects John Broome (Canada)
2 |SharePoint Saturday New York City
Operating Systems Operating Systems - Winter 2011 Dr. Melanie Rieback Design and Implementation.
Operating Systems Operating Systems - Winter 2012 Dr. Melanie Rieback Design and Implementation.
VOORBLAD.
Sample Service Screenshots Enterprise Cloud Service 11.3.
1 RA III - Regional Training Seminar on CLIMAT&CLIMAT TEMP Reporting Buenos Aires, Argentina, 25 – 27 October 2006 Status of observing programmes in RA.
Basel-ICU-Journal Challenge18/20/ Basel-ICU-Journal Challenge8/20/2014.
1..
CONTROL VISION Set-up. Step 1 Step 2 Step 3 Step 5 Step 4.
© 2012 National Heart Foundation of Australia. Slide 2.
Services Course Windows Live SkyDrive Participant Guide.
SLP – Endless Possibilities What can SLP do for your school? Everything you need to know about SLP – past, present and future.
Global Analysis and Distributed Systems Software Architecture Lecture # 5-6.
By CA. Pankaj Deshpande B.Com, FCA, D.I.S.A. (ICA) 1.
Macromedia Dreamweaver MX 2004 – Design Professional Dreamweaver GETTING STARTED WITH.
Model and Relationships 6 M 1 M M M M M M M M M M M M M M M M
Analyzing Genes and Genomes
© 2007 Cisco Systems, Inc. All rights reserved.Cisco Public 1 Addressing the Network – IPv4 Network Fundamentals – Chapter 6.
©Brooks/Cole, 2001 Chapter 12 Derived Types-- Enumerated, Structure and Union.
Essential Cell Biology
1 Phase III: Planning Action Developing Improvement Plans.
Intracellular Compartments and Transport
PSSA Preparation.
Essential Cell Biology
SESSION ID: Continuous Monitoring with the 20 Critical Security Controls SPO1-W02 Wolfgang Kandek CTO.
The DDS Benchmarking Environment James Edmondson Vanderbilt University Nashville, TN.
- 1 - Defense Security Service Background: During the Fall of 2012 Defense Security Service will be integrating ISFD with the Identity Management (IdM)
Energy Generation in Mitochondria and Chlorplasts
TCP/IP Protocol Suite 1 Chapter 18 Upon completion you will be able to: Remote Login: Telnet Understand how TELNET works Understand the role of NVT in.

Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
FutureGrid Cyberinfrastructure for Computational Research.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Presentation transcript:

2010 FutureGrid User Advisory Meeting Architecture Roadmap Status Now – October –Next year 11:15-12:00, Monday, August 2, 2010 Pittsburgh, PA Gregor von Laszewski Representing the Software Architecture Committee esentations/software/tg2010/ fg-tg2010-arch-timeline.pptx 1

Milestones Oct Project Start Nov SC Demo Mar Network Completed July 2010 Hardware available to early users Oct Hardware available to general users Nov SC Demo Oct.2011 Integration into TeraGrid Oct Project end Phase I Phase II Phase III Acceptance tests 2

Approach: Spiral Phase instead of prototype Established method Document templates Task tracking 3

Phases Phase I: Enabling the early adopters to use the system. The system will only have limited services available, this includes queuing systems and basic cloud services. Establishing essential administrative services for supporting the FG team, FG developers, FG users, and FG operations. Phase II: Delivering a first operated FG deployment for users. focusing on the delivery of basic dynamic provision activities in FG. Improving the administrative services for supporting the FG team, FG developers, FG users, and FG operations. Phase III: Delivers a first officially operated FG deployment for users focusing on the delivery of basic dynamic provision activities in FG. Integration into the TeraGrid. We may add additional Phases as needed. 4

Now = Phase I Now = within this month Aug & Sept. Gradually improved – With outlook for dynamic provisioning. 5

Selected Use Cases Supported in Phase I Providing support and information: – In order for FG to function we must provide sufficient support infrastructure to deliver information to the uesers, the ability to interact with system staff through Forums/Mailinglists, tracking Issues, and contributing to the dissemination and sharing of information. Ability to run a HPC job – In order to compare the performance of different non HPC frameworks we need to also provide the ability to run HPC jobs on the same hardware as the other experiments. Ability to use clouds such as Nimbus and Eucalyptus – In order to test out other than HPC jobs, we need to allow the user to use a cloud environment such as Nimbus or Eucalyptus. Ability to store images – Users need to be able to store some of the images the use as part of the cloud experiments Ability authenticate to FG computational resources – Users need the ability to authenticate to the FG resources to utilize them (all computational resources are managed through an account service so users have only one account). Nimbus service, Eucalyptus service, hpc service, web site, authentication 6

Phase I: Conceptual View 7

Phase I: Service View = Now 8

Services: Admin Administrative Services Web server for FG developers – Ticket system for developers – Wiki like service for contributing documentation Web server for FG users – Ticket system for user ticket – Account Service – Hardware Monitor via Inca 9

Services: Computing Computational Resources and Services IU iDataPlex: india.futuregrid.org – HPC Queue IU Cray: xray.futuregrid.org – HPC Queue – Eucalyptus Cloud SDSC iDataPlex: sierra.futuregrid.org – HPC Queue – Eucalyptus Cloud – Nimbus Cloud UC iDataPlex: hotel.futuregrid.org – HPC Queue – Nimbus Cloud UF iDataPlex: foxtrot.futuregrid.org – HPC Queue – Nimbus Cloud 10

Layered Architecture View 11

Deployment View Robert Henshel: HPC software – Tasks: describe which HPC software should be installed where. Determine who is installing the software (consult David H.). David Hancock: Queuing systems, Eucalyptus – Outstanding tasks: Complete the Queuing system install for the HPC related jobs, the Nimbus, and Eucalyptus system. Document if it is possible already at this stage to shrink and increase the number of nodes used for the various split usage between HPC, Nimbus, and Eucalyptus. Rich Knepper: Operational Services – Outstanding tasks: Describe how the operational services have been deployed so that they can be maintained by a group of system administrators. Describe and develop processes for managing accounts and other operational activities. List if additional software development activities are needed. Jonathan Bolte: Support Services and KB – Outstanding tasks: Describe which support software is available for Phase I, how is it deployed and operator remote groups are responsible for furthering their deployment in collaboration with the above group. Leads at the organizations are – UF: Mauricio – UC: Ti Legget – SDSC: Shava Smallen + some people at UC 12

Deployment view Security: based on ssh keys – Managed through LDAP, keys are populated into the local file system of each machine Filesystem: – Details from David Hancock Queueing system – Moab, each machine has their own queue 13

Status: Admin A FG web server that provides documentation about FG ( The web server is based on Drupal. A wiki that provides internal information to FG developers ( The wiki is based on Mediawiki. An list server for internal and external lists. The server is based on Mailman. A project tracking system for managing tasks ( The project tracking system is based on JIRA. A ticketing system for tracking user support requests (send to The user ticket system is RT. Where appropriate tickets, will be transfered to other systems (e.g. JIRA) and tracked there. An account management system allowing to register projects and account based on LDAP is available. Scripts populate the account information to the resources. A public SVN repository in Sourceforge was created. A private SVN repository to manage security related code has been created. 14

Status: Computing We have enabled a status report of FG to the portal. – status page (updated by the system administrators) status page – inca page (updated by the inca system automatically) inca page The exact status can be found at on the FutureGrid status page.FutureGrid status page Sites – IU iDataPlex india.fuguregrid.org Torque/Moab – IU Cray xray.futuregrid.org Torque/Moab SDSC iDataPlex sierra.futuregrid.org Torque/Moab – UC iDataPlex hotel.futuregrid.org Torque/Moab – UF iDataPlex foxtrot.futuregrid.org Torque/Moab An exact status of the systems is conducted by Greg Pike – (versions, functionality) – Status of Nimbus and Eucalyptus is being tested. 15

Risks Systems just became available Cray is a bit different to users (statically binding) Nimbus needs user CA We let users access the system what is considered a really early user Documentation must be of high priority – No task can be resolved without making sure it is documented properly. Not everything has been tested or is working – This includes Moab, Nimbus, Eucalyptus – Outdated as of yesterday. 16

Phase I: Service View (August) = Now = August 17

Phase I: Service View (Sept.) = Now = August = Sept. 18

Moving between a service and HPC by administrator 19

October: Phase II 20

Moving between a service and HPC by user 21

Conceptual View 22

Layered View 23

Dynamic Provisioning 24

Image Repository 25

Performance Subsystem 26

Current - Inca Functional Monitoring Google Map View FutureGrid Core Services Tests FutureGrid Basic Cluster Tests FutureGrid Cluster Storage Tests 27

Performance Monitoring Service Collect performance and load data from all FutureGrid components – use passive measurements when possible – (Passive) Instrumentation of FutureGrid components via NetLogger (in testing) – (Active) Automated benchmarking via Inca – HPCC (deployed), SPEC, UTKs Grid Benchmark Challenge – Other hardware monitoring systems - Ganglia, Nagios, perfSONAR, etc. - (planned) Comparative studies with other clouds and grids 28

Current - Performance Monitoring HPCC metrics HPCC metrics (history) Automated benchmarking (HPCC) Global NOC Network Monitoring 29

Performance Tools Provide full support of partner tools Provide best effort support of external tools (UTK) (TU-D) (Tau – U of Oregon) (LBNL, SDSC) (HLRS)(JSC) 30

Current - Vampir FutureGrid install in progress - available today on IU Quarry and BigRed machines Vampir architecture Vampir GUI screenshots 31

General Milestones Allow dynamic provisioning of Eucalyptus Clouds Allow dynamic provisioning of Nimbus Clouds Demonstrate the use of non Cloud technologies 32

Next Year: Phase III 33

34

Portal Subsystem 35

FutureGrid Architecture 36

General Milestones Allow dynamic provisioning of other stacks – Dryad, other frameworks, gather feedback from users Demonstrate how to do performance comparisons Identify an example for using the Network Interruption Device Distributed Image Repository Integration with TG 37

Milestones 38 Oct Project Start Nov SC Demo Mar Network Completed July 2010 Hardware available to early users Oct Hardware available to general users Oct.2011 Integration into TeraGrid Oct Project end