DOE Perspective on Cyberinfrastructure - LBNL Gary Jung Manager, High Performance Computing Services Lawrence Berkeley National Laboratory Educause CCI.

Slides:



Advertisements
Similar presentations
Learner-Centered Education Course Redesign Initiative Builds upon work of prior LCE grants Will award grants of $40,000 - $50,000, with the option.
Advertisements

Intergovernmental Science-Policy Platform for Biodiversity and Ecosystem Services (IPBES) Importance for Africa ipBes side event at AMCEN. May 28th 2009,
MRSEC Directors Meeting Focus on MRSEC Working Groups NSF, Thursday June 24, 2010 Sean Jones, Tom Rieker, and Charles Ying MRSEC Program Directors.
Internet2 Infrastructure. An advanced networking consortium whose members include: – 221 U.S. universities – 45 leading corporations – 66 government agencies,
Founded in 2010: UCL, Southampton, Oxford and Bristol Key Objectives of the Consortium: Prove the concept of shared, regional e-infrastructure services.
Knowledge Management at the Gordon – Staff Portal Project Presented by Deirdre Carmichael 12 September 2008.
Financial System Replacement Project
1. 2 August Recommendation 9.1 of the Strategic Information Technology Advisory Committee (SITAC) report initiated the effort to create an Administrative.
Windows® Deployment Services
The Office of Information Technology ITA Meeting May 6, 2010 Harris Room (UC 2.212)
STFC and the UK e-Infrastructure Initiative The Hartree Centre Prof. John Bancroft Project Director, the Hartree Centre Member, e-Infrastructure Leadership.
Rob Allan Daresbury Laboratory NW-GRID Training Event 26 th January 2007 NW-GRID Future Developments R.J. Allan CCLRC Daresbury Laboratory.
WHICH TO CHOOSE RIGHT SERVER FOR THE RIGHT JOB. Today’s business environment demands that small and midsize businesses do more with less. The large majority.
Michael Hrybyk President/CEO An Overview.
Discussion of Infrastructure Clouds A NERSC Magellan Perspective Lavanya Ramakrishnan Lawrence Berkeley National Lab.
Joint CASC/CCI Workshop Report Strategic and Tactical Recommendations EDUCAUSE Campus Cyberinfrastructure Working Group Coalition for Academic Scientific.
The Internet2 NET+ Services Program Jerry Grochow Interim Vice President CSG January, 2012.
Advancing Alternative Energy Technologies Glenn MacDonell Director, Energy Industry Canada Workshop on Alternatives to Conventional Generation Technologies.
Towards a Virtual European Supercomputing Infrastructure Vision & issues Sanzio Bassini
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
University of California, Berkeley Opportunities and Challenges Michael Mundrane University of California, Berkeley 11 October 2010 Campus Leadership Engagement.
Information Technology Center Introduction to High Performance Computing at KFUPM.
Cisco and NetApp Confidential. Distributed under non-disclosure only. Name Date FlexPod Entry-level Solution FlexPod Value, Sized Right for Smaller Workloads.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
1 Scientific Cluster Support Program Steering Committee August 25, 2003 SCS Project Team.
Building a Cluster Support Service Implementation of the SCS Program UC Computing Services Conference Gary Jung SCS Project Manager
Academic and Research Technology (A&RT)
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Server and Short to Mid Term Storage Funding Research Computing Funding Issues.
Prepare for Change Ideas for Today and Tomorrow. Change is inevitable: Internal Factors Aging infrastructures Aging workforce Projects vs. programs New.
Cloud Computing Bill Turnbull Associate CIO for Advanced Technology and Systems Integration Department of Energy.
Status Report on Tier-1 in Korea Gungwon Kang, Sang-Un Ahn and Hangjin Jang (KISTI GSDC) April 28, 2014 at 15th CERN-Korea Committee, Geneva Korea Institute.
Research Cyberinfrastructure Alliance Working in partnership to enable computationally intensive, innovative, interdisciplinary research for the 21 st.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
CI Days: Planning Your Campus Cyberinfrastructure Strategy Russ Hobby, Internet2 Internet2 Member Meeting 9 October 2007.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Advancing Computational Science in Academic Institutions Organisers: Dan Katz – University of Chicago Gabrielle Allen – Louisiana State University Rob.
Outline IT Organization SciComp Update CNI Update
U.S. Department of Energy Office of Science Advanced Scientific Computing Research Program NERSC Users Group Meeting Department of Energy Update September.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
SG - OSG Improving Campus Research CI Through Leveraging and Integration: Developing a SURAgrid-OSG Collaboration John McGee, RENCI/OSG Engagement Coordinator.
What is Cyberinfrastructure? Russ Hobby, Internet2 Clemson University CI Days 20 May 2008.
Top Issues Facing Information Technology at UAB Sheila M. Sanders UAB Vice President Information Technology February 8, 2007.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Alessandra CiocioApril 6, CSAC meeting1 Mid-Range Computing Working Group Report CSAC and ITSD are working in partnership to determine the value.
A Data Centre for Science and Industry Roadmap. INNOVATION NETWORKING DATA PROCESSING DATA REPOSITORY.
Center for Advanced Energy Studies Harold S. Blackman Interim Director, CAES July 18, 2007.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
Adoption and Use of Electronic Medical Records (in Federally Qualified Health Centers) and Supporting an ASP Community Care Network of Virginia, Inc.
Funding: Staffing for Research Computing What staffing models does your institution use for research computing? How does your institution pay for the staffing.
Tony Doyle - University of Glasgow Introduction. Tony Doyle - University of Glasgow 6 November 2006ScotGrid Expression of Interest Universities of Aberdeen,
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
A. CiocioITSD/CSAC Retreat March 3, Scientific Cluster Support Program SCS Steering Committee Report.
Lawrence H. Landweber National Science Foundation SC2003 November 20, 2003
Seaborg Decommission James M. Craw Computational Systems Group Lead NERSC User Group Meeting September 17, 2007.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Cloud Computing ENG. YOUSSEF ABDELHAKIM. Agenda :  The definitions of Cloud Computing.  Examples of Cloud Computing.  Which companies are using Cloud.
Scrum BUT Research Adapting Agile principles in a research environment Gary Morgan Chief Science and Technology Office March 16, 2009 PNNL-SA
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
A Brief Introduction to NERSC Resources and Allocations
What is HPC? High Performance Computing (HPC)
Strategy for Complex Networking, IT, & Telecommunication Decisions
Volunteer Computing for Science Gateways
Scaling Science Communities Lessons learned by and future plans of the Open Science Grid Frank Würthwein OSG Executive Director Professor of Physics UCSD/SDSC.
Educational Information System Replacement Plan Overview
ESnet and Science DMZs: an update from the US
Introduce yourself Presented by
Presentation transcript:

DOE Perspective on Cyberinfrastructure - LBNL Gary Jung Manager, High Performance Computing Services Lawrence Berkeley National Laboratory Educause CCI Working Group Meeting November 5, 2009

2 November 5, 2009 Midrange Computing DOE ASCR hosted a workshop in Oct 2008 to assess the role of mid-range computing in the Office of Science and revealed that this computation continues to play an increasingly important role in enabling the Office of Science. Although it is not part of ASCR's mission, midrange computing, and the associated data management play a vital and growing role in advancing science in disciplines where capacity is as important as capability. Demand for midrange computing services is… o growing rapidly at many sites (>30% growth annually at LBNL) o the direct expression of a broad scientific need Midrange computing is a necessary adjunct to leadership-class facilities

3 November 5, 2009 Berkeley Lab Computing Gap between desktop and National Centers Midrange Computing Working Group 2001 Cluster support program started in 2002 o Services for PI-owned clusters include: Pre purchase consulting; development of specs and RFP, facilities planning, installation and configuration, ongoing cluster support, user services consulting, cybersecurity, computer room colocation Currently 32 clusters in production, over 1400 nodes, 6500 processor cores Funding: Institution provides support for infrastructure costs, technical development. Researchers pay for cluster and incremental cost of support.

4 November 5, 2009 Cluster Support Phase II: Perceus Metacluster All clusters interconnected into shared cluster infrastructure o Permits sharing of resources, storage  Global home file system o One ‘super master’ node, used to boot nodes across all clusters  multiple system images supported o One master job scheduler, submitting to all clusters o Simplifies provisioning new systems and ongoing support Metacluster model made possible by Perceus software o successor to Warewulf ( o can run jobs across clusters, recapturing stranded capacity.

5 November 5, 2009

6 Laboratory-Wide Cluster - Drivers “Computation lets us understand everything we do.” – LBNL Acting Lab Director Paul Alivisatos 38% of scientists depend on cluster computing for research. 69% of scientists are interested in cycles on a Lab-owned cluster. o early-career scientists twice as likely to be ‘very interested’ than later-career peers Why do scientists at LBNL need midrange computing resources? o ‘on ramp’ activities in preparation for running at supercomputing centers (development, debugging, benchmarking, optimization) o scientific inquiry not connected with ‘on ramp’ activities

7 November 5, 2009 Laboratory-Wide Cluster “Lawrencium” Overhead funded program o Capital equipment dollars shifted from business computing o Overhead funded staffing - 2 FTE Production in Fall 2008 General purpose Linux cluster suitable for a wide range of applications o 198-nodes, 1584 cores, DDR Infiniband interconnect o 40TB NFS home directory storage; 100TB Lustre parallel scratch o Commercial job scheduler and banking system o #500 on the Nov 2008 Top500 Open to all LBNL PIs and collaborators on their project Users are required to complete a survey when applying for accounts and later provide feedback on science results No user allocations at this time. This has been successful to date.

8 November 5, 2009 Networking - LBLNet Peer at 10GBE with ESNET 10GbE at core. Moving to 10GbE to the buildings Goal is sustained high speed data flows with cybersecurity Network based IDS approach - traffic is innocent until proven guilty o Reactive firewall o Does not impede data flow. no stateful firewall. o Bro cluster allows us to scale our IDS to 10GBE

9 November 5, 2009 Communications and Governance General announcements at IT council Steering committees used for scientific computing o Small group of stakeholders, technical experts, decision makers o Helps to validate and communicate decisions o Accountability

10 November 5, 2009 Challenges Funding (past) o Difficult for IT to shift funding from other areas of computing to support for science o Recharge can constrain adoption. Full cost recovery definitely will. New Technology (ongoing) Facilities (current) o Computer room is approaching capacity despite upgrades  Environmental Monitoring  Plenum in ceiling converted to hot air return  Tricks to boost underfloor pressure  Water cooled doors o Underway  DCIE measurement in process  Tower and heat exchanger replacement  Data Center container investigation

11 November 5, 2009 Next Steps Opportunities presented by cloud computing o Amazon investigation earlier this year. Others ongoing  Latency sensitive applications ran poorly as expected  Performance dependent of specific use case  Data migration. Economics of storing vs moving  Certain LBNL factors favor costs for build instead of buy Large storage and computation for data analysis GPU investigation

12 November 5, 2009 Points of Collaboration UC Berkeley HPCC o Recent high profile joint projects between UCB and LBNL encourages close collaboration o 25-30% of scientists have dual appointment o UC Berkeley proximity to LBNL facilitates the use of cluster services University of California Shared Research Computing Services pilot (SRCS) o LBNL and SDSC joint pilot for the ten UC campuses o Two 272-node clusters located at UC Berkeley and SDSC o Shared computing is more cost-effective o Dedicated CENIC L3 connecting network for integration o Pilot consists of 24 research projects

13 November 5, 2009