State of HCC 2012 Dr. David R. Swanson Director, Holland Computing Center.

Slides:



Advertisements
Similar presentations
October 2011 David Toback, Texas A&M University Research Topics Seminar 1 David Toback January 2015 Big Computing and the Mitchell Institute for Fundamental.
Advertisements

High Performance Computing Course Notes Grid Computing.
Building Campus HTC Sharing Infrastructures Derek Weitzel University of Nebraska – Lincoln (Open Science Grid Hat)
Research CU Boulder Cyberinfrastructure & Data management Thomas Hauser Director Research Computing CU-Boulder
Campus High Throughput Computing (HTC) Infrastructures (aka Campus Grids) Dan Fraser OSG Production Coordinator Campus Grids Lead.
Cloud Computing PRESENTED BY- Rajat Dixit (rd2392)
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
Holland Computing Center David R. Swanson, Ph.D. Director.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
NICLS: Development of Biomedical Computing and Information Technology Infrastructure Presented by Simon Sherman August 15, 2005.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Illinois Campus Cluster Program User Forum October 24, 2012 Illini Union Room 210 2:00PM – 3:30PM.
An Introduction to Cloud Computing. The challenge Add new services for your users quickly and cost effectively.
GridPP Steve Lloyd, Chair of the GridPP Collaboration Board.
IPlant Collaborative Tools and Services Workshop iPlant Collaborative Tools and Services Workshop Collaborating with iPlant.
SICSA student induction day, 2009Slide 1 Social Simulation Tutorial Session 6: Introduction to grids and cloud computing International Symposium on Grid.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
José M. Hernández CIEMAT Grid Computing in the Experiment at LHC Jornada de usuarios de Infraestructuras Grid January 2012, CIEMAT, Madrid.
100G R&D at Fermilab Gabriele Garzoglio (for the High Throughput Data Program team) Grid and Cloud Computing Department Computing Sector, Fermilab Overview.
1 Florida Cyberinfrastructure Development: SSERCA Fall Internet2 Meeting Raleigh, Va October 3, 2011 Paul Avery University of Florida
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
The Sharing and Training of HPC Resources at the University of Arkansas Amy Apon, Ph.D. Oklahoma Supercomputing Symposium October 4, 2006.
RNA-Seq 2013, Boston MA, 6/20/2013 Optimizing the National Cyberinfrastructure for Lower Bioinformatic Costs: Making the Most of Resources for Publicly.
The Future of the iPlant Cyberinfrastructure: Coming Attractions.
Discussion Topics DOE Program Managers and OSG Executive Team 2 nd June 2011 Associate Executive Director Currently planning for FY12 XD XSEDE Starting.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
Developing & Managing A Large Linux Farm – The Brookhaven Experience CHEP2004 – Interlaken September 27, 2004 Tomasz Wlodek - BNL.
Tools for collaboration How to share your duck tales…
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Open Science Grid (OSG) Introduction for the Ohio Supercomputer Center Open Science Grid (OSG) Introduction for the Ohio Supercomputer Center February.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Slide David Britton, University of Glasgow IET, Oct 09 1 Prof. David Britton GridPP Project leader University of Glasgow UK-T0 Meeting 21 st Oct 2015 GridPP.
CERN - IT Department CH-1211 Genève 23 Switzerland t High Availability Databases based on Oracle 10g RAC on Linux WLCG Tier2 Tutorials, CERN,
DiRAC-3 – The future Jeremy Yates, STFC DiRAC HPC Facility.
Regional Cyberinfrastructure Planning Great Plains Network Greg Monaco, Ph.D. Director for Research and Cyberinfrastructure Initiatives
Galaxy Community Conference July 27, 2012 The National Center for Genome Analysis Support and Galaxy William K. Barnett, Ph.D. (Director) Richard LeDuc,
LHC Computing, CERN, & Federated Identities
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 TCS Confidential. 2 Objective : In this session we will be able to learn:  What is Cloud Computing?  Characteristics  Cloud Flavors  Cloud Deployment.
A. Mohapatra, T. Sarangi, HEPiX-Lincoln, NE1 University of Wisconsin-Madison CMS Tier-2 Site Report D. Bradley, S. Dasu, A. Mohapatra, T. Sarangi, C. Vuosalo.
Derek Weitzel Grid Computing. Background B.S. Computer Engineering from University of Nebraska – Lincoln (UNL) 3 years administering supercomputers at.
Pathway to Petaflops A vendor contribution Philippe Trautmann Business Development Manager HPC & Grid Global Education, Government & Healthcare.
The RAL PPD Tier 2/3 Current Status and Future Plans or “Are we ready for next year?” Chris Brew PPD Christmas Lectures th December 2007.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
IPCEI on High performance computing and big data enabled application: a pilot for the European Data Infrastructure Antonio Zoccoli INFN & University of.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Campus Grid Technology Derek Weitzel University of Nebraska – Lincoln Holland Computing Center (HCC) Home of the 2012 OSG AHM!
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
March 2014 Open Science Grid Operations A Decade of HTC Infrastructure Support Kyle Gross Operations Support Lead Indiana University / Research Technologies.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Computing infrastructures for the LHC: current status and challenges of the High Luminosity LHC future Worldwide LHC Computing Grid (WLCG): Distributed.
How to get the needed computing Tuesday afternoon, 1:30pm Igor Sfiligoi Leader of the OSG Glidein Factory Operations University of California San Diego.
Elastic Cyberinfrastructure for Research Computing
HCC Fall KickStart University of Nebraska – Lincoln Holland Computing Center David Swanson, Emelie Harstad, Jingchao Zhang, Adam Caprez, Derek Weitzel,
A Brief Introduction to NERSC Resources and Allocations
What is HPC? High Performance Computing (HPC)
Ian Bird WLCG Workshop San Francisco, 8th October 2016
Clouds , Grids and Clusters
Scaling Science Communities Lessons learned by and future plans of the Open Science Grid Frank Würthwein OSG Executive Director Professor of Physics UCSD/SDSC.
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
Storage Trends: DoITT Enterprise Storage
Introduce yourself Presented by
This work is supported by projects Research infrastructure CERN (CERN-CZ, LM ) and OP RDE CERN Computing (CZ /0.0/0.0/1 6013/ ) from.
Presentation transcript:

State of HCC 2012 Dr. David R. Swanson Director, Holland Computing Center

Nature Communications, July 17, 2012 Nebraska Supercomputing Symposium 2012

HCC CPU Hour Usage 2012 Nebraska Supercomputing Symposium 2012 Zeng (Quant Chem) 4.5M Starace (AMO Phys) 2.7M Rowe (Climate) 2.0M NanoScience 6.4M Comp Bio 3.0M Comp Sci 1.7M Physics 0.7M Mech E 0.4M

High Performance Computing Xiao Zeng, Chemistry, UNL (prior slide) DFT and Car Parrinello MD HPC – tightly coupled codes Requires expensive low-latency local network (infiniband) Requires high-performance storage (Panasas, Lustre) Requires highly reliable hardware Nebraska Supercomputing Symposium 2012

Eureka! A Higgs! (or at least something currently indistinguishable) "I think we have it. We have discovered a particle that is consistent with a Higgs boson." –CERN Director-General Rolf Heuer Nebraska Supercomputing Symposium 2012

US CMS Tier2 Computing Nebraska Supercomputing Symposium 2012

Compact Muon Solenoid (CMS) 5.5 mi Large Hadron Collider Nebraska Supercomputing Symposium 2012

CMS Grid Computing Model Nebraska Supercomputing Symposium 2012

Eureka! A Higgs! (or at least something currently indistinguishable) Ca. 50 PB of CMS data in entirety Over 1 PB currently at HCC’s “Tier2”, 3500 cores Collaboration at many scales –HCC and Physics Department –Over 2700 scientists worldwide –International Grid Computing Infrastructure –Data grid as well –UNL closely linked to KU, KSU physicists via a jointly hosted “Tier3” Nebraska Supercomputing Symposium 2012

Data Intensive HTC Huge database Requires expensive high-bandwidth wide area network (dwdm fiber) Requires high-capacity storage (HDFS, dCache) HTC – loosely coupled codes Requires hardware Nebraska Supercomputing Symposium 2012

Outline HCC Overview New User report HCC-Go Moving Forward (after break) –Next purchase –It’s the Data, stupid… –Other Issues Nebraska Supercomputing Symposium 2012

Outline New User report HCC-Go Moving Forward (next section) –Next purchase (motivation) –New Communities –PIVOT –It’s the Data, stupid… Nebraska Supercomputing Symposium 2012

HOLLAND COMPUTING CENTER OVERVIEW Nebraska Supercomputing Symposium 2012

NU Holland Computing Center has a University-wide mission to –Facilitate and perform computational and data intensive research –Engage and train NU researchers, students, and other state communities –This includes you! –HCC would be delighted to collaborate Nebraska Supercomputing Symposium 2012

Computational Science – 3 rd Pillar Experiment Theory Computation/Data Nebraska Supercomputing Symposium 2012

Lincoln Resources 10 staff Red Sandhills 5,000 compute cores 3 PetaBytes storage in HDFS Nebraska Supercomputing Symposium 2012

Sandhills “Condominium Cluster” 44 nodes X 32-core, 128 GB, IB Lustre (175 TB) Priority Access –$HW + $50/month –4 groups currently SLURM Nebraska Supercomputing Symposium 2012

Omaha Resources 3 Staff Firefly Tusker 10,000 compute cores 500 TB storage New offices soon: 158J PKI Nebraska Supercomputing Symposium 2012

Tusker 106*64= 6784 cores 256 GB/node 2 nodes w/ 512 GB 360 TB Lustre –100 TB more en route QDR IB 43 TFlop Nebraska Supercomputing Symposium 2012

Tusker ¼ footprint of Firefly ¼ the power 2X the TFLOPS 2X the storage Fully utilized Maui/Torque Nebraska Supercomputing Symposium 2012

In between … HCC (UNL) to Internet2: 10 gbps HCC (Schorr) to HCC (PKI): 20 gbps Allows us to do some interesting things –“overflow” jobs to/from Red –DYNES project –Xrootd mechanism Nebraska Supercomputing Symposium 2012

HCC Staff HPC Applications Specialists –Dr. Adam Caprez –Dr. Ashu Guru –Dr. Jun Wang –Dr. Nicholas Palermo System Administrators –Dr. Carl Lundstedt –Garhan Attebury –Tom Harvill –John Thiltges –Josh Samuelson –Dr. Brad Hurst Nebraska Supercomputing Symposium 2012

HCC Staff Other Staff –Dr. Brian Bockelman –Joyce Young GRAs –Derek Weitzel –Chen He –Kartik Vedalaveni –Zhe Zhang Undergraduates –Carson Crawford –Kirk Miller –Avi Knecht –Phil Brown –Slav Ketsman –Nicholas Nachtigal –Charles Cihacek Nebraska Supercomputing Symposium 2012

HCC Campus Grid Holland Computing Center resources are combined into an HTC campus grid –10,000 cores, 500 TB in Omaha –5,000 cores, 3 PB in Lincoln –All tied together via a single submission protocol using OSG software stack –Straightforward to expand to OSG sites across the country, as well as to EC2 (cloud) –HPC jobs get priority; HTC ensures high utilization Nebraska Supercomputing Symposium 2012

HCC Model for a Campus Grid Me, my friends and everyone else Grid Campus Local 25 Nebraska Supercomputing Symposium 2012

HCC & Open Science Grid National, distributed computing partnership for data- intensive research –Opportunistic computing –Over 100,000 cores –Supports the LHC experiments, other science –Funded for 5 more years –Over 100 sites in the Americas –Ongoing support for 2.5 (+3) FTE at HCC Nebraska Supercomputing Symposium 2012

It Works! Nebraska Supercomputing Symposium 2012

HCC Networking Monitoring Nebraska Supercomputing Symposium 2012

OSG Resources Nebraska Supercomputing Symposium 2012

Working philosophy Use what we buy –These pieces of infrastructure are linked, but improve asynchronously –Depreciation is immediate –Leasing is still more expensive (for now) –Buying at fixed intervals mitigates risk, increases ROI –Space, Power and Cooling have a longer life span Share what we aren’t using –Share opportunistically – retain local ownership –Consume opportunistically – there is more to gain! –Collaborators, not just consumers –Greater good vs. squandered opportunity Nebraska Supercomputing Symposium 2012

Working philosophy A Data deluge is upon us Support is essential –If you only build it, they still may not come –Build incrementally and buy time for user training –Support can grow more gradually than hardware Links to national and regional infrastructure are critical –Open Source Community –GPN access to Internet2 –Access to OSG, XSEDE resources –Collaborations with fellow OSG experts –LHC Nebraska Supercomputing Symposium 2012

HCC New Users FY UNL- City UNL- East UNOUNMC Outside NU system (74)33 (10)75 (19)30 (17)112 (26) (95)50 (17)105 (30)35 (5)130 (18) Nebraska Supercomputing Symposium 2012

New User Communities Theatre, Fine Arts/Digital Media, Architecture Psychology, Finance UNMC Puerto Rico PIVOT collaborators Nebraska Supercomputing Symposium 2012

HCC NEW USER REPORT: HEATH ROEHR Nebraska Supercomputing Symposium 2012

HCC-GO : DR. ASHU GURU Nebraska Supercomputing Symposium 2012

MOVING FORWARD Nebraska Supercomputing Symposium 2012

NEW PURCHASE Nebraska Supercomputing Symposium 2012

$2M for … More computing –need ca. 100 TF to hit Top500 for Jun 2013 –Likely use all of funds to hit that amount More storage –Near-line archive (9 PB) –HDFS Specialty hardware –GPGPU/Viz –Mic hardware Nebraska Supercomputing Symposium 2012

More computing How much RAM/core? Currently almost always oversubscribed Large scale jobs almost impossible (> 2000 core) Safest investment – will use right away Firefly due to be retired soon – EOL Nebraska Supercomputing Symposium 2012

More computing Nebraska Supercomputing Symposium 2012

More Computing Nebraska Supercomputing Symposium 2012

More storage Most rapidly growing demand Growing contention, can’t just queue up Largest unmet need (?) Nebraska Supercomputing Symposium 2012

Storage for $2M $2M HDFS cluster –250 nodes –4000 cores (Intel) –9.0 PB (RAW) –128 GB / node Nebraska Supercomputing Symposium 2012

Other options GPGPUs most Green option for computing Highest upside for raw power (Top500) Mic even compatible with x86 codes SMP uniquely meets some needs, easiest to use/program Bluegene, Tape silo, … Nebraska Supercomputing Symposium 2012

HCC personnel timeline Nebraska Supercomputing Symposium 2012

HCC networking timeline Nebraska Supercomputing Symposium 2012

HCC cpu timeline 900X Nebraska Supercomputing Symposium 2012

HCC storage timeline 30,000X Nebraska Supercomputing Symposium 2012

Composite Timeline Data increase/ CPU Cores = 33 Data increase/ WAN bandwidth = 150 It takes a month to move 3 PB at 10 Gb/sec Power < 100X increase, largely constant last 3 years Nebraska Supercomputing Symposium 2012

Storage at HCC Affordable, Reliable, High Performance, High Capacity –Pick 2 –So multiple options /home /work /shared Currently, no /archive Nebraska Supercomputing Symposium 2012

/home Reliable Low performance –No W from workers ZFS Rsync’ed pair, one in Omaha, one in Lincoln Backed up incrementally, requires severe quotas Nebraska Supercomputing Symposium 2012

/work High performance High(er) capacity Not permanent storage Lenient quotas More robust, more reliable “scratch space” Subject to purge as needed Nebraska Supercomputing Symposium 2012

/share Purchased by given group Exported to both Lincoln and Omaha machines Usually for capacity, striped for some reliability Nebraska Supercomputing Symposium 2012

Storage Strategy Maintain /home for precious files –Could be global Maintain /work for runtime needs –Remain local to cluster Create /share for near-line archive –3-5 year time frame (or less) –Use for accumulating intermediate data, then purge –Global access Nebraska Supercomputing Symposium 2012

Storage strategy Permanent archival has 3 options –1) library –2) Amazon glacier Currently $120/TB/year –3) tape system Nebraska Supercomputing Symposium 2012

HCC Data Visualizations Fish! HadoopViz OSG Google Earth Web-based monitoring – – Nebraska Supercomputing Symposium 2012

Other discussion topics Maui vs. SLURM Queue length policy Education approaches –This (!) –Tutuorials (next!) –Afternoon workshops –Semester courses –Individual presentations/meetings –Online materials Nebraska Supercomputing Symposium 2012

©2007 The Board of Regents of the University of Nebraska NU Administration (UNL, NRI) NSF, DOE, EPSCoR, OSG Holland Foundation CMS: Ken Bloom, Aaron Dominguez HCC: Drs. Brian Bockelman, Adam Caprez, Ashu Guru, Brad Hurst, Carl Lundstedt, Nick Palmero, Jun Wang. Garhan Attebury, Tom Harvill, Josh Samuelson, John Thiltges Chen He, Derek Weitzel