Research Computing University Of South Florida Providing Advanced Computing Resources for Research and Instruction through Collaboration.

Slides:



Advertisements
Similar presentations
Founded in 2010: UCL, Southampton, Oxford and Bristol Key Objectives of the Consortium: Prove the concept of shared, regional e-infrastructure services.
Advertisements

Statewide IT Conference30-September-2011 HPC Cloud Penguin on David Hancock –
Take your CMS to the cloud to lighten the load Brett Pollak Campus Web Office UC San Diego.
CURRENT AND FUTURE HPC SOLUTIONS. T-PLATFORMS  Russia’s leading developer of turn-key solutions for supercomputing  Privately owned  140+ employees.
HPCC Mid-Morning Break High Performance Computing on a GPU cluster Dirk Colbry, Ph.D. Research Specialist Institute for Cyber Enabled Discovery.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
IDC HPC User Forum Conference Appro Product Update Anthony Kenisky, VP of Sales.
Duke Atlas Tier 3 Site Doug Benjamin (Duke University)
WEST VIRGINIA UNIVERSITY HPC and Scientific Computing AN OVERVIEW OF HIGH PERFORMANCE COMPUTING RESOURCES AT WVU.
LinkSCEEM-2: A computational resource for the development of Computational Sciences in the Eastern Mediterranean Mostafa Zoubi SESAME SESAME – LinkSCEEM.
1 Intellectual Architecture Leverage existing domain research strengths and organize around multidisciplinary challenges Institute for Computational Research.
Academic and Research Technology (A&RT)
Bill Wrobleski Director, Technology Infrastructure ITS Infrastructure Services.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
Institutional Research Computing at WSU: Implementing a community-based approach Exploratory Workshop on the Role of High-Performance Computing in the.
Communicating with Users about HTCondor and High Throughput Computing Lauren Michael, Research Computing Facilitator HTCondor Week 2015.
Server and Short to Mid Term Storage Funding Research Computing Funding Issues.
Project Overview:. Longhorn Project Overview Project Program: –NSF XD Vis Purpose: –Provide remote interactive visualization and data analysis services.
Research Cyberinfrastructure Alliance Working in partnership to enable computationally intensive, innovative, interdisciplinary research for the 21 st.
LARGE SCALE DEPLOYMENT OF DAP AND DTS Rob Kooper Jay Alemeda Volodymyr Kindratenko.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
DRAFT 1 Institutional Research Computing at WSU: A community-based approach Governance model, access policy, and acquisition strategy for consideration.
A Cloud is a type of parallel and distributed system consisting of a collection of inter- connected and virtualized computers that are dynamically provisioned.
Research Support Services Research Support Services.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
27 May 2004 C.N. Papanicolas EGEE and the role of IASA ( In close collaboration with UOA ) IASA GRID Steering Committee: George Kallos Lazaros.
Library Publishing Services: Strategies for Success Charleston Library Conference November 3, 2011.
1 Florida Cyberinfrastructure Development: SSERCA Fall Internet2 Meeting Raleigh, Va October 3, 2011 Paul Avery University of Florida
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
The Research Computing Center Nicholas Labello
IPlant cyberifrastructure to support ecological modeling Presented at the Species Distribution Modeling Group at the American Museum of Natural History.
Taking the Complexity out of Cluster Computing Vendor Update HPC User Forum Arend Dittmer Director Product Management HPC April,
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Grid Computing at The Hartford Condor Week 2008 Robert Nordlund
Looking Ahead: A New PSU Research Cloud Architecture Chuck Gilbert - Systems Architect and Systems Team Lead Research CI Coordinating Committee Meeting.
SouthGrid SouthGrid SouthGrid is a distributed Tier 2 centre, one of four setup in the UK as part of the GridPP project. SouthGrid.
CCS Overview Rene Salmon Center for Computational Science.
UCSB Projects & Progress 2011 UC Santa Barbara Projects & Progress 2010 A brief look at some of the things we’ve been working on this past year.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
The LBNL Perceus Cluster Infrastructure Next Generation Cluster Provisioning and Management October 10, 2007 Internet2 Fall Conference Gary Jung, SCS Project.
Cyberinfrastructure: An investment worth making Joe Breen University of Utah Center for High Performance Computing.
11/15/04PittGrid1 PittGrid: Campus-Wide Computing Environment Hassan Karimi School of Information Sciences Ralph Roskies Pittsburgh Supercomputing Center.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
ComPASS Summary, Budgets & Discussion Panagiotis Spentzouris, Fermilab ComPASS PI.
Galaxy Community Conference July 27, 2012 The National Center for Genome Analysis Support and Galaxy William K. Barnett, Ph.D. (Director) Richard LeDuc,
What’s Coming? What are we Planning?. › Better docs › Goldilocks – This slot size is just right › Storage › New.
National Computational Science Ky PACS at the University of Kentucky April 2000 –Advanced Computing Resources –EPSCoR Outreach –SURA Liaison –John.
The Evolution of the Italian HPC Infrastructure Carlo Cavazzoni CINECA – Supercomputing Application & Innovation 31 Marzo 2015.
Architecture of a platform for innovation and research Erik Deumens – University of Florida SC15 – Austin – Nov 17, 2015.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
Berkeley Lab Software Distribution Site NLIT Dan Pulsifer - Engineering May 11 th, 2008.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
An Brief Introduction Charlie Taylor Associate Director, Research Computing UF Research Computing.
NIIF HPC services for research and education
Elastic Cyberinfrastructure for Research Computing
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
What is HPC? High Performance Computing (HPC)
Heterogeneous Computation Team HybriLIT
Clouds of JINR, University of Sofia and INRNE Join Together
Low-Cost High-Performance Computing Via Consumer GPUs
GGF15 – Grids and Network Virtualization
XSEDE’s Campus Bridging Project
System G And CHECS Cal Ribbens
Shared Research Computing Policy Advisory Committee (SRCPAC)
Introduce yourself Presented by
Storing and Accessing G-OnRamp’s Assembly Hubs outside of Galaxy
Introduction to High Performance Computing Using Sapelo2 at GACRC
Welcome to (HT)Condor Week #19 (year 34 of our project)
Presentation transcript:

Research Computing University Of South Florida Providing Advanced Computing Resources for Research and Instruction through Collaboration

Mission Provide advanced computing resources required by a major research university o Software o Hardware o Training o Support

User Base 40 Research groups 6 Colleges 100 faculty 300 students

Hardware System was build on the condominium model and consists of 300 Nodes 2400 Processors o University provides infrastructure and some computational resources o Faculty funding provides bulk of computational resources

Software Over 50 scientific codes o Installation o Integration o Upgrades o Licensing

Support Personnel Provide all systems administration Software support One-on-one consulting System efficiency improvements Users are no longer just the traditional “number crunchers

Current Projects Consolidating the last standalone cluster (of appreciable size) Advanced Visualization Center o Group of 19 Faculty applied for funding o Personnel o Training o Large Resolution 3D display

Current Projects New computational resources o Approximately 100 nodes o GPU resources o Upgrade parallel file system Virtual Clusters o HPC for the other 90 % FACC

Florida State University's Shared HPC Building and Maintaining Sustainable Research Computing at FSU

Shared-FSU HPC Mission Support multidisciplinary research Provide a general access computing platform Encourage cost sharing by departments with dedicated computing needs Provide a broad base of support and training opportunities

Turn-key Research Solution Participation is Voluntary University provides staffing University provides general infrastructure o Network fabrics o Racks o Power/Cooling Additional buy-in incentives o Leverage better pricing as a group o Matching funds Offer highly flexible buy-in options o Hardware purchase only o Short-term Service Level Agreements o Long-term Service Level Agreements Shoot for 50% of hardware costs covered by Buy-in

Research FSU 500 plus users 33 Academic Units 5 Colleges

HPC Owner Groups 2007 o Department of Scientific Computing o Center for Ocean-Atmosphere Prediction Studies o Department of Meteorology 2008 o Gunzburger Group (Applied Mathematics) o Taylor Group (Structural Biology) o Department of Scientific Computing o Kostov Group (Chemical & Biomedical Engineering) 2009 o Department of Physics (HEP, Nuclear, etc.) o Institute of Molecular Biophysics o Bruschweiler Group (National High Magnetic Field Laboratory) o Center for Ocean-Atmosphere Prediction Studies (with the Department of Oceanography) o Torrey Pines Institute of Molecular Studies 2010 o Chella Group (Chemical Engineering) o Torrey Pines Institute of Molecular Studies o Yang Group (Institute of Molecular Biophysics) o Meteorology Department o Bruschweiler Group o Fajer Group (Institute of Molecular Biophysics) o Bass Group (Biology)

Research FSU Publications o Macromolecules o Bioinformatics o Systematic Biology o Journal of Biogeography o Journal of Applied Remote Sensing o Journal of Chemical Theory and Computation o Physical Review Letters o Journal of Physical Chemistry o Proceeding of the National Academy of Science o Biophysical Journal o Journal Chemical Theory Computation o Journal: J. Phys. Chem. o PLoS Pathogens o Journal of Virology o Journal of the American Chemical Society o The Journal of Chemical Physics o PLoS Biology o Ocean Modeling o Journal of Computer-Aided Molecular Design

Sliger Data Center Shared- HPC pfs FSU’s Shared-HPC Stage 1: Infiniband Connected Cluster

Single and Multiprocessor Usage Year 1

DSL Building Sliger Data Center Shared- HPC pfs Condor FSU’s Shared-HPC Stage 2: Alternative Backfilling

Backfilling Single Proc Jobs on Non- HPC Resources Using Condor

Condor Usage ~1000 processor cores available for single processor computations 2,573,490 processor hours used since Condor was made available to all HPC users in September Seven users have been using Condor from HPC Dominate users are Evolutionary Biology, Molecular Dynamics, and Statistics (same users that were submitting numerous single proc. jobs) Two workshop introducing it to HPC users

Single vs. Multi-processor Jobs Year 2

Single vs. Multi-processor Jobs Year 3

DSL Building Sliger Data Center Shared- HPC pfs Condor SMP FSU’s Shared-HPC Stage 3: Scalable SMP

One MOAB Queue for SMP or very large memory jobs Three “nodes” o M905 blade with 16 cores and 64GB mem o M905 blade with 24 cores and 64GB mem o 3Leaf system with up to 132 cores and 528 GB mem

DSL Building DSL Data Center Sliger Data Center Shared- HPC pfs Condor SMP 2° fs Vis

Interactive Cluster Functions Facilitates data exploration Provides venue for software not well suited for a batch scheduled environment o (e.g., some MatLab, VMD, R, Python, etc.) Provides access to hardware not typically found on standard desktops/laptops/mobile devises (e.g. lots of memory, high-end GPUs) Provides licensing and configuration support for software applications and libraries

Interactive Cluster Hardware Layout 8 high-end CPU based host nodes o Multi-core Intel or AMD processors o 4 to 8 GB of memory per core o 16X PCIe connectivity o QDR IB connectivity to Luster storage o IP (read-only) connectivity to Panasas o 10 Gbps connectivity to campus network backbone One C410x external PCI chassis o Compact o IPMI management o Supports up to 16 NVIDIA Tesla M2050  Up to teraflops

DSL Building DSL Data Center Sliger Data Center Shared- HPC pfs Condor SMP 2° fs Vis Db.Web

Web/Database Hardware Function Facilitates creation of Data analysis Pipelines/Workflows Favored by external funding agencies o Demonstrated cohesive Cyberinfrastructure o Fits well into required Data Management Plans (NSF) Intended to facilitate access to data on Secondary storage or cycles on owner share of HPC Basic Software Install, no development support Bare Metal or VM

Web/Database Hardware Examples

FSU Research CI HPC HTC SMP 1° storage 2° Storage Vis and interactive DB and Web

Florida State University's Shared HPC Universities are by design multifaceted and lack a singular focus of support Local HPC resources should also be multifaceted and have a broad basis of support

HPC Summit University of Florida HPC Center

HPC Summit Short history Started in Phase I: CLAS – Avery – OIT 2005 Phase IIb: o COE – 9 investors 2007 Phase IIb: o COE – 3 investors 2009 Phase III: o DSR – 17 investors - ICBR - IFAS 2011 Phase IV: o 22 investors

HPC Summit Budget Total budget o $0.7 M o $1.8 M o $0.3 M o $1.2 M o $1.6 M o $0.4 M o $0.9 M

HPC Summit Hardware 4,500 cores 500 TB storage InfiniBand connected In three machine rooms o Connected by 20 Gbit/sec Campus Research Network

HPC Summit System software RedHat Enterprise Linux o through free CentOS distribution o upgrade once per year Lustre file system o mounted on all nodes o Scratch only o Provide backup through CNS service  Requires separate agreement between researcher and CNS

HPC Summit Other software Moab scheduler (commercial license) Intel compilers (commercial license) Numerous applications o Open and commercial

HPC Summit Operation Shared cluster some hosted systems 300 users 90% - 95% utilization

HPC Summit Investor Model Normalized Computing Unit o $400 per NCU o Is one core o In fully functional system (RAM, disk, shared file system) o For 5 years

HPC Summit Investor Model Optional Storage Unit o $140 per OSU o 1 TB of file storage (RAID) on one of a few global parallel file systems (Lustre) o For 1 year

HPC Summit Other options Hosted system o Buy all hardware, we operate o No sharing Pay as you go o Agree to pay monthly bill o Equivalent (almost) to $400 NCU prorated on a monthly basis Or rates are cents per hour o Cheaper than Amazon Elastic Cloud

Mission Statement UM CCS is establishing nationally and internationally recognized research programs, focusing on those of an interdisciplinary nature, and actively engaging in computational research to solve the complex technological problems of modern society. We provide a framework for promoting collaborative and multidisciplinary activities across the University and beyond

CCS overview Started in June 2007 Faculty Senate approval in 2008 Four Founding Schools: A&S, CoE, RSMAS, Medical Offices in all Campus ~30 FTEs Data Center at the NAP of Americas

UM CCS Research Programs and Cores Physical Science & Engineering Computational Biology & Bioinformatics Data Mining Visualization Computational Chemistry Software Engineering High Performance Computing Social Systems Informatics

Quick Facts Over 1,000 UM users 5,200 cores of Linux Based Cluster 1,500 cores of Power-based Cluster ~2.0 PT of Storage 4.0 PT of Back-up More at: o C4 o

High Performance Computing UM Wide Resource Provides Academic Community & Research Partners with Comprehensive HPC Resources: o Hardware & Scientific Software Infrastructure o Expertise in Designing & Implementing HPC Solutions o Designing & Porting Algorithms & Programs to Parallel Computing Models Open Access of compute processing (first come serve) o Peer Review for large projects – Allocation Committee o Cost Center for priority access HPC services o Storage Cloud o Visualization and Data Analysis Cloud o Processing Cloud