A High-Performance Campus-Scale Cyberinfrastructure: The Technical, Political, and Economic Presentation by Larry Smarr to the NSF Campus Bridging Workshop.

Slides:



Advertisements
Similar presentations
Health Sciences Driving UCSD Research Cyberinfrastructure Invited Talk UCSD Health Sciences Faculty Council UC San Diego April 3, 2012 Dr. Larry Smarr.
Advertisements

High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting Stem Cell Research Invited Presentation Sanford Consortium for Regenerative.
“End-to-end Optical Fiber Cyberinfrastructure for Data-Intensive Research: Implications for Your Campus” Featured Speaker EDUCAUSE 2010 Anaheim Convention.
Sequencing Genomics: The New Big Data Driver IntermezzoTalk SURFnet7, Part of GigaPort3 Utrecht, Netherlands December 7, 2011 Dr. Larry Smarr Director,
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biomedical Sciences Joint Presentation UCSD School of Medicine Research Council.
Calit2-Living in the Future " Keynote Sharecase 2006 University of California, San Diego March 29, 2006 Dr. Larry Smarr Director, California Institute.
Bringing Mexico Into the Global LambdaGrid Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E. Gruber.
High Performance Cyberinfrastructure Enables Data-Driven Science in the Globally Networked World Keynote Presentation Sequencing Data Storage and Management.
High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC.
Why Optical Networks Are Emerging as the 21 st Century Driver Scientific American, January 2001.
The First Year of Cal-(IT) 2 Report to The University of California Regents UCSF San Francisco, CA March 13, 2002 Dr. Larry Smarr Director, California.
UNIVERSITY OF CALIFORNIA, SAN DIEGO SAN DIEGO SUPERCOMPUTER CENTER UC Grid Summit -- April 1, 2009 UC San Diego Campus Grid Update Shava Smallen San Diego.
“Building US/Mexico Collaborations Using Optical Networks” Opening Workshop Welcome Big Data Big Network 2 Calit2’s Qualcomm Institute February 10, 2014.
SAN DIEGO SUPERCOMPUTER CENTER Emerging HIPAA and Protected Data Requirements for Research Computing at SDSC Ron Hawkins Director of Industry Relations.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO HPEC September 16, 2010 Dr. Mike Norman, PI Dr. Allan Snavely, Co-PI.
SAN DIEGO SUPERCOMPUTER CENTER Choonhan Youn Viswanath Nandigam, Nancy Wilkins-Diehr, Chaitan Baru San Diego Supercomputer Center, University of California,
SAN DIEGO SUPERCOMPUTER CENTER Niches, Long Tails, and Condos Effectively Supporting Modest-Scale HPC Users 21st High Performance Computing Symposia (HPC'13)
PRISM: High-Capacity Networks that Augment Campus’ General Utility Production Infrastructure Philip Papadopoulos, PhD. Calit2 and SDSC.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Status Update TeraGrid Science Advisory Board Meeting July 19, 2010 Dr. Mike.
Cyber UC San Diego Elazar C. Harel May 14, 2008.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
© UC Regents 2010 Extending Rocks Clusters into Amazon EC2 Using Condor Philip Papadopoulos, Ph.D University of California, San Diego San Diego Supercomputer.
“Introduction to UC San Diego’s Integrated Digital Infrastructure” Opening Talk IDI Showcase 2015 University of California, San Diego May 6-7, 2015 Dr.
“High Performance Cyberinfrastructure Enables Data-Driven Science in the Globally Networked World” Invited Speaker Grand Challenges in Data-Intensive Discovery.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
“How LambdaGrids are Transforming Science" Keynote iGrid2005 La Jolla, CA September 29, 2005 Dr. Larry Smarr Director, California Institute.
“Mapping the Human Gut Microbiome in Health and Disease Using Sequencing, Supercomputing, and Data Analysis” Invited Talk Delivered by Mehrdad Yazdani,
Institute for Digital Research and Education (IDRE) UCLA’s CI Vision Research CI DataNet Cyber Learning Institute for Digital Research and Education (IDRE)
Source: Jim Dolgonas, CENIC CENIC is Removing the Inter-Campus Barriers in California ~ $14M Invested in Upgrade Now Campuses Need to Upgrade.
“An Integrated Science Cyberinfrastructure for Data-Intensive Research” Panel CISCO Executive Symposium San Diego, CA June 9, 2015 Dr. Larry Smarr Director,
“Calit2: A UC Experiment for Living in the Future" Talk to UCSD Near You La Jolla, CA April 11, 2006 Dr. Larry Smarr Director, California Institute.
“Creating a High Performance Cyberinfrastructure to Support Analysis of Illumina Metagenomic Data” DNA Day Department of Computer Science and Engineering.
Developing a North American Global LambdaGrid Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E.
“Comparative Human Microbiome Analysis” Remote Video Talk to CICESE Big Data, Big Network Workshop Ensenada, Mexico October 10, 2013 Dr. Larry Smarr Director,
National Science Foundation DMR ITR Computational Review and Workshop: ITR and beyond Daryl Hess, NSF Bruce Taggart, NSF June 17-19, 2004 Urbana, IL.
Cal-(IT) 2 : A Public-Private Partnership in Southern California U.S. Business Council for Sustainable Development Year-End Meeting December 11, 2003 Institute.
Introduction to Calit2 Visit by NASA Ames February 29, 2008 Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology.
Chicago/National/International OptIPuter Infrastructure Tom DeFanti OptIPuter Co-PI Distinguished Professor of Computer Science Director, Electronic Visualization.
Information Resources and Communications University of California, Office of the President UC-Wide Activities in Support of Research and Scholarship David.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA, SAN DIEGO Michael L. Norman Principal Investigator Interim Director, SDSC Allan Snavely.
Innovative Research Alliances Invited Talk IUCRP Fellows Seminar UCSD La Jolla, CA July 10, 2006 Dr. Larry Smarr Director, California Institute for Telecommunications.
Using Photonics to Prototype the Research Campus Infrastructure of the Future: The UCSD Quartzite Project Philip Papadopoulos Larry Smarr Joseph Ford Shaya.
SoCal Infrastructure OptIPuter Southern California Network Infrastructure Philip Papadopoulos OptIPuter Co-PI University of California, San Diego Program.
Russ Hobby Program Manager Internet2 Cyberinfrastructure Architect UC Davis.
A High-Performance Campus-Scale Cyberinfrastructure For Effectively Bridging End-User Laboratories to Data-Intensive Sources Presentation by Larry Smarr.
Information Technology Infrastructure Committee (ITIC) Report to the NAC March 8, 2012 Larry Smarr Chair ITIC.
SAN DIEGO SUPERCOMPUTER CENTER SDSC's Data Oasis Balanced performance and cost-effective Lustre file systems. Lustre User Group 2013 (LUG13) Rick Wagner.
“Big Data” and Data-Intensive Science (eScience) Ed Lazowska Bill & Melinda Gates Chair in Computer Science & Engineering University of Washington July.
“ Calit2-Living in the Future " Briefing The Future in Review (FiRe) 2006 Conference University of California, San Diego May 15, 2006 Dr. Larry Smarr Director,
UC San Diego Research CyberInfrastructure Strategy Charlotte Klock CSG Meeting June 15-17, 2011.
Project GreenLight Overview Thomas DeFanti Full Research Scientist and Distinguished Professor Emeritus California Institute for Telecommunications and.
Ocean Sciences Cyberinfrastructure Futures Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technologies Harry E.
The OptIPuter Project Tom DeFanti, Jason Leigh, Maxine Brown, Tom Moher, Oliver Yu, Bob Grossman, Luc Renambot Electronic Visualization Laboratory, Department.
OptIPuter Networks Overview of Initial Stages to Include OptIPuter Nodes OptIPuter Networks OptIPuter Expansion OPtIPuter All Hands Meeting February 6-7.
“The Pacific Research Platform: a Science-Driven Big-Data Freeway System.” Big Data for Information and Communications Technologies Panel Presentation.
“CAMERA Goes Live!" Presentation with Craig Venter National Press Club Washington, DC March 13, 2007 Dr. Larry Smarr Director, California Institute for.
Cyberinfrastructure Overview Russ Hobby, Internet2 ECSU CI Days 4 January 2008.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
“The UCSD Big Data Freeway System” Invited Short Talk Workshop on “Enriching Human Life and Society” UC San Diego February 6, 2014 Dr. Larry Smarr Director,
“ OptIPuter Year Five: From Research to Adoption " OptIPuter All Hands Meeting La Jolla, CA January 22, 2007 Dr. Larry Smarr Director, California.
Southern California Infrastructure Philip Papadopoulos Greg Hidley.
“Genomics: The CAMERA Project" Invited Talk 5 th Annual ON*VECTOR International Photonics Workshop UCSD February 28, 2006 Dr. Larry Smarr Director,
“Pacific Research Platform Science Drivers” Opening Remarks PRP Science Driver PI Workshop UC Davis March 23, 2016 Dr. Larry Smarr Director, California.
High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC.
“OptIPuter: From the End User Lab to Global Digital Assets" Panel UC Research Cyberinfrastructure Meeting October 10, 2005 Dr. Larry Smarr.
“ Building an Information Infrastructure to Support Microbial Metagenomic Sciences " Presentation to the NBCR Research Advisory Committee UCSD La Jolla,
Campus Cyberinfrastructure
Optical SIG, SD Telecom Council
The OptIPortal, a Scalable Visualization, Storage, and Computing Termination Device for High Bandwidth Campus Bridging Presentation by Larry Smarr to.
Presentation transcript:

A High-Performance Campus-Scale Cyberinfrastructure: The Technical, Political, and Economic Presentation by Larry Smarr to the NSF Campus Bridging Workshop October 11, 2010 Anaheim, CA Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technology Harry E. Gruber Professor, Dept. of Computer Science and Engineering Jacobs School of Engineering, UCSD Follow me on Twitter: lsmarr

Academic Research “OptIPlatform” Cyberinfrastructure: An End-to-End 10Gbps Lightpath Cloud National LambdaRail Campus Optical Switch Data Repositories & Clusters HPC HD/4k Video Images HD/4k Video Cams End User OptIPortal 10G Lightpaths HD/4k Telepresence Instruments

“Blueprint for the Digital University”--Report of the UCSD Research Cyberinfrastructure Design Team Focus on Data-Intensive Cyberinfrastructure research.ucsd.edu/documents/rcidt/RCIDTReportFinal2009.pdf No Data Bottlenecks --Design for Gigabit/s Data Flows April 2009

Broad Campus Input to Build the Plan and Support for the Plan Campus Survey of CI Needs-April 2008 –45 Responses (Individuals, Groups, Centers, Depts) –#1 Need was Data Management –80% Data Backup –70% Store Large Quantities of Data –64% Long Term Data Preservation –50% Ability to Move and Share Data Vice Chancellor of Research Took the Lead Case Studies Developed from Leading Researchers Broad Research CI Design Team –Chaired by Mike Norman and Phil Papadopoulos –Faculty and Staff: –Engineering, Oceans, Physics, Bio, Chem, Medicine, Theatre –SDSC, Calit2, Libraries, Campus Computing and Telecom

Why Invest in Campus Research CI? Competitive Advantage Growing Campus Demand Leadership Opportunities Complementarity With National Programs Preservation of Digital Knowledge is Vital to the Scientific Method Institutional Obligations to Preserve Data: –OMB Circular A ‐ 110/CFR Part 215 –Preserve Data for 3 Years After Grant Escalating Energy/Space Demands Integration with UC ‐ Wide Initiatives

Why Invest Now? Doing More With Less Exploit UCSD’s Under ‐ Developed Synergies SDSC Deployment of the Triton Resource The Longer We Wait – The Harder It Will Get –The More Opportunities Will Be Lost

Implementing the Campus Research CI Plan Cyberinfrastructure Planning & Operations Committee –Chancellor Fox Appoints Fall 2009 –Mission: Develop a Business Plan for the Self-Sustaining Operations of a Research Cyberinfrastructure –Report Delivered April 2010 Business Plan Components –Direct Campus Investment –Energy Savings –PI Contributions –ICR –Separate Budgets for Startup and Sustaining Create an RCI Oversight Committee

UCSD Campus Investment in Fiber and Networks Enables High Performance Campus Bridging CI DataOasis (Central) Storage OptIPortal Tile Display Wall Campus Lab Cluster Digital Data Collections Triton – Petadata Analysis Gordon – HPC System Cluster Condo Scientific Instruments N x 10Gbe CENIC, NLR, I2DCN Source: Philip Papadopoulos, SDSC, UCSD

UCSD Planned Optical Networked Biomedical Researchers and Instruments Cellular & Molecular Medicine West National Center for Microscopy & Imaging Biomedical Research Center for Molecular Genetics Pharmaceutical Sciences Building Cellular & Molecular Medicine East CryoElectron Microscopy Facility Radiology Imaging Lab Bioengineering San Diego Supercomputer Center Connects at 10 Gbps : –Microarrays –Genome Sequencers –Mass Spectrometry –Light and Electron Microscopes –Whole Body Imagers –Computing –Storage

Triton Resource Large Memory PSDAF 256/512 GB/sys 9TB Total 128 GB/sec ~ 9 TF x28 Shared Resource Cluster 24 GB/Node 6TB Total 256 GB/sec ~ 20 TF x256 Campus Research Network UCSD Research Labs Large Scale Storage 2 PB 40 – 80 GB/sec 3000 – 6000 disks Phase 0: 1/3 TB, 8GB/s Moving to a Shared Enterprise Data Storage and Analysis Resource:Triton SDSC

Rapid Evolution of 10GbE Port Prices Makes Campus-Scale 10Gbps CI Affordable $80K/port Chiaro (60 Max) $ 5K Force 10 (40 max) $ 500 Arista 48 ports ~$1000 (300+ Max) $ 400 Arista 48 ports Port Pricing is Falling Density is Rising – Dramatically Cost of 10GbE Approaching Cluster HPC Interconnects Source: Philip Papadopoulos, SDSC, UCSD

10G Switched Data Analysis Resource: Data Oasis (RFP Underway) 2 32 OptIPuter 32 Colo RCN CalRen Existing Storage 1500 – 2000 TB > 40 GB/s Triton 8 Dash 100 Gordon Oasis Procurement (RFP) Minimum 40 GB/sec for Lustre Nodes must be able to function as Lustre OSS (Linux) or NFS (Solaris) Connectivity to Network is 2 x 10GbE/Node Likely Reserve dollars for inexpensive replica servers 40

High Performance Computing (HPC) vs. High Performance Data (HPD) AttributeHPCHPD Key HW metricPeak FLOPSPeak IOPS Architectural featuresMany small-memory multicore nodes Fewer large-memory vSMP nodes Typical applicationNumerical simulationDatabase query Data mining ConcurrencyHigh concurrencyLow concurrency or serial Data structuresData easily partitioned e.g. grid Data not easily partitioned e.g. graph Typical disk I/O patternsLarge block sequentialSmall block random Typical usage modeBatch processInteractive Source: SDSC

G RAND C HALLENGES IN D ATA -I NTENSIVE S CIENCES O CTOBER 26-28, 2010 S AN D IEGO S UPERCOMPUTER C ENTER, UC S AN D IEGO Confirmed conference topics and speakers : Needs and Opportunities in Observational Astronomy - Alex Szalay, JHU Transient Sky Surveys – Peter Nugent, LBNL Large Data-Intensive Graph Problems – John Gilbert, UCSB Algorithms for Massive Data Sets – Michael Mahoney, Stanford U. Needs and Opportunities in Seismic Modeling and Earthquake Preparedness - Tom Jordan, USC Needs and Opportunities in Fluid Dynamics Modeling and Flow Field Data Analysis – Parviz Moin, Stanford U. Needs and Emerging Opportunities in Neuroscience – Mark Ellisman, UCSD Data-Driven Science in the Globally Networked World – Larry Smarr, UCSD