Cosmic-Scale Applications for Cyberinfrastructure NSF MPS Cyberscience Workshop NSF Headquarters Arlington, VA April 21, 2004 Dr. Larry Smarr Director,

Slides:



Advertisements
Similar presentations
The Australian Virtual Observatory e-Science Meeting School of Physics, March 2003 David Barnes.
Advertisements

Ed Seidel Assistant Director Directorate for Mathematical and Physical Sciences National Science Foundation October 1, 2010.
Background Chronopolis Goals Data Grid supporting a Long-term Preservation Service Data Migration Data Migration to next generation technologies Trust.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CF21) IRNC Kick-Off Workshop July 13,
SDSC Computing the 21st Century Talk Given to the NSF Sugar Panel May 27, 1998.
The Open Science Grid: Bringing the power of the Grid to scientific research
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
Universe in a box: simulating formation of cosmic structures Andrey Kravtsov Department of Astronomy & Astrophysics Center for Cosmological Physics (CfCP)
2. Point Cloud x, y, z, … Complete LiDAR Workflow 1. Survey 4. Analyze / “Do Science” 3. Interpolate / Grid USGS Coastal & Marine.
Quasar & Black Hole Science for GSMT Central question: Why do quasars evolve?
1 Observing the Most Violent Events in the Universe Virgo Barry Barish Director, LIGO Virgo Inauguration 23-July-03 Cascina 2003.
GridSphere for GridLab A Grid Application Server Development Framework By Michael Paul Russell Dept Computer Science University.
Astro-DISC: Astronomy and cosmology applications of distributed super computing.
X-ray Optical microwave Cosmology at KIPAC. The Survey 5000 square degrees (overlap with SPT and VISTA) Five-band (grizY) + VISTA (JHK) photometry to.
Science, Data, & Art in the Imax Film “Hubble 3D” Frank Summers, Greg Bacon, Lisa Frattare, Zolt Levay Space Telescope Science Institute January 21, 2010.
UNIVERSITY of MARYLAND GLOBAL LAND COVER FACILITY High Performance Computing in Support of Geospatial Information Discovery and Mining Joseph JaJa Institute.
Computing in Atmospheric Sciences Workshop: 2003 Challenges of Cyberinfrastructure Alan Blatecky Executive Director San Diego Supercomputer Center.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
Why Build Image Mosaics for Wide Area Surveys? An All-Sky 2MASS Mosaic Constructed on the TeraGrid A. C. Laity, G. B. Berriman, J. C. Good (IPAC, Caltech);
The NASA/NExScI/IPAC Star and Exoplanet Database 14 May 2009 David R. Ciardi on behalf of the NStED Team.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
The Cosmic Simulator Daniel Kasen (UCB & LBNL) Peter Nugent, Rollin Thomas, Julian Borrill & Christina Siegerist.
Amit Chourasia Visualization Scientist Visualization Services Presented at : Florida State University, Nov 20 th 2006 Scientific Visualization of Large.
1 Galaxies at Cosmic Dawn Revealed in the First Year of the Hubble Frontier Fields Initiative Dr. Gabriel Brammer (ESA/AURA, STScI) Hubble Science Briefing.
National Center for Supercomputing Applications The Computational Chemistry Grid: Production Cyberinfrastructure for Computational Chemistry PI: John Connolly.
National Center for Supercomputing Applications Observational Astronomy NCSA projects radio astronomy: CARMA & SKA optical astronomy: DES & LSST access:
Physics Steven Gottlieb, NCSA/Indiana University Lattice QCD: focus on one area I understand well. A central aim of calculations using lattice QCD is to.
Changing Science and Engineering: the impact of HPC Sept 23, 2009 Edward Seidel Assistant Director, Mathematical and Physical Sciences, NSF (Director,
Einstein’s elusive waves
Applications for the Grid Here at GGF1: Gabrielle Allen, Thomas, Dramlitsch, Gerd Lanfermann, Thomas Radke, Ed Seidel Max Planck Institute for Gravitational.
Jennifer Lotz Hubble Science Briefing Jan. 16, 2014 Exploring the Depths of the Universe.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Where to find LiDAR: Online Data Resources.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
Research Networks and Astronomy Richard Schilizzi Joint Institute for VLBI in Europe
Geosciences - Observations (Bob Wilhelmson) The geosciences in NSF’s world consists of atmospheric science, ocean science, and earth science Many of the.
THEORETICAL ASTROPHYSICS AND THE US-NVO INITIATIVE D. S. De Young National Optical Astronomy Observatory.
Renaissance: Formation of the first light sources in the Universe after the Dark Ages Justin Vandenbroucke, UC Berkeley Physics 290H, February 12, 2008.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
1. G O D D A R D S P A C E F L I G H T C E N T E R 2 New Budget Initiatives for NASA in FY04.
1 Building Bridges: CGWA Inauguration 15 December 2003 Lazarus Approach to Binary Black Hole Modeling John Baker Laboratory for High Energy Astrophysics.
Middleware Camp NMI (NSF Middleware Initiative) Program Director Alan Blatecky Advanced Networking Infrastructure and Research.
Possibility of detecting CHRISTODOULOU MEMORY of GRAVITATIONAL WAVES by using LISA (Laser Interferometer Space Antenna) Thus, the final form of the memory.
Breakout # 1 – Data Collecting and Making It Available Data definition “ Any information that [environmental] researchers need to accomplish their tasks”
Center for Computational Visualization University of Texas, Austin Visualization and Graphics Research Group University of California, Davis Molecular.
A GRID solution for Gravitational Waves Signal Analysis from Coalescing Binaries: preliminary algorithms and tests F. Acernese 1,2, F. Barone 2,3, R. De.
The OptIPuter Project Tom DeFanti, Jason Leigh, Maxine Brown, Tom Moher, Oliver Yu, Bob Grossman, Luc Renambot Electronic Visualization Laboratory, Department.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
“The UCSD Big Data Freeway System” Invited Short Talk Workshop on “Enriching Human Life and Society” UC San Diego February 6, 2014 Dr. Larry Smarr Director,
Energy Balance in Clusters of Galaxies Patrick M. Motl & Jack O. Burns Center for Astrophysics and Space Astronomy University of Colorado at Boulder X-ray.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Albert-Einstein-Institut Exploring Distributed Computing Techniques with Ccactus and Globus Solving Einstein’s Equations, Black.
FIRST LIGHT A selection of future facilities relevant to the formation and evolution of galaxies Wavelength Sensitivity Spatial resolution.
LIGO-G M Press Conference Scientific Operation of LIGO Gary H Sanders Caltech (on behalf of a large team) APS April Meeting Philadelphia 6-April-03.
The Submillimeter Array 1 David J. Wilner
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
CMB & LSS Virtual Research Community Marcos López-Caniego Enrique Martínez Isabel Campos Jesús Marco Instituto de Física de Cantabria (CSIC-UC) EGI Community.
High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC.
“OptIPuter: From the End User Lab to Global Digital Assets" Panel UC Research Cyberinfrastructure Meeting October 10, 2005 Dr. Larry Smarr.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
LISA Laser Interferometer Space Antenna: The Mission Mike Cruise For the LISA Team.
The Quest for Gravitational Waves: a global strategy
GW Policy: The Future: G3 Detectors
Tools and Services Workshop
Joslynn Lee – Data Science Educator
The Cactus Team Albert Einstein Institute
GW150914: The first direct detection of gravitational waves
TeraScale Supernova Initiative
Presentation transcript:

Cosmic-Scale Applications for Cyberinfrastructure NSF MPS Cyberscience Workshop NSF Headquarters Arlington, VA April 21, 2004 Dr. Larry Smarr Director, California Institute for Telecommunications and Information Technologies Harry E. Gruber Professor, Dept. of Computer Science and Engineering Jacobs School of Engineering, UCSD

Cosmic-Scale Science: Cyberinfrastructure Links Theory with Observation Two Examples –Formation of Structures in the Early Universe –Black Hole Collisions and Gravitational Radiation Common Features Emerge –$ Billions of New Instruments Generating Data –Much More Powerful Supercomputers Needed –Sophisticated Software Key –eg, Automatic Mesh Refinement Cyberinfrastructure Required for Data Produced –Federated Repositories –Data Grid Middleware –Local Laboratory Standards-Based Clusters

NASA, ESA, S. Beckwith (STScI) and the HUDF Team Hubble Ultra Deep Field Fundamental Physics Challenge: Formation of First Galaxies and Clusters: Faintest galaxies ~ 1 billion years old Galaxy population is strongly evolving 380,000 yr NASA WMAP Source: Mike Norman, UCSD

Formation & Evolution of Galaxies: $Billions of New Digital Observatories Nature and Occurrence of the First Galaxies “First Light” (JWST, ALMA) Properties of High-Z Galaxies (HST, ALMA)  Galaxy Building Blocks? Source(s) of Early Reionization (WMAP) Star Formation History of Galaxies (Spitzer) Emergence of the Hubble Types (DEEP2) Influence of Environment on Galaxy Type and Large Scale Structure (SDSS) Supermassive Black Hole Formation and AGN/QSO Phenomena In Galaxies (SDSS, HST, CXO) Many Open Questions Are Being Investigated Observationally Source: Mike Norman, UCSD

Cosmic Simulator with a Billion Zone and Gigaparticle Resolution Compare with Sloan Survey Source: Mike Norman, UCSD SDSC Blue Horizon

Why Does the Cosmic Simulator Need Cyberinfrastructure? One Gigazone Run: –Generates ~10 TeraByte of Output –A “Snapshot” is 100 GB –Need to Visually Analyze as We Create SpaceTimes Visual Analysis Daunting –Single Frame is About 8GB –A Smooth Animation of 1000 Frames is 1000 x 8 GB=8TB –Stage on Rotating Storage to High Res Displays Can Run Evolutions Faster than We can Archive Them –File Transport Over Shared Internet ~50 Mbit/s –4 Hours to Move ONE Snapshot! –Many Scientists Will Need Access for Analysis Source: Mike Norman, UCSD

Limitations of Uniform Grids for Complex Scientific and Engineering Problems Source: Greg Bryan, Mike Norman, NCSA 512x512x512 Run on 512-node CM-5 Gravitation Causes Continuous Increase in Density Until There is a Large Mass in a Single Grid Zone

Solution: Develop Automatic Mesh Refinement (AMR) to Resolve Mass Concentrations Source: Greg Bryan, Mike Norman, John Shalf, NCSA 64x64x64 Run with Seven Levels of Adaption on SGI Power Challenge, Locally Equivalent to 8192x8192x8192 Resolution

Background Image Shows Grid Hierarchy Used –Key to Resolving Physics is More Sophisticated Software –Evolution is from 10Myr to Present Epoch Every Galaxy > M solar in 100 Mpc/H Volume Adaptively Refined With AMR –256 3 Base Grid –Over 32,000 Grids At 7 Levels Of Refinement –Spatial Resolution of 4 kpc at Finest –150,000 CPU-hr On NCSA Origin2000 –Completed In AMR or Unigrid Now Feasible –8-64 Times The Mass Resolution –Can Simulate First Galaxies AMR Allows Digital Exploration of Early Galaxy and Cluster Core Formation Source: Mike Norman, UCSD

Hydrodynamic Cosmology Simulation of Galaxy Formation Using Parallel Adaptive Mesh Refinement (Enzo) Image credit: Donna Cox, Bob Patterson (NCSA) Simulation: M. Norman (UCSD)

Cosmic Simulator: Thresholds of Capability and Discovery 2000: Formation of Galaxy Cluster Cores (1 TFLOP/s) 2006: Properties of First Galaxies (40 TFLOP/s) 2010: Emergence of Hubble Types (150 TFLOP/s) 2014: Large Scale Distribution Of Galaxies By Luminosity And Morphology (500 TFLOP/s) Hubble types LSS Source: Mike Norman, UCSD

Proposed Galaxy Simulation Cyber-Grid User Grid Modelers Observers Visualizers Developer Grid Enzo Code Data Mgmt Analysis Tools Visualization Middleware Observational Survey Partners SDSS DEEP2 SWIRE Outreach Tutorials Animations PBS Nova Production Simulated Galaxy Grid Enzo Data Grid Enzo Simulation Code Enzo Data Analysis Tools Portal Interface Simulated Galaxy Archive NSF NMI, PI: M. Norman, UCSD

LIGO, VIRGO, GEO and LISA Search for Gravitational Waves $1B Being Spent On Ground-Based LIGO/VIRGO/GEO and Space-Based LISA –Use Laser Interferometers To Detect Waves Matched Filtering of Waveforms Requires Large Numbers of Simulations –Stored In Federated Repositories LISA’s Increased Sensitivity Vastly Opens Parameter Space: –Many Orders Of Magnitude More Parameter Space to be Searched! LIGO-Hanford Virgo-Pisa Source: Ed Seidel, LSU

Two Body Problem in General Relativity - The Collision of Two Black Holes Numerical Solution of Einstein Equations Required Problem Solution Started 40 Years Ago, 10 More to Go Wave Forms Critical for NSF LIGO Gravitational Wave Detector A PetaFLOPS-Class Grand Challenge Oct. 10, 1995 Matzner, Seidel, Shapiro, Smarr, Suen, Teukolsky, Winicuor

MegaflopGigaflopTeraflopKiloflop Lichnerowicz The Numerical Two Black Hole Problem Spans the Digital Computer Era Hahn & Lindquist DeWitt/Misner -Chapel Hill DeWitt-LLNL Cadez Thesis Eppley Thesis Smarr Thesis Modern Era

Relative Amount of Floating Point Operations for Three Epochs of the 2BH Collision Problem 9,000,000X 30,000X 1999 Seidel & Suen, et al. SGI Origin 256 Processors Each 500 Mflops 40 Hours 1977 Eppley & Smarr CDC 7600 One Processor Each 35 Mflops 5 Hours 300X 1963 Hahn & Lindquist IBM 7090 One Processor Each 0.2 Mflops 3 Hours 10,000x More Required!

What is Needed to Finish the Computing Job Current Black Hole Jobs –Grid: 768 X 768 X 384 Memory Used: 250+ GB –Runtime:~ Day Or More Output: Multi-TB+ (Disk Limited) Inspiraling BH Simulations Are Volume Limited –Scale As N 3-4 Low-Resolution Simulations of BH Collisions: –Currently Require O(10 15 ) FLOPS: High-Resolution Inspiraling Binaries Need: –Increased Simulation Volume, Evolution Time, And Resolution - And O( ) Flops –50-100TF With Adaptive Meshes Will Make This Possible Source: Ed Seidel, LSU

Why Black Hole Simulations Need Cyberinfrastructure Software Development is Key –Use Adaptive Meshes to Accurately Resolve Metric –~10 Levels Of Refinement, –Several Machine-Days Per Spacetime Output –Minimal TB For Full Analysis (Multiple Orbits) of: –Gravitational Waves –Event Horizon Structure Evolution Real-Time Scheduling Needed Across Multiple Resources For Collaborative Distributed Computing –Spawning (For Analysis, Steering Tasks), Migration –Interactive Viz From Distributed Collaborations –Implies Need for Dedicated Gigabit Light Pipes (Lambdas) Source: Ed Seidel, LSU

Ensembles Of Simulations Needed for LIGO, GEO, LISA Gravitational Wave Astronomy Variations for Internal Approximations –Accuracy, Sensitivity Analysis To Gauge Parameters, Resolution, Algorithms –Dozen Simulations Per Physical Scenario Variations In Physical Scenarios-->Waveform Catalogs –Masses, Spins, Orbital Characteristics Varied –Huge Parameter Space To Survey In Total: Simulations Needed –Potentially Generating 25TB Each –Stored In Federated Repositories Data Analysis Of LIGO, GEO, LISA Signals –Interacting With Simulation Data –Managing Parameter Space/Signal Analysis Source: Ed Seidel, LSU

To a Grid “Supercomputers” are Just High Performance Data Generators Similar to Particle Accelerators, Telescopes, Ocean Observatories, Microscopes, etc. All Require: –Web Portal Access for Real-Time Instrument Control –Grid Middleware for Security, Scheduling, Reservations –Federated Repositories for Data Archiving –Data Grids for Data Replication and Management –High Performance Networking to Deal With Data Floods –Local Visualization and Analysis Facilities –Multi-Site Multi-Modal Collaboration Software That is—a Cyberinfrastructure!

NSF Must Increase Funding for Community Software/Toolkit Development Major Problem To Enable Community –Modern Software Engineering –Training –User Support Require Toolkits For: –Sharing/Developing Of Community Codes –Algorithmic Libraries, e.g. AMR –Local Compute, Storage, Visualization, & Analysis –Federated Repositories –Grid Middleware –Lambda Provisioning

LambdaGrid Required to Support the Distributed Collaborative Teams Grand Challenge-Like Teams Involving US and International Collaborations –Example: GWEN (Gravitational Wave European Network) Involves 20 Groups! Simulation Data Stored Across Geographically Distributed Spaces –Organization, Access, Mining Issues Collaborative Data Spaces to Support Interaction with: –Colleagues, Data, Simulations Need Lambda Provisioning For: –Coupling Supercomputers and Data Grid –Remote Visualization And Monitoring Of Simulations –Analysis Of Federated Data Sets By Virtual Organizations Source: Ed Seidel, LSU

Special Thanks to: Ed Seidel –Director, Center for Computation and Technology, –Department of Physics and Astronomy, –Louisiana State University –& Albert-Einstein-Institut –Potsdam, Germany –Representing dozens of scientists Michael Norman –Director, Laboratory for Computational Astrophysics –Physics Department, –UC San Diego Members of the OptIPuter Team