CS267 - April 26th, 2011 Big Bang, Big Iron High Performance Computing and the Cosmic Microwave Background Julian Borrill Computational Cosmology Center,

Slides:



Advertisements
Similar presentations
Dark Matter, Dark Energy, and the Current State of Cosmology
Advertisements

Primordial perturbations and precision cosmology from the Cosmic Microwave Background Antony Lewis CITA, University of Toronto
Planck 2013 results, implications for cosmology
A Scientific History of the Universe. How do we predict the conditions of the early universe? What are the different eras in the early universe? What.
Chapter 17 The Beginning of Time
Foreground cleaning in CMB experiments Carlo Baccigalupi, SISSA, Trieste.
Systematic effects in cosmic microwave background polarization and power spectrum estimation SKA 2010 Postgraduate Bursary Conference, Stellenbosch Institute.
Universe in a box: simulating formation of cosmic structures Andrey Kravtsov Department of Astronomy & Astrophysics Center for Cosmological Physics (CfCP)
Cosmic Microwave Background Data Analysis : From Time-Ordered Data To Power Spectra Julian Borrill Computational Research Division, Berkeley Lab & Space.
Applications for K42 Initial Brainstorming Paul Hargrove and Kathy Yelick with input from Lenny Oliker, Parry Husbands and Mike Welcome.
CMB as a physics laboratory
CMB polarisation results from QUIET
© 2010 Pearson Education, Inc. Chapter 23 The Beginning of Time.
X What is the shape of our universe? By Sandro Berndt.
Cosmology I & II Expanding universe Hot early universe Nucleosynthesis Baryogenesis Cosmic microwave background (CMB) Structure formation Dark matter,
Evolution of the Universe (continued)
Cosmic Microwave Background (CMB) Peter Holrick and Roman Werpachowski.
P olarized R adiation I maging and S pectroscopy M ission Probing cosmic structures and radiation with the ultimate polarimetric spectro-imaging of the.
The Cosmic Simulator Daniel Kasen (UCB & LBNL) Peter Nugent, Rollin Thomas, Julian Borrill & Christina Siegerist.
Eric V. Linder (arXiv: v1). Contents I. Introduction II. Measuring time delay distances III. Optimizing Spectroscopic followup IV. Influence.
Hubble’s Law Our goals for learning What is Hubble’s Law?
Bill Reach 2009 May 14 Greater IPAC Technology Symposium.
CMB observations and results Dmitry Pogosyan University of Alberta Lake Louise, February, 2003 Lecture 1: What can Cosmic Microwave Background tell us.
Probing fundamental physics with CMB B-modes Cora Dvorkin IAS Harvard (Hubble fellow) Status and Future of Inflationary Theory workshop August 2014, KICP.
CS267-April 20th, 2010 Big Bang, Big Iron High Performance Computing and the Cosmic Microwave Background Julian Borrill Computational Cosmology Center,
CS267 – April 30th, 2013 Big Bang, Big Data, Big Iron High Performance Computing and the Cosmic Microwave Background Julian Borrill Computational Cosmology.
AS2001 / 2101 Chemical Evolution of the Universe Keith Horne Room 315A
Parallel I/O Performance: From Events to Ensembles Andrew Uselton National Energy Research Scientific Computing Center Lawrence Berkeley National Laboratory.
MAPping the Universe ►Introduction: the birth of a new cosmology ►The cosmic microwave background ►Measuring the CMB ►Results from WMAP ►The future of.
Cosmic Microwave Background Carlo Baccigalupi, SISSA CMB lectures at TRR33, see the complete program at darkuniverse.uni-hd.de/view/Main/WinterSchoolLecture5.
AS2001 Chemical Evolution of the Universe Keith Horne 315a
Cosmic Microwave Background Carlo Baccigalupi, SISSA CMB lectures at TRR33, see the complete program at darkuniverse.uni-hd.de/view/Main/WinterSchoolLecture5.
PHY306 1 Modern cosmology 3: The Growth of Structure Growth of structure in an expanding universe The Jeans length Dark matter Large scale structure simulations.
The measurement of q 0 If objects are observed at large distances of known brightness (standard candles), we can measure the amount of deceleration since.
Cosmic Microwave Background Carlo Baccigalupi, SISSA CMB lectures at TRR33, see the complete program at darkuniverse.uni-hd.de/view/Main/WinterSchoolLecture5.
EBEx foregrounds and band optimization Carlo Baccigalupi, Radek Stompor.
Racah Institute of physics, Hebrew University (Jerusalem, Israel)
The Planck Satellite Hannu Kurki-Suonio University of Helsinki Finnish-Japanese Workshop on Particle Cosmology, Helsinki
Testing the slow roll inflation paradigm with the Big Bang Observer
The Beginning of Time: Evidence for the Big Bang & the Theory of Inflation.
Adventures in Parameter Estimation Jason Dick University of California, Davis.
Astro-2: History of the Universe Lecture 10; May
Chapter 18: Chapter 18: Cosmology. WHAT DO YOU THINK? What does the universe encompass? Is the universe expanding, fixed in size, or contracting? Will.
Comprehensive Scientific Support Of Large Scale Parallel Computation David Skinner, NERSC.
US Planck Data Analysis Review 1 Julian BorrillUS Planck Data Analysis Review 9–10 May 2006 Computing Facilities & Capabilities Julian Borrill Computational.
The Cosmic Microwave Background
Cosmic Microwave Background Data Analysis At NERSC Julian Borrill with Christopher Cantalupo Theodore Kisner.
Cosmology -- the Origin and Structure of the Universe Cosmological Principle – the Universe appears the same from all directions. There is no preferred.
CMB, lensing, and non-Gaussianities
The Planck Satellite Matthew Trimble 10/1/12. Useful Physics Observing at a redshift = looking at light from a very distant object that was emitted a.
The Planck Mission: Looking into the Past to Learn about Our Future Courtney Nickle, Stephanie Clark and Taylor Phillips Astronomy, Spring 2011 Abstract.
BICEP2 Results & Its Implication on inflation models and Cosmology Seokcheon Lee 48 th Workshop on Gravitation & NR May. 16 th
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Constraints on primordial non-Gaussianity.
CMB physics Zong-Kuan Guo 《现代宇宙学》
Cosmic Microwave Background
Harrison B. Prosper Florida State University YSP
Testing Primordial non-Gaussianities in CMB Anisotropies
Julian Borrill Computational Cosmology Center, LBL
Chapter 30 Section 4 Big Bang Theory.
(for the Algorithm Development Group of the US Planck Team)
Data Issues Julian Borrill
12th Marcel Grossman Meeting,
The Cosmic Microwave Background and the WMAP satellite results
BIG BANG THEORY.
Physics Seminar Measurement of the Cosmic Microwave Background anisotropies and polarization with Planck Assoc. Prof. Guillaume Patanchon Astroparticle.
Cosmology.
Homework #10 is due tonight, 9:00 pm.
BIG BANG THEORY.
CMB Anisotropy 이준호 류주영 박시헌.
Separating E and B types of CMB polarization on an incomplete sky Wen Zhao Based on: WZ and D.Baskaran, Phys.Rev.D (2010) 2019/9/3.
Presentation transcript:

CS267 - April 26th, 2011 Big Bang, Big Iron High Performance Computing and the Cosmic Microwave Background Julian Borrill Computational Cosmology Center, LBL Space Sciences Laboratory, UCB and the BOOMERanG, MAXIMA, Planck, EBEX & PolarBear collaborations

CS267 - April 26th, 2011 The Cosmic Microwave Background About 400,000 years after the Big Bang, the expanding Universe cools through the ionization temperature of hydrogen: p + + e - => H. Without free electrons to scatter off, the photons free-stream to us today. COSMIC - filling all of space. MICROWAVE - redshifted by the expansion of the Universe from 3000K to 3K. BACKGROUND - primordial photons coming from “behind” all astrophysical sources.

CS267 - April 26th, 2011 CMB Science Primordial photons give the earliest possible image of the Universe. The existence of the CMB supports a Big Bang over a Steady State cosmology (NP1). Tiny fluctuations in the CMB temperature (NP2) and polarization encode the fundamentals of – Cosmology geometry, topology, composition, history, … – Highest energy physics grand unified theories, the dark sector, inflation, … Current goals: – definitive T measurement provides complementary constraints for all dark energy experiments. – detection of cosmological B-mode gives energy scale of inflation from primordial gravity waves. (NP3)

CS267 - April 26th, 2011 The Concordance Cosmology Supernova Cosmology Project (1998): Cosmic Dynamics (   -  m ) BOOMERanG & MAXIMA (2000): Cosmic Geometry (   +  m ) 70% Dark Energy + 25% Dark Matter + 5% Baryons 95% Ignorance What (and why) is the Dark Universe ?

CS267 - April 26th, 2011 Observing the CMB With very sensitive, very cold, detectors. Scanning all of the sky from space, or just some of it from the stratosphere or high dry ground.

CS267 - April 26th, 2011 Analysing The CMB

CS267 - April 26th, 2011 CMB Satellite Evolution Evolving science goals require (i) higher resolution & (ii) polarization sensitivity.

CS267 - April 26th, 2011 CMB Data Analysis In principle very simple – Assume Guassianity and maximize the likelihood 1.of maps given the observations and their noise statistics (analytic). 2.of power spectra given maps and their noise statistics (iterative). In practice very complex – Foregrounds, glitches, asymmetric beams, non-Gaussian noise, etc. – Algorithm & implementation scaling with evolution of CMB data-set size HPC architecture

CS267 - April 26th, 2011 The CMB Data Challenge Extracting fainter signals (polarization, high resolution) from the data requires: – larger data volumes to provide higher signal-to-noise. – more complex analyses to control fainter systematic effects. ExperimentStart DateObservationsPixels* COBE BOOMERanG WMAP Planck PolarBear QUIET-II CMBpol x increase in data volume over next 15 years – need linear analysis algorithms to scale through next 10 M-foldings !

CS267 - April 26th, 2011 CMB Data Analysis Evolution Data volume & computational capability dictate analysis approach. DateDataSystemMapPower Spectrum B98 Cray T3E x 700 Explicit Maximum Likelihood (Matrix Invert - N p 3 ) Explicit Maximum Likelihood (Matrix Cholesky + Tri-solve - N p 3 ) B2K2 IBM SP3 x 3,000 Explicit Maximum Likelihood (Matrix Invert - N p 3 ) Explicit Maximum Likelihood (Matrix Invert + Multiply - N p 3 ) Planck SF IBM SP3 x 6,000 PCG Maximum Likelihood (band-limited FFT – few N t ) Monte Carlo (Sim + Map - many N t ) Planck AF EBEX Cray XT4 x 40,000 PCG Maximum Likelihood (band-limited FFT – few N t ) Monte Carlo (SimMap - many N t ) Planck MC PolarBear Cray XE6 x 150,000 PCG Maximum Likelihood (band-limited FFT – few N t ) Monte Carlo (Hybrid SimMap - many N t )

CS267 - April 26th, 2011 Scaling In Practice 2000: BOOMERanG T-map – 10 8 samples => 10 5 pixels – 128 Cray T3E processors; 2005: Planck T-map – samples => 10 8 pixels – 6000 IBM SP3 processors; 2008: EBEX T/P-maps – samples, 10 6 pixels – Cray XT4 cores. 2010: Planck Monte Carlo 1000 noise T-maps – samples => pixels – Cray XT4 cores.

CS267 - April 26th, 2011 Planck Sim/Map Target For Planck to publish its results in time, by mid-2012 we need to be able to simulate and map – O(10 4 ) realizations of the entire mission 74 detectors x 2.5 years ~ O(10 16 ) samples – On O(10 5 ) cores – In O(10) wall-clock hours WAIT ~ 1 day : COST ~ 10 6 CPU-hrs

CS267 - April 26th, 2011 TARGET: 10 4 maps 9 freqs 2.5 years 10 5 cores 10 hours 12x217 FFP1 M3/ GCP CTP3 OTFS Hybrid/ Peta-Scaling Hybrid/ Peta-Scaling

CS267 - April 26th, 2011 Simulation & Mapping: Calculations Given the instrument noise statistics & beams, a scanning strategy, and a sky: 1)SIMULATION: d t = n t + s t = n t + P tp s p – A realization of the piecewise stationary noise time-stream: Pseudo-random number generation & FFT – A signal time-stream scanned & beam-smoothed from the sky map: SHT 2)MAPPING: (P T N -1 P) d p = P T N -1 d t (A x = b) – Build the RHS FFT & sparse matrix-vector multiply – Solve for the map PCG over FFT & sparse matrix-vector multiply

CS267 - April 26th, 2011 Simulation & Mapping: Scaling In theory such analyses should scale – Linearly with the number of observations. – Perfectly to arbitrary numbers of cores. In practice this does not happen because of – IO (reading pointing; writing time-streams reading pointing & timestreams; writing maps) – Communication (gathering maps from all processes) – Calculation inefficiency (linear operations only) Code development has been an ongoing history of addressing these challenges anew with each new data volume and system concurrency.

CS267 - April 26th, 2011 IO - Before For each MC realization For each detector Read detector pointingSim Write detector timestream For all detectors Read detector timestream & pointingMap Write map  Read: 56 x Realizations x Detectors x Observations bytes Write: 8 x Realizations x (Detectors x Observations + Pixels) bytes E.g. for Planck, read 500PB & write 70PB.

CS267 - April 26th, 2011 IO - Optimizations Read sparse telescope pointing instead of dense detector pointing – Calculate individual detector pointing on the fly. Remove redundant write/read of time-streams between simulation & mapping – Generate simulations on the fly only when map-maker requests data. Put MC loop inside map-maker – Amortize common data reads over all realizations.

CS267 - April 26th, 2011 IO – After Now Read telescope pointing For each detector Calculate detector pointing For each MC realizationSimMap For all detectors Simulate time-stream Write map  Read: 24 x Sparse Observations bytes Write: 8 x Realizations x Pixels bytes E.g. for Planck, read 2GB & write 70TB (10 8 read & 10 3 write compression).

CS267 - April 26th, 2011 Communication Details The time-ordered data from all the detectors are distributed over the processes subject to: – Load-balance – Common telescope pointing Each process therefore holds – some of the observations – for some of the pixels. In each PCG iteration, each process solves with its observations. At the end of each iteration, each process needs to gather the total result for all of the pixels in its subset of the observations.

CS267 - April 26th, 2011 Communication - Before Initialize a process & MPI task on every core Distribute time-stream data & hence pixels After each PCG iteration – Each process creates a full map vector by zero-padding – Call MPI_Allreduce(map, world) – Each process extracts the pixels of interest to it & discards the rest

CS267 - April 26th, 2011 Communication – Optimizations Reduce the number of MPI tasks – Use threads for on-node communication – Only use MPI for off-node communication Minimize the total volume of the messages – Determine processes’ pair-wise pixel overlap – If the data volume is smaller, use gathers in place of reduces

CS267 - April 26th, 2011 Communication – After Now Initialize a process & MPI task on every node Distribute time-stream data & hence pixels Calculate common pixels for every pair of processes After each PCG iteration – If most pixels are common to most processes use MPI_Allreduce(map, world) as before – Else Each process prepares its send buffer Call MPI_Alltoallv(sbuffer, rbuffer, world) Each process receives contributions only to the pixels of interest to it & sums them.

CS267 - April 26th, 2011 Communication - Impact Fewer communicators & smaller message volume:

CS267 - April 26th, 2011 HPC System Evaluation Well-characterized & -instrumented science application codes can be a powerful tool for whole-system performance evaluation. Compare – unthreaded/threaded – allreduce/allgather on Cray XT4, XT5, XE6 on 200 – cores

CS267 - April 26th, 2011 Current Status Calculation scale with #observations. IO & communication scale with #pixels. Observations/pixel ~ S/N: science goals will help scaling! – Planck: O(10 3 ) observations per pixel – PolarBear: O(10 6 ) observations per pixel For each experiment, fixed data volume => strong scaling. Between experiments, growing data volume => weak scaling.

CS267 - April 26th, 2011 HPC System Evolution Clock speed is no longer able to maintain Moore’s Law. Multi-core CPU and GPGPU are two major approaches. Both of these will require – significant code development – performance experiments & auto-tuning E.g. NERSC’s new XE6 system Hopper – 6384 nodes – 2 sockets per node – 2 NUMA nodes per socket – 6 cores per NUMA node What is the best way to run hybrid code on such a system?

CS267 - April 26th, 2011

Conclusions The CMB provides a unique window onto the early Universe – investigate fundamental cosmology & physics. The CMB data sets we gather and the HPC systems we analyze them on are both evolving. CMB data analysis is a long-term computationally-challenging problem requiring state-of-the-art HPC capabilities. The science we can extract from present and future CMB data sets will be determined by the limits on a)our computational capability, and b)our ability to exploit it.