The Statistical Properties of Large Scale Structure Alexander Szalay Department of Physics and Astronomy The Johns Hopkins University.

Slides:



Advertisements
Similar presentations
Independent Component Analysis
Advertisements

Noise & Data Reduction. Paired Sample t Test Data Transformation - Overview From Covariance Matrix to PCA and Dimension Reduction Fourier Analysis - Spectrum.
November 12, 2013Computer Vision Lecture 12: Texture 1Signature Another popular method of representing shape is called the signature. In order to compute.
Exploiting Sparse Markov and Covariance Structure in Multiresolution Models Presenter: Zhe Chen ECE / CMR Tennessee Technological University October 22,
ECE 472/572 - Digital Image Processing Lecture 8 - Image Restoration – Linear, Position-Invariant Degradations 10/10/11.
An introduction to Principal Component Analysis (PCA)
Dimensional reduction, PCA
lecture 5, Sampling and the Nyquist Condition Sampling Outline  FT of comb function  Sampling Nyquist Condition  sinc interpolation Truncation.
Prénom Nom Document Analysis: Data Analysis and Clustering Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Data Basics. Data Matrix Many datasets can be represented as a data matrix. Rows corresponding to entities Columns represents attributes. N: size of the.
Independent Component Analysis (ICA) and Factor Analysis (FA)
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Visual Recognition Tutorial1 Random variables, distributions, and probability density functions Discrete Random Variables Continuous Random Variables.
Modeling the 3-point correlation function Felipe Marin Department of Astronomy & Astrophysics University of Chicago arXiv: Felipe Marin Department.
ICA Alphan Altinok. Outline  PCA  ICA  Foundation  Ambiguities  Algorithms  Examples  Papers.
Laurent Itti: CS599 – Computational Architectures in Biological Vision, USC Lecture 7: Coding and Representation 1 Computational Architectures in.
Introduction to Power Spectrum Estimation Lloyd Knox (UC Davis) CCAPP, 23 June 2010.
(1) A probability model respecting those covariance observations: Gaussian Maximum entropy probability distribution for a given covariance observation.
Modern Navigation Thomas Herring
Survey on ICA Technical Report, Aapo Hyvärinen, 1999.
Review of Probability.
Linear Algebra and Image Processing
EE513 Audio Signals and Systems Statistical Pattern Classification Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Weak Lensing 3 Tom Kitching. Introduction Scope of the lecture Power Spectra of weak lensing Statistics.
G. Cowan Lectures on Statistical Data Analysis Lecture 3 page 1 Lecture 3 1 Probability (90 min.) Definition, Bayes’ theorem, probability densities and.
Random Sampling, Point Estimation and Maximum Likelihood.
Mathematical Preliminaries. 37 Matrix Theory Vectors nth element of vector u : u(n) Matrix mth row and nth column of A : a(m,n) column vector.
7-1 Introduction The field of statistical inference consists of those methods used to make decisions or to draw conclusions about a population. These.
Binary Stochastic Fields: Theory and Application to Modeling of Two-Phase Random Media Steve Koutsourelakis University of Innsbruck George Deodatis Columbia.
Galaxy clustering II 2-point correlation function 5 Feb 2013.
Basics of Neural Networks Neural Network Topologies.
The Theory/Observation connection lecture 2 perturbations Will Percival The University of Portsmouth.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Course 2 Image Filtering. Image filtering is often required prior any other vision processes to remove image noise, overcome image corruption and change.
NASSP Masters 5003F - Computational Astronomy Lecture 14 Reprise: dirty beam, dirty image. Sensitivity Wide-band imaging Weighting –Uniform vs Natural.
CY3A2 System identification1 Maximum Likelihood Estimation: Maximum Likelihood is an ancient concept in estimation theory. Suppose that e is a discrete.
23 Sep The Feasibility of Constraining Dark Energy Using LAMOST Redshift Survey L.Sun Peking Univ./ CPPM.
Log Likelihood Estimate the log likelihood in the KL basis, by rotating into the diagonal eigensystem, and rescaling with the square root of the eigenvalues.
ENEE631 Digital Image Processing (Spring'04) Basics on 2-D Random Signal Spring ’04 Instructor: Min Wu ECE Department, Univ. of Maryland, College Park.
6dF Workshop April Sydney Cosmological Parameters from 6dF and 2MRS Anaïs Rassat (University College London) 6dF workshop, AAO/Sydney,
The Feasibility of Constraining Dark Energy Using LAMOST Redshift Survey L.Sun.
3rd International Workshop on Dark Matter, Dark Energy and Matter-Antimatter Asymmetry NTHU & NTU, Dec 27—31, 2012 Likelihood of the Matter Power Spectrum.
Elements of Pattern Recognition CNS/EE Lecture 5 M. Weber P. Perona.
Discrete-time Random Signals
Dark Energy and baryon oscillations Domenico Sapone Université de Genève, Département de Physique théorique In collaboration with: Luca Amendola (INAF,
Stochastic Process Theory and Spectral Estimation Bijan Pesaran Center for Neural Science New York University.
 Introduction to reciprocal space
Stochastic Process Theory and Spectral Estimation
Spatial Point Processes Eric Feigelson Institut d’Astrophysique April 2014.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Feasibility of detecting dark energy using bispectrum Yipeng Jing Shanghai Astronomical Observatory Hong Guo and YPJ, in preparation.
New Approaches to Modeling Nonlinear Structure Formation Nuala McCullagh Johns Hopkins University Cosmology on the Beach Cabo San Lucas, Mexico January.
Carlos Hernández-Monteagudo CE F CA 1 CENTRO DE ESTUDIOS DE FÍSICA DEL COSMOS DE ARAGÓN (CE F CA) J-PAS 10th Collaboration Meeting March 11th 2015 Cosmology.
Chapter 13 Discrete Image Transforms
Geology 6600/7600 Signal Analysis 04 Sep 2014 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
Power Spectrum Estimation in Theory and in Practice Adrian Liu, MIT.
Environmental Data Analysis with MatLab 2 nd Edition Lecture 22: Linear Approximations and Non Linear Least Squares.
Inh Jee University of Texas at Austin Eiichiro Komatsu & Karl Gebhardt
Part 3: Estimation of Parameters. Estimation of Parameters Most of the time, we have random samples but not the densities given. If the parametric form.
CS Statistical Machine learning Lecture 7 Yuan (Alan) Qi Purdue CS Sept Acknowledgement: Sargur Srihari’s slides.
Biointelligence Laboratory, Seoul National University
Probability Theory and Parameter Estimation I
Background on Classification
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
The Chinese University of Hong Kong
Outline Linear Shift-invariant system Linear filters
Outline Linear Shift-invariant system Linear filters
EE513 Audio Signals and Systems
Feature space tansformation methods
Probabilistic Surrogate Models
Presentation transcript:

The Statistical Properties of Large Scale Structure Alexander Szalay Department of Physics and Astronomy The Johns Hopkins University

Outline Lecture 1: –Basic cosmological background –Growth of fluctuations –Parameters and observables Lecture 2: –Statistical concepts and definitions –Practical approaches –Statistical estimators Lecture 3: –Applications to the SDSS –Angular correlations with photometric redshifts –Real-space power spectrum

Lecture #2 Statistical concepts and definitions –Correlation function –Power spectrum –Smoothing kernels –Window functions Statistical estimators

Basic Statistical Tools Correlation functions –N-point and N th -order –Defined in real space –Easy to compute, direct physical meaning –Easy to generalize to higher order Power spectrum –Fourier space equivalent of correlation functions –Directly related to linear theory –Origins in the Big Bang –Connects the CMB physics to redshift surveys Most common are the 2 nd order functions: –Variance  8 2 –Two-point correlation function  (r) –Power spectrum P(k)

The Galaxy Correlation Function First measured by Totsuji and Kihara, then Peebles etal Mostly angular correlations in the beginning Later more and more redshift space Power law is a good approximation Correlation length r 0 =5.4 h -1 Mpc Exponent is around  =1.8 Corresponding angular correlations

The Overdensity We can observe galaxy counts n(x), and compare to the expected counts Overdensity Fourier transform Inverse Wave number

The Power Spectrum Consider the ensemble average Change the origin by R Translational invariance Power spectrum Rotational invariance

Correlation Function Defined through the ensemble average Can be expressed through the Fourier transform Using the translational invariance in Fourier space The correlation function only depends on the distance

P(k) vs  (r) The Rayleigh expansion of a plane wave gives Using the rotational invariance of P(k) The power per logarithmic interval The power spectrum and correlation function form a Fourier transform pair

Filtering the Density Effect of a smoothing kernel K(x), where Convolution theorem Filtered power spectrum Filtered correlation function

Variance At r=0 separation we get the variance: Usual kernel is a ‘top-hat’ with an R=8h -1 Mpc radius The usual normalization of the power spectrum is using this window

Selection Window We always have an anisotropic selection window, both over the sky and along the redshift direction The observed overdensity is Using the convolution theorem

The Effect of Window Shape The lines in the spectrum are at least as broad as the window – the PSF of measuring the power spectrum! The shape of the window in Fourier space is the conjugate to the shape in real space The larger the survey volume, the sharper the k-space window => survey design

1-Point Probability Distribution Overdensity is a superposition from Fourier space Each  depends on a large number of modes Variance (usually filtered at some scale R) Central limit theorem: Gaussian distribution

N-point Probability Distribution Many ‘soft’ pixels, smoothed with a kernel Raw dataset x Parameter vector  Joint probability distribution is a multivariate Gaussian M is the mean, C is the correlation matrix C depends on the parameters 

Fisher Information Matrix Measures the sensitivity of the probability distribution to the parameters Kramer-Rao theorem: one cannot measure a parameter more accurately than

Higher Order Correlations One can define higher order correlation functions Irreducible correlations represented by connected graphs For Gaussian fields only 2-point, all other =0 Peebles conjecture: only tree graphs are present Hierarchical expansion Important on small scales

Correlation Estimators Expectation value Rewritten with the density as Often also written as Probability of finding objects in excess of random The two estimators above are NOT EQUIVALENT

Edge Effects The objects close to the edge are different The estimator has an excess variance (Ripley) If one is using the first estimator, these cancel in first order (Landy and Szalay 1996)

Discrete Counts We can measure discrete galaxy counts The expected density is a known, fractional The overdensity is If we define cells with counts N i

Power Spectrum Naïve estimator for a discrete density field is FKP (Feldman, Kaiser and Peacock) estimator The Fourier space equivalent to LS

Wish List Almost lossless for the parameters of interest Easy to compute, and test hypotheses (uncorrelated errors) Be computationally feasible Be able to include systematic effects

The Karhunen-Loeve Transform Subdivide survey volume into thousands of cells Compute correlation matrix of galaxy counts among cells from fiducial P(k)+ noise model Diagonalize matrix –Eigenvalues –Eigenmodes Expand data over KL basis Iterate over parameter values: –Compute new correlation matrix –Invert, then compute log likelihood Vogeley and Szalay (1996)

Eigenmodes

Optimal weighting of cells to extract signal represented by the mode Eigenvalues measure S/N Eigenmodes orthogonal In k-space their shape is close to window function Orthogonality = repulsion Dense packing of k-space => filling a ‘Fermi sphere’

Truncated expansion Use less than all the modes: truncation Best representation in the rms sense Optimal subspace filtering, throw away modes which contain only noise

Correlation Matrix The mean correlation between cells Uses a fiducial power spectrum Iterate during the analysis

Whitening Transform Remove expected count n i ‘Whitened’ counts Can be extended to other types of noise => systematic effects Diagonalization: overdensity eigenmodes Truncation optimizes the overdensity

Truncation Truncate at 30 Mpc/h –avoid most non-linear effects –keep decent number of modes