Bayesian analysis for Pulsar Timing Arrays Rutger van Haasteren (Leiden) Yuri Levin (Leiden) Pat McDonald (CITA) Ting-Ting Lu (Toronto)

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Bayesian Belief Propagation
University of Joensuu Dept. of Computer Science P.O. Box 111 FIN Joensuu Tel fax Gaussian Mixture.
Image Modeling & Segmentation
CSCE643: Computer Vision Bayesian Tracking & Particle Filtering Jinxiang Chai Some slides from Stephen Roth.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
CS479/679 Pattern Recognition Dr. George Bebis
Pattern Recognition and Machine Learning
Bayesian inference “Very much lies in the posterior distribution” Bayesian definition of sufficiency: A statistic T (x 1, …, x n ) is sufficient for 
Bayesian Estimation in MARK
1 Parametric Sensitivity Analysis For Cancer Survival Models Using Large- Sample Normal Approximations To The Bayesian Posterior Distribution Gordon B.
Probabilistic Robotics
Segmentation and Fitting Using Probabilistic Methods
Rob Fergus Courant Institute of Mathematical Sciences New York University A Variational Approach to Blind Image Deconvolution.
Aspects of Conditional Simulation and estimation of hydraulic conductivity in coastal aquifers" Luit Jan Slooten.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
CHAPTER 16 MARKOV CHAIN MONTE CARLO
Symposium CoRoT 2009 – Cité Universitaire – 2 February 2009 Analysis of power spectra of Sun-like stars using a Bayesian Approach Thierry Appourchaux Symposium.
Visual Recognition Tutorial
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Bayesian estimation Bayes’s theorem: prior, likelihood, posterior
Computational statistics 2009 Random walk. Computational statistics 2009 Random walk with absorbing barrier.
Part 4 b Forward-Backward Algorithm & Viterbi Algorithm CSE717, SPRING 2008 CUBS, Univ at Buffalo.
Predictive Automatic Relevance Determination by Expectation Propagation Yuan (Alan) Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani.
Lecture 5: Learning models using EM
Nonlinear and Non-Gaussian Estimation with A Focus on Particle Filters Prasanth Jeevan Mary Knox May 12, 2006.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Search for the Gravitational Wave Memory effect with the Parkes Pulsar Timing Array Jingbo Wang 1,2,3, Hobbs George 3, Dick Manchester 3, Na Wang 1,4 1.
Statistical Analysis of Systematic Errors and Small Signals Reinhard Schwienhorst University of Minnesota 10/26/99.
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
1 Bayesian methods for parameter estimation and data assimilation with crop models Part 2: Likelihood function and prior distribution David Makowski and.
PATTERN RECOGNITION AND MACHINE LEARNING
Chi-squared distribution  2 N N = number of degrees of freedom Computed using incomplete gamma function: Moments of  2 distribution:
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Computer vision: models, learning and inference Chapter 19 Temporal models.
Harrison B. Prosper Workshop on Top Physics, Grenoble Bayesian Statistics in Analysis Harrison B. Prosper Florida State University Workshop on Top Physics:
Lecture note for Stat 231: Pattern Recognition and Machine Learning 4. Maximum Likelihood Prof. A.L. Yuille Stat 231. Fall 2004.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 11: Bayesian learning continued Geoffrey Hinton.
Bayesian Inversion of Stokes Profiles A.Asensio Ramos (IAC) M. J. Martínez González (LERMA) J. A. Rubiño Martín (IAC) Beaulieu Workshop ( Beaulieu sur.
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Probability and Measure September 2, Nonparametric Bayesian Fundamental Problem: Estimating Distribution from a collection of Data E. ( X a distribution-valued.
Efficient computation of Robust Low-Rank Matrix Approximations in the Presence of Missing Data using the L 1 Norm Anders Eriksson and Anton van den Hengel.
PROCESS MODELLING AND MODEL ANALYSIS © CAPE Centre, The University of Queensland Hungarian Academy of Sciences Statistical Model Calibration and Validation.
Bayesian Reasoning: Tempering & Sampling A/Prof Geraint F. Lewis Rm 560:
Additional Topics in Prediction Methodology. Introduction Predictive distribution for random variable Y 0 is meant to capture all the information about.
Guest lecture: Feature Selection Alan Qi Dec 2, 2004.
Lecture 2: Statistical learning primer for biologists
Detecting Gravitational Waves with a Pulsar Timing Array LINDLEY LENTATI CAMBRIDGE UNIVERSITY.
Trond Reitan (Division of statistics and insurance mathematics, Department of Mathematics, University of Oslo) Statistical modelling and latent variables.
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
6. Population Codes Presented by Rhee, Je-Keun © 2008, SNU Biointelligence Lab,
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Markov Chain Monte Carlo in R
Alan Qi Thomas P. Minka Rosalind W. Picard Zoubin Ghahramani
Multidimensional Integration Part I
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
J. Ellis, F. Jenet, & M. McLaughlin
Learning From Observed Data
Probabilistic Surrogate Models
Uncertainty Propagation
Stochastic Methods.
Presentation transcript:

Bayesian analysis for Pulsar Timing Arrays Rutger van Haasteren (Leiden) Yuri Levin (Leiden) Pat McDonald (CITA) Ting-Ting Lu (Toronto)

Pulsar Timing Array GW timing residuals: Multidimentional Gaussian process; coherence matrix C (A, n)=G(t –t ) Q ai bj Jenet et al 04 Hill & Benders 1981 amplitude slope Phinney 01 Jaffe & Backer 03 Wyithe & Loeb 03 j i ab GWB spectrum geometry

Pulsar Timing Array C= C (A, n)+noise +noise +…+noise ai bj Real coherence matrix: Bayesian solution: parametrize each pulsar noise reasonably: N exp[-n f]+σ construct multidimantional probability distribution marginalize over quadradic spindowns – analytical marginalize over pulsar noises – numerical P(A,n, N,n | data)=exp[-X (2C) X - (1/2) log(det{C})] (Prior/Norm) x x where X=(timing residuals) – (quadratic spidown) a a a a ai

Markov Chain Monte-Carlo Direct integration unrealistic Markoff chain cleverly explores parameter space, dwelling in high-probability regions Typically need few points for reliable convergence Can do white pulsar noises-last year’s talk However, problems with chain convergence when one allows for colored pulsar noises. Need something different!

Maximum-likelihood method Find global maximum of log[P(A, n, N )] Run a chain in the neighbourhood until enough points to fit a quadratic form: log(P)=log(P ) – (p –p ) Q (p – p ) Approximate P as a Gaussian and marginalize over pulsar noises a 0i i i j jj 00

Results 10 pulsars 500 ns, 70 timings each over 9 yr

Results 10 pulsars 500 ns, 70 timings each over 9 yr

Results 10 pulsars 100 ns, 70 timings each over 9 yr

Results 10 pulsars 100 ns, 70 timings each over 9 yr

Results 10 pulsars, 50 ns timing error, 5 years, every 2.5 weeks A=10 E-15 n=-7/3

our algorithm: Does not rely on estimators – explores the full multi-dimensional likelihood function Measures simultaneously amplitude AND slope of the gravitational-wave background Deals easily with unevenly sampled data, variable number of tracked pulsars, etc. Deals easily with systematics-quadratic spindowns, zero resets, pointing errors, and human errors of known functional form.

Example problem: finding the white noise amplitude b Pulsar observer: b =(b + … +b )/N N Error = b/N 0.5

Example problem: finding the white noise amplitude b Bayesian Theorist: P(data|b)=exp[(b + … +b )/2b -.5 log(b)] N 2 P(b|data)=(1/K) P(data|b) P (b) 0

Example problem: finding the white noise amplitude b Bayesian Theorist: P(data|b)=exp[(b + … +b )/2b -.5 log(b)] N 2 P(b|data)=(1/K) P(data|b) P (b) 0 normalization prior

Example problem: finding the white noise amplitude b Bayesian Theorist: b P

Complication: white noise + jump a

Pulsar observer: fit for a Lazy Bayesian Theorist: 1.Find P(a,b|data) 2.Integrate over a

Complication: white noise + jump a Pulsar observer: fit for a Lazy Bayesian Theorist: 1.Find P(a,b|data) 2.Integrate over a ANALYTICAL!

Complication: white noise + jump a Pulsar observer: fit for a Lazy Bayesian Theorist: 1.Find P(a,b|data) 2.Integrate over a 3. Get expression P(b|data), insensitive to jumps!

Jump removal:

Does not have to be jumps. ANYTHING of known functional form, i.e.: Quadratic/cubic pulsar spindowns Annual variations Periodicity due to Jupiter Zero resets ISM variations, if measured independently can be removed analytically when writing down P(b). Don’t care if pre-fit by observers or not.

Pulsar Timing Array C (A, n) ai bj Bayesian analysis: compute P(A,n| data), after “removing” unwanted components of known functional form easy

Pulsar Timing Array C (A, n) ai bj Bayesian analysis: compute P(A.n| data), after “removing” unwanted components of known functional form Complication 1: low-frequency cut-off

Pulsar Timing Array C (A, n) ai bj Bayesian analysis: compute P(A.n| data), after “removing” unwanted components of known functional form Complication 1: low-frequency cut-off

Pulsar Timing Array C (A, n) ai bj Bayesian analysis: compute P(A.n| data), after “removing” unwanted components of known functional form Complication 2: pulsar noises, measured concurrently with GWs. This is the real difficulty with the Bayesian Method.

Results

Strengths of B. approach Philosophy No loss of info, no need to choose optimal estimator No noise whitening, etc. Irregular time intervals, etc. Easy removal of unwanted functions Weaknesses: Computational cost Need better algorithms!

PhD position in Leiden Supported by 5-yr VIDI grant Collaboration with observers/other theorists essential ….