BXA robust parameter estimation & practical model comparison

Slides:



Advertisements
Similar presentations
Part 2: Unsupervised Learning
Advertisements

J. Daunizeau Institute of Empirical Research in Economics, Zurich, Switzerland Brain and Spine Institute, Paris, France Bayesian inference.
MCMC estimation in MlwiN
Probabilistic models Haixu Tang School of Informatics.
Bayesian inference of normal distribution
Biointelligence Laboratory, Seoul National University
Bayesian Estimation in MARK
Efficient Cosmological Parameter Estimation with Hamiltonian Monte Carlo Amir Hajian Amir Hajian Cosmo06 – September 25, 2006 Astro-ph/
Flipping A Biased Coin Suppose you have a coin with an unknown bias, θ ≡ P(head). You flip the coin multiple times and observe the outcome. From observations,
Markov-Chain Monte Carlo
1 Removing Camera Shake from a Single Photograph Rob Fergus, Barun Singh, Aaron Hertzmann, Sam T. Roweis and William T. Freeman ACM SIGGRAPH 2006, Boston,
Rob Fergus Courant Institute of Mathematical Sciences New York University A Variational Approach to Blind Image Deconvolution.
1 Vertically Integrated Seismic Analysis Stuart Russell Computer Science Division, UC Berkeley Nimar Arora, Erik Sudderth, Nick Hay.
Visual Recognition Tutorial
Computing the Posterior Probability The posterior probability distribution contains the complete information concerning the parameters, but need often.
EE 290A: Generalized Principal Component Analysis Lecture 6: Iterative Methods for Mixture-Model Segmentation Sastry & Yang © Spring, 2011EE 290A, University.

Bayesian Analysis of X-ray Luminosity Functions A. Ptak (JHU) Abstract Often only a relatively small number of sources of a given class are detected in.
Lecture 5: Learning models using EM
Lecture 16 – Thurs, Oct. 30 Inference for Regression (Sections ): –Hypothesis Tests and Confidence Intervals for Intercept and Slope –Confidence.
Visual Recognition Tutorial
Arizona State University DMML Kernel Methods – Gaussian Processes Presented by Shankar Bhargav.
Bayesian Learning Rong Jin.
G. Cowan RHUL Physics Bayesian Higgs combination page 1 Bayesian Higgs combination using shapes ATLAS Statistics Meeting CERN, 19 December, 2007 Glen Cowan.
Chapter Two Probability Distributions: Discrete Variables
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
Bayesian parameter estimation in cosmology with Population Monte Carlo By Darell Moodley (UKZN) Supervisor: Prof. K Moodley (UKZN) SKA Postgraduate conference,
Model Inference and Averaging
Mixture Models, Monte Carlo, Bayesian Updating and Dynamic Models Mike West Computing Science and Statistics, Vol. 24, pp , 1993.
Learning Theory Reza Shadmehr Linear and quadratic decision boundaries Kernel estimates of density Missing data.
- 1 - Bayesian inference of binomial problem Estimating a probability from binomial data –Objective is to estimate unknown proportion (or probability of.
July 11, 2006Bayesian Inference and Maximum Entropy Probing the covariance matrix Kenneth M. Hanson T-16, Nuclear Physics; Theoretical Division Los.
LIGO- G Z Analyzing Event Data Lee Samuel Finn Penn State University Reference: T030017, T
The generalization of Bayes for continuous densities is that we have some density f(y|  ) where y and  are vectors of data and parameters with  being.
MCMC reconstruction of the 2 HE cascade events Dmitry Chirkin, UW Madison.
Bayesian Travel Time Reliability
Reducing MCMC Computational Cost With a Two Layered Bayesian Approach
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Fitting normal distribution: ML 1Computer vision: models, learning and inference. ©2011 Simon J.D. Prince.
Chapter 2: Bayesian hierarchical models in geographical genetics Manda Sayler.
Introduction: Metropolis-Hasting Sampler Purpose--To draw samples from a probability distribution There are three steps 1Propose a move from x to y 2Accept.
DirectFit reconstruction of the Aya’s two HE cascade events Dmitry Chirkin, UW Madison Method of the fit: exhaustive search simulate cascade events with.
Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability Primer Bayesian Brain Probabilistic Approaches to Neural Coding 1.1 A Probability.
Hierarchical Models. Conceptual: What are we talking about? – What makes a statistical model hierarchical? – How does that fit into population analysis?
Prediction and Missing Data. Summarising Distributions ● Models are often large and complex ● Often only interested in some parameters – e.g. not so interested.
Bayesian Inference: Multiple Parameters
Generalization Performance of Exchange Monte Carlo Method for Normal Mixture Models Kenji Nagata, Sumio Watanabe Tokyo Institute of Technology.
Crash course in probability theory and statistics – part 2 Machine Learning, Wed Apr 16, 2008.
Markov Chain Monte Carlo in R
Bayesian estimation for time series models
CEE 6410 Water Resources Systems Analysis
Figure Legend: From: Bayesian inference for psychometric functions
Model Inference and Averaging
Smithsonian Astrophysical Observatory
Special Topics In Scientific Computing
Remember that our objective is for some density f(y|) for observations where y and  are vectors of data and parameters,  being sampled from a prior.
Estimating the Spatial Sensitivity Function of A Light Sensor N. K
Location-Scale Normal Model
Multidimensional Integration Part I
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Biointelligence Laboratory, Seoul National University
LECTURE 07: BAYESIAN ESTIMATION
A Flexible Bayesian Framework for Modeling Haplotype Association with Disease, Allowing for Dominance Effects of the Underlying Causative Variants  Andrew.
Bayesian inference J. Daunizeau
Mathematical Foundations of BME
Ch 3. Linear Models for Regression (2/2) Pattern Recognition and Machine Learning, C. M. Bishop, Previously summarized by Yung-Kyun Noh Updated.
Mixture Models with Adaptive Spatial Priors
Yalchin Efendiev Texas A&M University
Presentation transcript:

BXA robust parameter estimation & practical model comparison Johannes Buchner FONDECYT fellow http://astrost.at/istics/

BXA – Bayesian X-ray Analysis Plugin connecting xspec/sherpa with MultiNest https://github.com/JohannesBuchner/BXA Nested Sampling vs. MCMC for parameter estimation Model comparison with evidences/Bayes factors Population inference instead of samples statistics Buchner+14 Buchner+15

L, NH from X-ray spectrum Scattered Powerlaw component CTK flat spectrum + FeK line

L, NH from X-ray spectrum Probability cloud

L, NH from X-ray spectrum MCMC (emcee/GW) Nested Sampling (BXA/Multinest) Simple fitting: One minimum MCMC: One minimum convergence tests can be fooled

Missing ingredients MCMC: Insert tuned transition kernel NS: Insert constrained drawing algorithm General solutions: MultiNest, MCMC, HMCMC, Galilean, RadFriends, PolyChord

Buchner (2014)

Working with BXA Do not need to abandon xspec/sherpa Specify working parameter space log, linear, etc. ~ priors – or ML MCMC approach Run once, be done! No stepping, no convergence worries No wondering about other solutions Output: PDF of each parameter

BXA in practice (with pyxspec) s = Spectrum('example-file.fak') s.ignore("**"); s.notice("0.2-8.0") m = Model("pow") Set up spectrum & model Set up parameter space Run! m.powerlaw.norm.values = ",,1e-10,1e-10,1e1,1e1" # 10^-10 .. 10 m.powerlaw.PhoIndex.values = ",,1,1,3,3" # 1 .. 3 # define prior transformations = [ bxa.create_uniform_prior_for( m, m.powerlaw.PhoIndex), bxa.create_jeffreys_prior_for(m, m.powerlaw.norm), ] bxa.standard_analysis(transformations, outputfiles_basename = 'simple-example')

Parameter probability distributions That was parameter estimation – now for model comparison! Parameter correlations

Model comparison Z = p(D|M) = “marginal likelihood”, “evidence” … see Buchner+14 Z = p(D|M) = “marginal likelihood”, “evidence” … average likelihood over parameter space Z1/Z2: Probability for model M1 compared to M2 Example: M1: wabs*pl – photo-el absorbed powerlaw M2: table{sphere}: includes Compton scattering M1: M2: Z1/Z2 was 10 CDFS source 179

Comparison of several models Zi = p(Mi|D) see Buchner+14 Model i

Example applications Comparison between obscurer geometries (Buchner+14) Identifying relativistic broad lines: Narrow line model vs. Relativistic broad line models (Baronchelli+in prep.) Threshold for decisions? False selection rates?

False selection rates As with any other estimator! Simulate many times without effect measure false decisions Calibrate threshold Feasible with BXA

Population inference Example: Photon index, NH distribution of population of LGRBs, AGN sample statistics (selection effects – see Buchner+15, A1) Keep uncertainties, Upper limits Don’t: Plot histogram of means Stack posteriors

Hierarchical Bayesian inference Assume flexible population model Examples: Gaussian, Gaussian mixture, beta distribution Fold in posterior samples for object i Likelihood for population parameters - MCMC Importance sampling from posteriors need flat priors, see Buchner+17a for details Product over objects Mean over samples

Population inference Keeps uncertainties Fold in redshift uncertainties (unlike X-ray stacking!) Population uncertainties Include limited sample size Model can be 1d,2d,3d,… Special case of luminosity function analysis (Loredo2004)

Summary BXA: a MultiNest plugin for xspec/sherpa BXA: parameter estimation Robust to multiple solutions, convergence issues, etc. BXA: Bayesian model comparison Sensitive; can be interpreted; calibrated to FDR Population inference Understand the population from a limited sample Forward-fold all uncertainties more in: Buchner+14, +15, +17a + their refs