Applied Bayesian Inference for Agricultural Statisticians Robert J. Tempelman Department of Animal Science Michigan State University 1.

Slides:



Advertisements
Similar presentations
Bayesian Estimation in MARK
Advertisements

Part 24: Bayesian Estimation 24-1/35 Econometrics I Professor William Greene Stern School of Business Department of Economics.
Gibbs Sampling Qianji Zheng Oct. 5th, 2010.
Bayesian statistics – MCMC techniques
BAYESIAN INFERENCE Sampling techniques
Maximum likelihood (ML) and likelihood ratio (LR) test
Group analyses of fMRI data Methods & models for fMRI data analysis 26 November 2008 Klaas Enno Stephan Laboratory for Social and Neural Systems Research.
Mixed models Various types of models and their relation
1 Overview of Hierarchical Modeling Thomas Nichols, Ph.D. Assistant Professor Department of Biostatistics Mixed Effects.
Using ranking and DCE data to value health states on the QALY scale using conventional and Bayesian methods Theresa Cain.
Maximum likelihood (ML)
Standard Error of the Mean
Extension of Bayesian procedures to integrate and to blend multiple external information into genetic evaluations J. Vandenplas 1,2, N. Gengler 1 1 University.
Introduction to Multilevel Modeling Using SPSS
IE 594 : Research Methodology – Discrete Event Simulation David S. Kim Spring 2009.
1 CE 530 Molecular Simulation Lecture 7 David A. Kofke Department of Chemical Engineering SUNY Buffalo
Applied Bayesian Inference, KSU, April 29, 2012 §  / §❻ Hierarchical (Multi-Stage) Generalized Linear Models Robert J. Tempelman 1.
Genetic Regulatory Network Inference Russell Schwartz Department of Biological Sciences Carnegie Mellon University.
Fixed vs. Random Effects Fixed effect –we are interested in the effects of the treatments (or blocks) per se –if the experiment were repeated, the levels.
Estimating parameters in a statistical model Likelihood and Maximum likelihood estimation Bayesian point estimates Maximum a posteriori point.
7/16/2014Wednesday Yingying Wang
SPM Course Zurich, February 2015 Group Analyses Guillaume Flandin Wellcome Trust Centre for Neuroimaging University College London With many thanks to.
Applied Bayesian Inference, KSU, April 29, 2012 § ❷ / §❷ An Introduction to Bayesian inference Robert J. Tempelman 1.
Corinne Introduction/Overview & Examples (behavioral) Giorgia functional Brain Imaging Examples, Fixed Effects Analysis vs. Random Effects Analysis Models.
Group analyses of fMRI data Methods & models for fMRI data analysis November 2012 With many thanks for slides & images to: FIL Methods group, particularly.
Markov Random Fields Probabilistic Models for Images
Multilevel Modeling Software Wayne Osgood Crime, Law & Justice Program Department of Sociology.
Empirical Efficiency Maximization: Locally Efficient Covariate Adjustment in Randomized Experiments Daniel B. Rubin Joint work with Mark J. van der Laan.
“Users of scooters must transfer to a fixed seat” (Notice in Brisbane Taxi)
W w w. u o t t a w a. c a Mathematics and Statistics Faculty of Science Probability and Statistics Group Dept: (613) Fax: (613)
Mobile Robot Localization (ch. 7)
Mixed Effects Models Rebecca Atkins and Rachel Smith March 30, 2015.
Tracking Multiple Cells By Correspondence Resolution In A Sequential Bayesian Framework Nilanjan Ray Gang Dong Scott T. Acton C.L. Brown Department of.
The famous “sprinkler” example (J. Pearl, Probabilistic Reasoning in Intelligent Systems, 1988)
CHAPTER 17 O PTIMAL D ESIGN FOR E XPERIMENTAL I NPUTS Organization of chapter in ISSO –Background Motivation Finite sample and asymptotic (continuous)
Markov Chain Monte Carlo for LDA C. Andrieu, N. D. Freitas, and A. Doucet, An Introduction to MCMC for Machine Learning, R. M. Neal, Probabilistic.
Point Estimation of Parameters and Sampling Distributions Outlines:  Sampling Distributions and the central limit theorem  Point estimation  Methods.
INTRODUCTION TO CLINICAL RESEARCH Introduction to Statistical Inference Karen Bandeen-Roche, Ph.D. July 12, 2010.
Chapter 5 Multilevel Models
Statistics Presentation Ch En 475 Unit Operations.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Stats Term Test 4 Solutions. c) d) An alternative solution is to use the probability mass function and.
Bayesian Modelling Harry R. Erwin, PhD School of Computing and Technology University of Sunderland.
Bayesian statistics named after the Reverend Mr Bayes based on the concept that you can estimate the statistical properties of a system after measuting.
Statistical Methods. 2 Concepts and Notations Sample unit – the basic landscape unit at which we wish to establish the presence/absence of the species.
Density Estimation in R Ha Le and Nikolaos Sarafianos COSC 7362 – Advanced Machine Learning Professor: Dr. Christoph F. Eick 1.
STA302/1001 week 11 Regression Models - Introduction In regression models, two types of variables that are studied:  A dependent variable, Y, also called.
Markov Chain Monte Carlo in R
Group Analyses Guillaume Flandin SPM Course London, October 2016
MCMC Output & Metropolis-Hastings Algorithm Part I
Advanced Statistical Computing Fall 2016
Data Science and Statistical Agencies
Bayesian data analysis
Generalized Linear Models
Introduction to the bayes Prefix in Stata 15
Maximum Likelihood Estimation
(Very Brief) Introduction to Bayesian Statistics
'Linear Hierarchical Models'
OVERVIEW OF LINEAR MODELS
Pairwise Sequence Alignment (cont.)
Linear Hierarchical Modelling
OVERVIEW OF LINEAR MODELS
The general linear model and Statistical Parametric Mapping
Expectation-Maximization & Belief Propagation
Opinionated Lessons #39 MCMC and Gibbs Sampling in Statistics
CS639: Data Management for Data Science
WellcomeTrust Centre for Neuroimaging University College London
Maximum Likelihood Estimation (MLE)
Regression Models - Introduction
Presentation transcript:

Applied Bayesian Inference for Agricultural Statisticians Robert J. Tempelman Department of Animal Science Michigan State University 1

Outline of talk: Introduction Review of Likelihood Inference An Introduction to Bayesian Inference Empirical Bayes Inference – The Bayesian Connection to Generalized Linear Mixed Models (GLMM) The Bayesian Revolution: – Markov Chain Monte Carlo (MCMC) methods – Metropolis Hastings sampling Comparisons of Bayesian with conventional GLMM analyses of agricultural experiments. Extensions of GLMM analyses of agricultural experiments using hierarchical Bayesian inference. 2

Warning You won’t learn Bayesian data analysis in one day…and there is still a lot I don’t know. Some great resources for agricultural statisticians/data analysts 3 Sorensen and Gianola Gelman et al. Carlin and Louis

How did I get interested in Bayesian Statistics? : Masters of Science in Animal Breeding, University of Guelph – Additive & Dominance Genetic Variation for milk yield in dairy cows Y = X  + Z a u a + Z d u d + e; e nx1 ~ N (0,I  2 e ); u a(qx1) ~ N (0,A  2 a ); u d(qx1) ~ N (0,D  2 d ); Y nx1 q > n  : Fixed effects u a, u d : Random effects 4 A, D: known correlation matrices

Inference issues What was known: – E(u|  GLS ,  2 a,  2 d,  2 e,y) was Best Linear Unbiased Predictor (BLUP) of u. But typically, don’t know variance components (VC):  2 a,  2 d,  2 e – Default: Use REML (Restricted Maximum Likelihood) to estimate VC. – E(u|  E-GLS , REML (  2 a,  2 d,  2 e ),y) is Empirical Best Linear Unbiased Predictor (E-BLUP). – Use Henderson’s Mixed Model Equations to get this. – What are properties of E-BLUP, E-GLS based on REML estimates of VC?....don’t ask, don’t tell. 5

An even potentially bigger inference issue Generalized linear mixed models (GLMM). – i.e., for the analysis of non-normal data? Binary, Count, etc. – Inference in GLMM analyses is often asymptotic (based on behavior in “large samples”) → even when VC are known. – What are the implications if VC are unknown? 6

From last year’s KSU workshop (Walt Stroup) Generalized linear mixed models: What’s really important – Probability distributions – For non-normal data, the model equation form Y = X  + Zu + e is not useful..it’s counterproductive. Formal tests are based on asymptotic (“large sample”) approximations. – Nice properties when n is “large” When is n large enough? – Quasi-likelihood (PROC GENMOD)…what’s that? “vacuous” (Walt Stroup) for repeated measures specs.→ can’t even simulate data generation process. Can we do better? I think so. 7

Fall, 1989: Phd Program University of Illinois- here I come! Journal of Animal Science,

My motivation for learning/understanding Bayesian statistics? Pragmatic, not philosophical. – Animal breeders are incredibly eclectic…they just want to solve problems in animal genetics! “Physicists and engineers very often become immersed in the subject matter. In particular, they work hand in hand with neuroscientists and often become experimentalists themselves. Furthermore, engineers (and likewise computer scientists) are ambitious; when faced with problems, they tend to attack, sweeping aside impediments stemming from limited knowledge about the procedures that they apply” – From “What is Statistics” by Brown and Kass (2009) in The American Statistician. – This is also the culture of statistical genetics /genomics/ animal breeding….and is the culture of data analysts. 9

Bayesian statistics Why the fuss…its philosophy is so messy? We’ve been doing things ok already…right? What’s wrong with our current toolkit? – Linear mixed models (LMM). Nothing really for classical assumptions. – Generalized linear mixed models (GLMM) Depends on the distribution…binary’s the worst to deal with. – Nonlinear mixed models. (NLMM) Not much wrong for classical assumptions and n is “large enough” Won’t address in this workshop. 10

The real issues? 1. Asymptotics – Likelihood inference is often based on approximations. – “large n” is really more than sample size. Depends on p. Depends on data distribution (e.g. binary vs. cont.). Depends on model complexity (i.e., design). 2. Flexibility. – Can we go beyond the (G)(N)LMM? 11