Unfolding Problem: A Machine Learning Approach

Slides:



Advertisements
Similar presentations
Copula Representation of Joint Risk Driver Distribution
Advertisements

Ordinary Least-Squares
General Linear Model With correlated error terms  =  2 V ≠  2 I.
A. The Basic Principle We consider the multivariate extension of multiple linear regression – modeling the relationship between m responses Y 1,…,Y m and.
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
6-1 Introduction To Empirical Models 6-1 Introduction To Empirical Models.
Ch11 Curve Fitting Dr. Deshi Ye
N.D.GagunashviliUniversity of Akureyri, Iceland Pearson´s χ 2 Test Modifications for Comparison of Unweighted and Weighted Histograms and Two Weighted.
The General Linear Model. The Simple Linear Model Linear Regression.
Statistical Image Modelling and Particle Physics Comments on talk by D.M. Titterington Glen Cowan RHUL Physics PHYSTAT05 Glen Cowan Royal Holloway, University.
Development of Empirical Models From Process Data
7. Least squares 7.1 Method of least squares K. Desch – Statistical methods of data analysis SS10 Another important method to estimate parameters Connection.
Ordinary least squares regression (OLS)
Basics of regression analysis
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
AN ITERATIVE METHOD FOR MODEL PARAMETER IDENTIFICATION 4. DIFFERENTIAL EQUATION MODELS E.Dimitrova, Chr. Boyadjiev E.Dimitrova, Chr. Boyadjiev BULGARIAN.
1 Least squares procedure Inference for least squares lines Simple Linear Regression.
Stochastic Linear Programming by Series of Monte-Carlo Estimators Leonidas SAKALAUSKAS Institute of Mathematics&Informatics Vilnius, Lithuania
CS433: Modeling and Simulation Dr. Anis Koubâa Al-Imam Mohammad bin Saud University 15 October 2010 Lecture 05: Statistical Analysis Tools.
Unfolding jet multiplicity and leading jet p T spectra in jet production in association with W and Z Bosons Christos Lazaridis University of Wisconsin-Madison.
Progress in identification of damping: Energy-based method with incomplete and noisy data Marco Prandina University of Liverpool.
Introduction 1. Similarity 1.1. Mechanism and mathematical description 1.2. Generalized variables 1.3. Qualitative analysis 1.4. Generalized individual.
PROCESS MODELLING AND MODEL ANALYSIS © CAPE Centre, The University of Queensland Hungarian Academy of Sciences Statistical Model Calibration and Validation.
1 Iterative dynamically stabilized (IDS) method of data unfolding (*) (*arXiv: ) Bogdan MALAESCU CERN PHYSTAT 2011 Workshop on unfolding.
Monte-Carlo method for Two-Stage SLP Lecture 5 Leonidas Sakalauskas Institute of Mathematics and Informatics Vilnius, Lithuania EURO Working Group on Continuous.
Chapter 28 Cononical Correction Regression Analysis used for Temperature Retrieval.
Data Modeling Patrice Koehl Department of Biological Sciences National University of Singapore
Machine Learning 5. Parametric Methods.
Least Squares Estimate Additional Notes 1. Introduction The quality of an estimate can be judged using the expected value and the covariance matrix of.
FORECASTING METHODS OF NON- STATIONARY STOCHASTIC PROCESSES THAT USE EXTERNAL CRITERIA Igor V. Kononenko, Anton N. Repin National Technical University.
Spectrum Reconstruction of Atmospheric Neutrinos with Unfolding Techniques Juande Zornoza UW Madison.
Computacion Inteligente Least-Square Methods for System Identification.
Simple and multiple regression analysis in matrix form Least square Beta estimation Beta Simple linear regression Multiple regression with two predictors.
Econometrics III Evgeniya Anatolievna Kolomak, Professor.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION Statistical Interpretation of Least Squares ASEN.
I. Statistical Methods for Genome-Enabled Prediction of Complex Traits OUTLINE THE CHALLENGES OF PREDICTING COMPLEX TRAITS ORDINARY LEAST SQUARES (OLS)
Data Modeling Patrice Koehl Department of Biological Sciences
Chapter 7. Classification and Prediction
Compressive Coded Aperture Video Reconstruction
DEEP LEARNING BOOK CHAPTER to CHAPTER 6
LINEAR CLASSIFIERS The Problem: Consider a two class task with ω1, ω2.
Inference for Regression (Chapter 14) A.P. Stats Review Topic #3
(5) Notes on the Least Squares Estimate
Reduction of Variables in Parameter Inference
M. Kuhn, P. Hopchev, M. Ferro-Luzzi
Background on Classification
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
LECTURE 10: DISCRIMINANT ANALYSIS
How Good is a Model? How much information does AIC give us?
The Elements of Statistical Learning
Evgeniya Anatolievna Kolomak, Professor
Rutgers Intelligent Transportation Systems (RITS) Laboratory
Elementary Statistics
Going Backwards In The Procedure and Recapitulation of System Identification By Ali Pekcan 65570B.
LOCATION AND IDENTIFICATION OF DAMPING PARAMETERS
Matrices Definition: A matrix is a rectangular array of numbers or symbolic elements In many applications, the rows of a matrix will represent individuals.
6-1 Introduction To Empirical Models
Chapter 11: Inference for Distributions of Categorical Data
10701 / Machine Learning Today: - Cross validation,
By Viput Subharngkasen
OVERVIEW OF LINEAR MODELS
5.4 General Linear Least-Squares
OVERVIEW OF LINEAR MODELS
Generally Discriminant Analysis
Introduction to Unfolding
LECTURE 09: DISCRIMINANT ANALYSIS
Juande Zornoza UW Madison
Facultad de Ingeniería, Centro de Cálculo
Unfolding with system identification
Introduction to Machine learning
Presentation transcript:

Unfolding Problem: A Machine Learning Approach Nikolai Gagunashvili School of Computing, University of Akureyri, Iceland nikolai@unak.is 11/21/2018

Contents Introduction Basic equation System identification The unfolding procedure A numerical example Conclusions References 11/21/2018

Introduction 11/21/2018

Introduction (cont.) 11/21/2018

The unfolding problem is an underspecified problem. Introduction (cont.) The unfolding problem is an underspecified problem. Any approach to solve the problem requires a priori information about the solution. Different methods of unfolding differ, directly or indirectly, through the use of this a priori information. 11/21/2018

Dimensional independent unfolding with D-optimal system identification Basic equation We will use the linear model for a transformation of a true distribution to the measured one where f= (f1,f2,….,fm)T is vector of an experimentally measured histogram content, φ= (φ1, φ2,..,φn)T is vector of some true histogram content, ε = (ε1, ε2,.....εm)T is vector of random residual components with mean value E ε = 0, and a diagonal variance matrix Var ε = are the statistical errors of the measured distribution. 11/21/2018

Dimensional independent unfolding with D-optimal system identification is the matrix of transformation that transform true distribution to experimentally measured distribution 11/21/2018

Basic equation (cont.) 11/21/2018

Basic equation (cont.) A Least Squares Method can give an estimator for the true distribution where , the estimator, is called the unfolded distribution. The full matrix of errors of the unfolded distribution is given by according to the Least Squares Method. 11/21/2018

There are two stage of solving unfolding (inverse) problem: 1. Investigation and calculation matrix P is known as problem of identification system and may be defined as the process of determining a model of a dynamic system using observed input-output data. 2. Solution of equation (1) that gives unfolded function with complete matrix of statistical errors of . 11/21/2018

System identification or calculation of matrix P The Monte-Carlo simulation of a set-up can be used to get input-output data (training sample). Monte-Carlo Simulations • • • • To regularize the solution of the unfolding problem, let us use training set with a priori known from theory, or from other experiments distributions 11/21/2018

System identification or calculation of matrix P Assume we have q input distributions in training sample, and present them as matrix where each row represents one input training histogram. 11/21/2018

System identification or calculation of matrix P For each i-th row of the matrix P we can write the equation is a vector of i-th components of output distributions, where is i-th bin content of output distribution for the j-th training distribution is a vector of random residuals with mean value and variance where is the statistical error of i-th bin of output distribution for the j-th training distribution. 11/21/2018

System identification or calculation of matrix P Least Squares Method gives and estimator for pi Columns of matrix can correlate with each other. It means that transformation of the training distribution to the i-th bin of output distribution can be parameterized by subset of elements of the row pi. May be more than one subset that describes this transformation in sufficiently good manner. Example: 11/21/2018

System identification or calculation of matrix P Thus for each i-th output bin we will have Ni candidate rows, and for all output bins candidate matrices P. We need to choose a matrix P that is good, or optimal , in some sense. The most convenient in this case is the criteria of D-optimality that is related to the minimization of determinant of full matrix of errors of unfolded distribution 11/21/2018

System identification or calculation of matrix P Main advantages D-optimization Minimizes the volume of the confidence ellipsoid for an unfolded distribution. There are many computer algorithms for optimization. 11/21/2018

Basic equation solution A Least Squares Method give an estimator for the true distribution where , the estimator, is the unfolded distribution. The full matrix of errors of the unfolded distribution is given by according to the Least Squares Method. 11/21/2018

The unfolding procedure Initialization Define a binning for experimental data. Define a binning for the unfolding distribution. System identification Choose a set of training distributions. Calculate the D-optimal matrix P. Basic equation solution Calculate unfolded distribution with full matrix of errors Test of goodness of the unfolding Fit unfolded distribution and compare the experimental distribution and the reconstructed simulated distribution The holdout method, cross-validation method, bootstrap method. 11/21/2018

The unfolding procedure Initialization Define a binning for experimental data 11/21/2018

The unfolding procedure Initialization Define a binning for the unfolded distribution 11/21/2018

The unfolding procedure Selection criteria for set of training distributions — A training distribution has corresponding the output distribution that can be compared with the experimentally measured distribution by χ2 test. — Let us select for identification a training distribution that has a corresponding output distribution satisfying a χ2 <a selection criteria (the parameter a defines a significant level p(a) for the comparison of two histograms). Application of this selection criteria Increase the number of candidate matrix P Decrease value of determinant of full matrix of errors Decrease value of statistical errors of unfolded distribution. 11/21/2018

The unfolding procedure Selection criteria for set of training distributions Experimental distribution Monte-Carlo Set of training distributions Set of output distributions 11/21/2018

A numerical example We take a true distribution with parameters An experimentally measured distribution is defined as where the acceptance and is the detector resolution function with σ=1.5. 11/21/2018

An example of the true distribution φ(x), the acceptance function A(x) and the resolution function R(x,10) 11/21/2018

An example of the measured distribution f 11/21/2018

11/21/2018

A numerical example (cont.) A histogram of the measured distribution was obtained by simulating 104 events. Random parameters are generated uniformly on the intervals [1,3] for A1; [0.5,1.5] for A2; [8,12] for B1 ; [10,18] for B2; [0.5,1.5] for C1; [0.5,1.5] for C2; which define a training distribution for identification. 11/21/2018

Training distributions generated for system identification and an unfolded distribution for different χ2 cut 11/21/2018

Conclusions The proposed method use of a set of a priori distributions for identification to obtain stable solution of unfolding problem. D-optimization and the application of the Least Squares Method gives the possibility of minimizing the statistical errors of the solution. χ2 selection criteria permits to decrease the possible bias of the procedure. The procedure has no restriction due to dimensionality of the problem. The procedure can be applied for solving unfolding problem with smooth solution as well as non-smooth solution. Based only on a statistical approach the method has a good statistical interpretation. 11/21/2018

References V.Blobel, Unfolding methods in high-energy physics experiments, CERN 85-02 (1985). V.P.Zhigunov, Improvement of resolution function as an inverse problem, Nucl. Instrum. Meth. 216(1983)183. A.Höcker,V.Kartvelishvili, SVD approach to data unfolding, Nucl. Instrum. Meth. A372(1996)469. N.D.Gagunashvili, Unfolding of true distributions from experimental data distorted by detectors with finite resolutions, Nucl. Instrum. Meth. A451(1993)657. N.D.Gagunashvili, Unfolding with system identification, Proceedings of PHYSTAT 05, Oxford,UK. 11/21/2018