Juande Zornoza UW Madison

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

Pattern Recognition and Machine Learning
CS479/679 Pattern Recognition Dr. George Bebis
B.Macukow 1 Lecture 12 Neural Networks. B.Macukow 2 Neural Networks for Matrix Algebra Problems.
RooUnfold unfolding framework and algorithms
JET CHARGE AT LHC Scott Lieberman New Jersey Institute of Technology TUNL REU 2013 Duke High Energy Physics Group Working with: Professor Ayana Arce Dave.
Data mining and statistical learning - lecture 6
Maximum likelihood (ML) and likelihood ratio (LR) test
Machine Learning CMPT 726 Simon Fraser University
Energy Reconstruction Algorithms for the ANTARES Neutrino Telescope J.D. Zornoza 1, A. Romeyer 2, R. Bruijn 3 on Behalf of the ANTARES Collaboration 1.
Linear and generalised linear models
Linear and generalised linear models
Linear and generalised linear models Purpose of linear models Least-squares solution for linear models Analysis of diagnostics Exponential family and generalised.
Maximum likelihood (ML)
Adaptive Signal Processing
Chapter 15 Modeling of Data. Statistics of Data Mean (or average): Variance: Median: a value x j such that half of the data are bigger than it, and half.
Unfolding jet multiplicity and leading jet p T spectra in jet production in association with W and Z Bosons Christos Lazaridis University of Wisconsin-Madison.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
Progress in identification of damping: Energy-based method with incomplete and noisy data Marco Prandina University of Liverpool.
RooUnfold unfolding framework and algorithms Tim Adye Rutherford Appleton Laboratory ATLAS RAL Physics Meeting 20 th May 2008.
RooUnfold unfolding framework and algorithms Tim Adye Rutherford Appleton Laboratory Oxford ATLAS Group Meeting 13 th May 2008.
Data collected during the year 2006 by the first 9 strings of IceCube can be used to measure the energy spectrum of the atmospheric muon neutrino flux.
Chapter 28 Cononical Correction Regression Analysis used for Temperature Retrieval.
Unfolding in ALICE Jan Fiete Grosse-Oetringhaus, CERN for the ALICE collaboration PHYSTAT 2011 CERN, January 2011.
JET CHARGE AT LHC Scott Lieberman New Jersey Institute of Technology TUNL REU 2013 Duke High Energy Physics Group Working with: Professor Ayana Arce Dave.
2005 Unbinned Point Source Analysis Update Jim Braun IceCube Fall 2006 Collaboration Meeting.
8th December 2004Tim Adye1 Proposal for a general-purpose unfolding framework in ROOT Tim Adye Rutherford Appleton Laboratory BaBar Statistics Working.
Spectrum Reconstruction of Atmospheric Neutrinos with Unfolding Techniques Juande Zornoza UW Madison.
Chapter 61 Chapter 7 Review of Matrix Methods Including: Eigen Vectors, Eigen Values, Principle Components, Singular Value Decomposition.
Computacion Inteligente Least-Square Methods for System Identification.
Geology 5670/6670 Inverse Theory 6 Feb 2015 © A.R. Lowry 2015 Read for Mon 9 Feb: Menke Ch 5 (89-114) Last time: The Generalized Inverse; Damped LS The.
Muon Energy reconstruction in IceCube and neutrino flux measurement Dmitry Chirkin, University of Wisconsin at Madison, U.S.A., MANTS meeting, fall 2009.
IC40 Spectrum Unfolding 7/1/2016Warren Huelsnitz1 SVD Method described in A. Höcker and V. Kartvelishvili, NIM A 372 (1996) 469NIM A 372 (1996) 469 Implemented.
Introduction We consider the data of ~1800 phenotype measurements Each mouse has a given probability distribution of descending from one of 8 possible.
I. Statistical Methods for Genome-Enabled Prediction of Complex Traits OUTLINE THE CHALLENGES OF PREDICTING COMPLEX TRAITS ORDINARY LEAST SQUARES (OLS)
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 1: INTRODUCTION.
Data Modeling Patrice Koehl Department of Biological Sciences
Lesson 8: Basic Monte Carlo integration
Chapter 3: Maximum-Likelihood Parameter Estimation
Deep Feedforward Networks
Ch 12. Continuous Latent Variables ~ 12
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
LECTURE 03: DECISION SURFACES
CH 5: Multivariate Methods
Analyzing Redistribution Matrix with Wavelet
Lecture 8:Eigenfaces and Shared Features
Data Mining Lecture 11.
Latent Variables, Mixture Models and EM
Dr. Unnikrishnan P.C. Professor, EEE
Unfolding Problem: A Machine Learning Approach
Unfolding atmospheric neutrino spectrum with IC9 data (second update)
Modern Spectral Estimation
Modelling data and curve fitting
Discrete Event Simulation - 4
Filtering and State Estimation: Basic Concepts
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Instructor :Dr. Aamer Iqbal Bhatti
ILC Main Linac Alignment Simulations
Singular Value Decomposition SVD
Gaussian Mixture Models And their training with the EM algorithm
Unfolding performance Data - Monte Carlo comparison
Pattern Recognition and Machine Learning
5.4 General Linear Least-Squares
Statistical Methods for Particle Physics Lecture 4: unfolding
Principles of the Global Positioning System Lecture 11
Introduction to Unfolding
Parametric Methods Berlin Chen, 2005 References:
Unfolding with system identification
Random Neural Network Texture Model
Presentation transcript:

Juande Zornoza UW Madison Spectrum Reconstruction of Atmospheric Neutrinos with Unfolding Techniques Juande Zornoza UW Madison

Introduction We will review different approaches for the reconstruction of the energy spectrum of atmospheric neutrinos: Blobel / Singular Value Decomposition actually, both methods are basically the same, with only differences in issues not directly related with the unfolding but with the regularization and so on Iterative Method based on Bayes’ theorem

Spectrum reconstruction In principle, energy spectra can be obtained using the reconstructed energy of each event. However, this is not efficient in our case because of the combination of two factors: Fast decrease (power-law) of the flux. Large fluctuations in the energy deposition. For this reason, an alternative way has to be used: unfolding techniques.

Spectrum unfolding Quantity to obtain: y, which follows pdf→ftrue(y) Measured quantity: b, which follows pdf→fmeas(b) Both are related by the Fredholm integral equation of first kind: Matrix notation The response matrix takes into account three factors: -Limited acceptance -Finite resolution -Transformation The response matrix inversion gives useless solutions, due to the effect of statistical fluctuations

Dealing with instabilities Regularization: Solution with minimum curvature Principle of maximum entropy Iterative procedure, leading asymptotically to the unfolded distribution

Single Value Decomposition1 The response matrix is decomposed as: U, V: orthogonal matrices S: non-negative diagonal matrix (si, singular values) This can be also seen as a minimization problem: Or, introducing the covariance matrix to take into account errors: 1 A. Hoecker, Nucl. Inst. Meth. in Phys. Res. A 372:469 (1996)

SVD: normalization Actually, it is convenient to normalize the unknowns Aij contains the number of events, not the probability. Advantages: the vector w should be smooth, with small bin-to-bin variations avoid weighting too much the cases of 100% probability when only one event is in the bin

SVD: Rescaling Rotation of matrices: allows to rewrite the system with a covariance matrix equal to I, more convenient to work with:

Regularization Several methods have been proposed for the regularization. The most common is to add a curvature term add a curvature term Other option: principle of maximum entropy

Regularization We have transformed the problem in the optimization of the value of , which tunes how much regularization we include:  too large: physical information lost  too small: statistical fluctuations spoil the result In order to optimize the value of : Evaluation using MC information Maximum curvature of the L-curve Components of vector L-curve

Solution to the system Actually, the solution to the system with the curvature term can be expressed as a function of the solution without curvature: where (Tikhonov factors)

Tikhonov factors The non-zero tau is equivalent to change di by And this allows to find a criteria to find a good tau Components of d fun0 fun1 k fun2  = sk2

Differences between SVD and Blobel More simplified implementation Possibility of different number of bins in y and b (non square A) Different curvature term Selection of optimum tau B-splines used in the standard Blobel implementation

Bayesian Iterative Method2 If there are several causes (Ei) which can produce an effect Xj and we know the initial probability of the causes P(Ei), the conditional probability of the cause to be Ei when X is observed is: prior guess: iterative approach smearing matrix: MC The expected number of events to be assigned to each of the causes is: experimental data (simulated) The dependence on the initial probability P0(Ei) can be overcome by an iterative process. 2 G. D'Agostini NIM A362(1995) 487-498

Reconsructed spectrum Iterative algorithm Choose the initial distribution P0(E). For instance, a good guess could be the atmospheric flux (without either prompt neutrinos or signal). Calculate and . Compare to . Replace by and by . Go to step 2. P(Xj|Ei) Smearing matrix (MC) Reconsructed spectrum P(Ei|Xj) n(Ei) P(Ei) no(E) Po(E) Initial guess n(Xj) Experimental data

For IceCube Several parameters can be investigated: Number of channels Number of NPEs Reconstructed energy Neural network output… With IceCube, we will have much better statistics than with AMANDA But first, reconstruction with 9 strings will be the priority

Remarks First, a good agreement between data and MC is necessary Different unfolding methods will be compared (several internal parameters to tune in each method) Several regularization techniques are also available in the literature Also an investigation on the best variable for unfolding has to be done Maybe several variables can be used in a multi-D analysis

This is the end

B-splines Spline: piecewise continuous and differentiable function that connects two neighbor points by a cubic polynomial: from H. Greene PhD. B-spline: spline functions can be expressed by a finite superposition of base functions (B-spilines). (first order) (higher orders)

Maximum entropy In general, we want to minimise S can be the spikeness of the solution, or, the entropy of the system: Probability of any event of being in bin i is fi/N. Then, following the maximum entropy principle, we will minimize: Useful for sharp gradients, i.e. when the solution it is not expected to be very smooth