A Neural Network MonteCarlo approach to nucleon Form Factors parametrization Paris, 10-03-2011. 2° CLAS12 Europen Workshop In collaboration with: A. Bacchetta.

Slides:



Advertisements
Similar presentations
Design of Experiments Lecture I
Advertisements

3.3 Hypothesis Testing in Multiple Linear Regression
Computational Statistics. Basic ideas  Predict values that are hard to measure irl, by using co-variables (other properties from the same measurement.
Chapter 7 Statistical Data Treatment and Evaluation
Pattern Recognition and Machine Learning
Kriging.
Use of Kalman filters in time and frequency analysis John Davis 1st May 2011.
Prediction, Correlation, and Lack of Fit in Regression (§11. 4, 11
Model Assessment, Selection and Averaging
Error Component models Ric Scarpa Prepared for the Choice Modelling Workshop 1st and 2nd of May Brisbane Powerhouse, New Farm Brisbane.
Model assessment and cross-validation - overview
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Visual Recognition Tutorial
The Simple Linear Regression Model: Specification and Estimation
Pattern Recognition and Machine Learning
1 META PDFs as a framework for PDF4LHC combinations Jun Gao, Joey Huston, Pavel Nadolsky (presenter) arXiv: , Parton.
Prénom Nom Document Analysis: Parameter Estimation for Pattern Recognition Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
BA 555 Practical Business Analysis
Curve-Fitting Regression
G. Cowan Lectures on Statistical Data Analysis 1 Statistical Data Analysis: Lecture 8 1Probability, Bayes’ theorem, random variables, pdfs 2Functions of.
Ordinary Kriging Process in ArcGIS
1 BA 555 Practical Business Analysis Review of Statistics Confidence Interval Estimation Hypothesis Testing Linear Regression Analysis Introduction Case.
Objectives of Multiple Regression
1 CE 530 Molecular Simulation Lecture 7 David A. Kofke Department of Chemical Engineering SUNY Buffalo
Gaussian process modelling
2015 AprilUNIVERSITY OF HAIFA, DEPARTMENT OF STATISTICS, SEMINAR FOR M.A 1 Hastie, Tibshirani and Friedman.The Elements of Statistical Learning (2nd edition,
Biointelligence Laboratory, Seoul National University
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Introduction to Linear Regression
Modern Navigation Thomas Herring
1 Advances in the Construction of Efficient Stated Choice Experimental Designs John Rose 1 Michiel Bliemer 1,2 1 The University of Sydney, Australia 2.
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 3: LINEAR MODELS FOR REGRESSION.
Assumptions of value-added models for estimating school effects sean f reardon stephen w raudenbush april, 2008.
Managerial Economics Demand Estimation & Forecasting.
Copyright © 2005 by the McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Managerial Economics Thomas Maurice eighth edition Chapter 4.
Inference for Regression Simple Linear Regression IPS Chapter 10.1 © 2009 W.H. Freeman and Company.
1 Chapter Two: Sampling Methods §know the reasons of sampling §use the table of random numbers §perform Simple Random, Systematic, Stratified, Cluster,
1 E. Fatemizadeh Statistical Pattern Recognition.
1 Pattern Recognition: Statistical and Neural Lonnie C. Ludeman Lecture 24 Nov 2, 2005 Nanjing University of Science & Technology.
1 Chapter 7 Sampling Distributions. 2 Chapter Outline  Selecting A Sample  Point Estimation  Introduction to Sampling Distributions  Sampling Distribution.
Colorado Center for Astrodynamics Research The University of Colorado 1 STATISTICAL ORBIT DETERMINATION ASEN 5070 LECTURE 11 9/16,18/09.
Ronan McNulty EWWG A general methodology for updating PDF sets with LHC data Francesco de Lorenzi*, Ronan McNulty (University College Dublin)
Issues concerning the interpretation of statistical significance tests.
Question paper 1997.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Chapter1: Introduction Chapter2: Overview of Supervised Learning
Chapter 20 Classification and Estimation Classification – Feature selection Good feature have four characteristics: –Discrimination. Features.
פרקים נבחרים בפיסיקת החלקיקים אבנר סופר אביב
Autoregressive (AR) Spectral Estimation
Spectrum Sensing In Cognitive Radio Networks
Machine Learning 5. Parametric Methods.
Treatment of correlated systematic errors PDF4LHC August 2009 A M Cooper-Sarkar Systematic differences combining ZEUS and H1 data  In a QCD fit  In a.
The KOSMOSHOWS What is it ? The statistic inside What it can do ? Future development Demonstration A. Tilquin (CPPM)
Pattern recognition – basic concepts. Sample input attribute, attribute, feature, input variable, independent variable (atribut, rys, příznak, vstupní.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 7 Inferences Concerning Means.
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
Computational Intelligence: Methods and Applications Lecture 14 Bias-variance tradeoff – model selection. Włodzisław Duch Dept. of Informatics, UMK Google:
Chapter 4: Basic Estimation Techniques
Making inferences from collected data involve two possible tasks:
LECTURE 09: BAYESIAN ESTIMATION (Cont.)
Basic Estimation Techniques
Ch3: Model Building through Regression
CH 5: Multivariate Methods
Basic Estimation Techniques
Collaborative Filtering Matrix Factorization Approach
HESSIAN vs OFFSET method
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Simple Linear Regression
Generally Discriminant Analysis
Model generalization Brief summary of methods
Presentation transcript:

A Neural Network MonteCarlo approach to nucleon Form Factors parametrization Paris, ° CLAS12 Europen Workshop In collaboration with: A. Bacchetta – University of Pavia M. Guagnelli – INFN Pavia M. Guagnelli – INFN Pavia J. Rojo – INFN Milano J. Rojo – INFN Milano

Drawbacks  A big value of the final χ 2 (indicating a bad fit) does not directly translates into larger parameter errors, only driven by the form of the Hessian at the minimum estimates of the parameter errors, only driven by the form of the Hessian at the minimum  The standard statistical prescription Δχ 2 =1 often yields unrealistically small errors, due to incompatible data setssystematics and theoretical uncertainties the incompatible data sets or to the presence of systematics and theoretical uncertainties linear error propagation  The validity of linear error propagation is assumed heavily depend on the functional form  Error estimates heavily depend on the functional form chosen for the parametrizations  Error propagation from data to model parameters and to parameters to observables in not trivial! Standard Hessian method  the covariance matrix determines both parameter errors (diagonal elements, including correlations) and error propagation to generic observables

 The theoretical bias introduced by the specific functional form adopted to fit the exp. data is difficult to assess, but it may have significant impact in applications Caveats  Simple, physics-inspired functions (based on theoretical constraints at small and large Q 2, large model-dependence etc.) imply a large model-dependence for the corresponding predictions and error estimates the behaviour in the extrapolation regions is strictly determined by the choice of  In particular, the behaviour in the extrapolation regions is strictly determined by the choice of the model function the model function, and does not fully reflect the present degree of ignorance in those ranges Lagrange Multiplier method overcomes the linear and quadratic approximations, but still needs non-standard Δχ 2 tolerance criteria and requires a full refitting each time uncertainties on a different observable are wanted

 MonteCarlo approach: perform global fits on an ensemble of N rep artificial replicas of the original data, obtained by means of importance-sampled MC random generation it does not rely on linear error propagation it does not rely on linear error propagation  test the implications of non-gaussian distributions of the experimental data and of the fitted model functions a MC sampling of the probability measure in the function space of FFs it provides a MC sampling of the probability measure in the function space of FFs computationally demanding  computationally demanding (N rep =10 for central values, 100 for errors and 1000 for correlations) Expectation values of observables depending on FFs: functional integrals over function space approximated by ensemble averages over the set of N rep best fit model replicas

 Neural Networks: a class of particularly flexible non-linear statistical data modeling tools, used to describe complex relationships and find patterns in data unbiased parametrization of the by choosing a stable and sufficiently redundant architecture  unbiased parametrization of the data data, i.e. independent on model assumptions and on the particular functional form adopted extrapolation regions behaviour in extrapolation regions NOT driven by the functions shape in the data regions supervised learning training algorithms many efficient supervised learning training algorithms available in literature overlearning risk  overlearning risk must be properly tamed  regularization

2. MC generation of artificial data replicas 3. Neural Network fits to MC replicas, minimizing the error function w.r.t. weights-thresholds 1. Experimental data set (electric-magnetic p-n FFs) 4. Ensemble of best-fit model functions. Everything (central values, error bands, error propagation) descending from sample statistical estimators

 Spreading of the best fit curves Nrep=200, M=4, constraints at Q 2 =0 hard-wired in the model fuctions

 1σ and 68% CL error bands: Main differences in extrapolation regions, pointing at a NON-gaussian distribution of best fit functions

Neural MonteCarlo approach  Neural MonteCarlo approach provides a powerful tool to parametrize form factors world data in a statistically rigorous way an unbiased global fit  It ensures an unbiased global fit, independent of the adopted functional form, thanks to NN redundancy statistical  Errors estimation and propagation is simply based on the statistical features of the best fit ensemble features of the best fit ensemble; no approximation is needed assess and include the effect of new data  It is possible to assess and include the effect of new data through a Bayesian reweighting, without the need to perform a full re-fit Many possible applications  Many possible applications: useful for observables highly sensitive to FFs