Aims: - evaluate typical properties in controlled model situations - gain general insights into machine learning problems - compare algorithms in controlled.

Slides:



Advertisements
Similar presentations
Capacity of MIMO Channels: Asymptotic Evaluation Under Correlated Fading Presented by: Zhou Yuan University of Houston 10/22/2009.
Advertisements

Let X 1, X 2,..., X n be a set of independent random variables having a common distribution, and let E[ X i ] = . then, with probability 1 Strong law.
Evaluating Classifiers
Rutgers CS440, Fall 2003 Review session. Rutgers CS440, Fall 2003 Topics Final will cover the following topics (after midterm): 1.Uncertainty & introduction.
Hydrologic Statistics
Dynamics of Learning VQ and Neural Gas Aree Witoelar, Michael Biehl Mathematics and Computing Science University of Groningen, Netherlands in collaboration.
3) Vector Quantization (VQ) and Learning Vector Quantization (LVQ)
10 Further Time Series OLS Issues Chapter 10 covered OLS properties for finite (small) sample time series data -If our Chapter 10 assumptions fail, we.
ELEC 303 – Random Signals Lecture 18 – Statistics, Confidence Intervals Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 10, 2009.
A gentle introduction to fluid and diffusion limits for queues Presented by: Varun Gupta April 12, 2006.
SUMS OF RANDOM VARIABLES Changfei Chen. Sums of Random Variables Let be a sequence of random variables, and let be their sum:
Time Series Basics Fin250f: Lecture 3.1 Fall 2005 Reading: Taylor, chapter
A gentle introduction to Gaussian distribution. Review Random variable Coin flip experiment X = 0X = 1 X: Random variable.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Evaluating Hypotheses
Efficient Training in high-dimensional weight space Theoretische Physik und Astrophysik Computational Physics Julius-Maximilians-Universität Würzburg.
Simulation Models as a Research Method Professor Alexander Settles.
Evaluating Classifiers Lecture 2 Instructor: Max Welling.
The moment generating function of random variable X is given by Moment generating function.
Random Variables and Probability Distributions
The Lognormal Distribution
Approximations to Probability Distributions: Limit Theorems.
1 10. Joint Moments and Joint Characteristic Functions Following section 6, in this section we shall introduce various parameters to compactly represent.
Continuous Probability Distribution  A continuous random variables (RV) has infinitely many possible outcomes  Probability is conveyed for a range of.
Radial Basis Function Networks
Sampling Distributions  A statistic is random in value … it changes from sample to sample.  The probability distribution of a statistic is called a sampling.
Review of Probability.
Probability Theory and Random Processes
FLUCTATION SCALING: TAYLOR’S LAW AND BEYOND János Kertész Budapest University of Technology and Economics.
Applications of Bayesian sensitivity and uncertainty analysis to the statistical analysis of computer simulators for carbon dynamics Marc Kennedy Clive.
Natural Gradient Works Efficiently in Learning S Amari (Fri) Computational Modeling of Intelligence Summarized by Joon Shik Kim.
Copyright ©2011 Nelson Education Limited The Normal Probability Distribution CHAPTER 6.
1 Lesson 3: Choosing from distributions Theory: LLN and Central Limit Theorem Theory: LLN and Central Limit Theorem Choosing from distributions Choosing.
MA-250 Probability and Statistics Nazar Khan PUCIT Lecture 26.
Lab 3b: Distribution of the mean
Week11 Parameter, Statistic and Random Samples A parameter is a number that describes the population. It is a fixed number, but in practice we do not know.
Chapter 7 Sampling and Sampling Distributions ©. Simple Random Sample simple random sample Suppose that we want to select a sample of n objects from a.
ELEC 303 – Random Signals Lecture 18 – Classical Statistical Inference, Dr. Farinaz Koushanfar ECE Dept., Rice University Nov 4, 2010.
Optimal Bayes Classification
Week 21 Stochastic Process - Introduction Stochastic processes are processes that proceed randomly in time. Rather than consider fixed random variables.
Dynamical Analysis of LVQ type algorithms, WSOM 2005 Dynamical analysis of LVQ type learning rules Rijksuniversiteit Groningen Mathematics and Computing.
7 sum of RVs. 7-1: variance of Z Find the variance of Z = X+Y by using Var(X), Var(Y), and Cov(X,Y)
BCS547 Neural Decoding. Population Code Tuning CurvesPattern of activity (r) Direction (deg) Activity
ECE 8443 – Pattern Recognition ECE 8527 – Introduction to Machine Learning and Pattern Recognition Objectives: Elements of a Discrete Model Evaluation.
Machine Learning 5. Parametric Methods.
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM )
Geology 6600/7600 Signal Analysis 09 Sep 2015 © A.R. Lowry 2015 Last time: Signal Analysis is a set of tools used to extract information from sequences.
One Function of Two Random Variables
The Unscented Particle Filter 2000/09/29 이 시은. Introduction Filtering –estimate the states(parameters or hidden variable) as a set of observations becomes.
Spatial Point Processes Eric Feigelson Institut d’Astrophysique April 2014.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Giansalvo EXIN Cirrincione unit #4 Single-layer networks They directly compute linear discriminant functions using the TS without need of determining.
Oliver Schulte Machine Learning 726
Large Sample Theory EC 532 Burak Saltoğlu.
Fall 2004 Perceptron CS478 - Machine Learning.
Probability Theory and Parameter Estimation I
Probability for Machine Learning
Appendix A: Probability Theory
Outline Introduction Signal, random variable, random process and spectra Analog modulation Analog to digital conversion Digital transmission through baseband.
Ch3: Model Building through Regression
Chapter 7: Sampling Distributions
Overview of Supervised Learning
Lecture 10: Observers and Kalman Filters
Random Sampling Population Random sample: Statistics Point estimate
Large Sample Theory EC 532 Burak Saltoğlu.
Hidden Markov Models Part 2: Algorithms
C14: The central limit theorem
Hydrologic Statistics
Lecture 2 – Monte Carlo method in finance
Central Limit Theorem: Sampling Distribution.
Presentation transcript:

aims: - evaluate typical properties in controlled model situations - gain general insights into machine learning problems - compare algorithms in controlled environments - optimize and develop novel training algorithm Theory of Learning - analysis of model scenarios specify… - learning problem and student complexity (student / teacher scenarios) - statistics of observed data - cost function and/or training algorithm complements other approaches - e.g. “assumption free” bounds on generalization behavior in VC-Theory etc.

essential ingredients: - large systems with many adaptive parameters - high-dimensional input data - perform average over stochastic training procedures - perform average over randomized data set Statistical Physics ? history: - equilibrium/dynamics of recurrent networks (Hopfield 1982) - physics of interactions between neurons (Gardner 1988) - learning a rule with a perceptron (Vallet 1989) - on-line learning dynamics (Kinzel/Rujan 1990) → obtain typical (average behavior of large systems in controlled model scenarios

The Central Limit Theorem (in its simplest form) consider: independent identically distributed (i.i.d.) random numbers x i The sum of many random numbers is a Gaussian with mean and variance The sum becomes for M→∞ a Gaussian random quantity with density

possible extensions: - non-zero mean values (trivial) - i-dependent variance (condition: finite, same order of magnitude) - sums of weakly correlated random numbers most important point: - details of the statistics of x i are irrelevant - could be Gaussians themselves, binary x i =±1, uniform in [-a,a] … CLT applies effectively already very quite small M (“>12”) important in the following: correlated sums, e.g. with given coefficients a i,b i