Human-centered Machine Learning

Slides:



Advertisements
Similar presentations
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Advertisements

Outline input analysis input analyzer of ARENA parameter estimation
Lecture 6 CS5661 Pairwise Sequence Analysis-V Relatedness –“Not just important, but everything” Modeling Alignment Scores –Coin Tosses –Unit Distributions.
Chapter 8 Random-Variate Generation Banks, Carson, Nelson & Nicol Discrete-Event System Simulation.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 10: The Bayesian way to fit models Geoffrey Hinton.
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
20 10 School of Electrical Engineering &Telecommunications UNSW UNSW 2. Laguerre Parameterised Hawkes Process To model spike train data,
Descriptive statistics Experiment  Data  Sample Statistics Sample mean Sample variance Normalize sample variance by N-1 Standard deviation goes as square-root.
Random-Variate Generation. Need for Random-Variates We, usually, model uncertainty and unpredictability with statistical distributions Thereby, in order.
Today Today: Chapter 9 Assignment: Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Discrete Event Simulation How to generate RV according to a specified distribution? geometric Poisson etc. Example of a DEVS: repair problem.
Making Sense Making Numbers Making Significance Ulf H. Olsson Professor of Statistics.
Descriptive statistics Experiment  Data  Sample Statistics Experiment  Data  Sample Statistics Sample mean Sample mean Sample variance Sample variance.
Today Today: Chapter 9 Assignment: 9.2, 9.4, 9.42 (Geo(p)=“geometric distribution”), 9-R9(a,b) Recommended Questions: 9.1, 9.8, 9.20, 9.23, 9.25.
Parametric modelling of cost data: some simulation evidence Andrew Briggs University of Oxford Richard Nixon MRC Biostatistics Unit, Cambridge Simon Dixon.
Analysis of Simulation Input.. Simulation Machine n Simulation can be considered as an Engine with input and output as follows: Simulation Engine Input.
Statistical analysis and modeling of neural data Lecture 4 Bijan Pesaran 17 Sept, 2007.
Space-time Modelling Using Differential Equations Alan E. Gelfand, ISDS, Duke University (with J. Duan and G. Puggioni)
Lecture 6 Data Collection and Parameter Estimation.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Statistical inference.
Gaussian Mixture Models and Expectation Maximization.
Likelihood probability of observing the data given a model with certain parameters Maximum Likelihood Estimation (MLE) –find the parameter combination.
 1  Outline  stages and topics in simulation  generation of random variates.
Stats for Engineers Lecture 9. Summary From Last Time Confidence Intervals for the mean t-tables Q Student t-distribution.
Modeling and Simulation CS 313
Combining Statistical Language Models via the Latent Maximum Entropy Principle Shaojum Wang, Dale Schuurmans, Fuchum Peng, Yunxin Zhao.
MEGN 537 – Probabilistic Biomechanics Ch.5 – Determining Distributions and Parameters from Observed Data Anthony J Petrella, PhD.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Selecting Input Probability Distribution. Simulation Machine Simulation can be considered as an Engine with input and output as follows: Simulation Engine.
Improved Cross Entropy Method For Estimation Presented by: Alex & Yanna.
Lecture 3: Statistics Review I Date: 9/3/02  Distributions  Likelihood  Hypothesis tests.
Intro. ANN & Fuzzy Systems Lecture 23 Clustering (4)
Machine Learning Saarland University, SS 2007 Holger Bast Max-Planck-Institut für Informatik Saarbrücken, Germany Lecture 8, Friday June 8 th, 2007 (introduction.
Chapter 8 Random-Variate Generation Banks, Carson, Nelson & Nicol Discrete-Event System Simulation.
Université d’Ottawa / University of Ottawa 2001 Bio 8100s Applied Multivariate Biostatistics L1a.1 Lecture 1a: Some basic statistical concepts l The use.
5. Maximum Likelihood –II Prof. Yuille. Stat 231. Fall 2004.
Week 41 How to find estimators? There are two main methods for finding estimators: 1) Method of moments. 2) The method of Maximum likelihood. Sometimes.
Stochastic Loss Reserving with the Collective Risk Model Glenn Meyers ISO Innovative Analytics Casualty Loss Reserving Seminar September 18, 2008.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
A Brief Maximum Entropy Tutorial Presenter: Davidson Date: 2009/02/04 Original Author: Adam Berger, 1996/07/05
CS Statistical Machine learning Lecture 25 Yuan (Alan) Qi Purdue CS Nov
1 BINARY CHOICE MODELS: LOGIT ANALYSIS The linear probability model may make the nonsense predictions that an event will occur with probability greater.
Bootstrapping James G. Anderson, Ph.D. Purdue University.
Stochasticity and Probability. A new approach to insight Pose question and think of the answer needed to answer it. Ask: How do the data arise? What is.
Modeling and Simulation CS 313
Manuel Gomez Rodriguez
Learning Deep Generative Models by Ruslan Salakhutdinov
Modeling and Simulation CS 313
Manuel Gomez Rodriguez
Improvement of Likelihood Model Testing
Maximum Likelihood Estimation
Generalization and adaptivity in stochastic convex optimization
Likelihood Ratio, Wald, and Lagrange Multiplier (Score) Tests
Simulation: Sensitivity, Bootstrap, and Power
Predictive Learning from Data
Probability & Statistics Probability Theory Mathematical Probability Models Event Relationships Distributions of Random Variables Continuous Random.
Discrete Event Simulation - 4
Modelling Input Data Chapter5.
Where did we stop? The Bayes decision rule guarantees an optimal classification… … But it requires the knowledge of P(ci|x) (or p(x|ci) and P(ci)) We.
Integration of sensory modalities
Biointelligence Laboratory, Seoul National University
Iterative resolution of multi-reads in multiple genomes
Parametric Methods Berlin Chen, 2005 References:
Mathematical Foundations of BME Reza Shadmehr
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
Human-centered Machine Learning
A Gentle Tutorial of the EM Algorithm and its Application to Parameter Estimation for Gaussian Mixture and Hidden Markov Models Jeff A. Bilmes International.
Human-centered Machine Learning
Presentation transcript:

Human-centered Machine Learning Sampling and parameter fitting with Hawkes Processes Human-centered Machine Learning http://courses.mpi-sws.org/hcml-ws18/

We will first sample and then fit. Recap: How to fit and why sample? Infer parameters (Learning) Sampling (Predicting) Raw Data Parametrize intensity Event times drawn from: Derive Log-likelihood: Helps with: Prediction Model Checking Sanity check Gaining Intuition Simulator Summary statistics Maximum likelihood estimation: We will first sample and then fit.

Recap: What are Hawkes processes? time History, Intensity of self-exciting (or Hawkes) process: Observations: Clustered (or bursty) occurrence of events Intensity is stochastic and history dependent

Recap: How to fit a Hawkes process? time Maximum likelihood The max. likelihood is jointly convex in and (use CVX!)

Recap: How to sample from a Hawkes process time Thinning procedure (similar to rejection sampling): Sample from Poisson process with intensity Inversion sampling Generate Keep sample with prob. Keep the sample if

Coding assignment overview Parameter fitting (Assignment 2) Sampler (Assignment 1) Outline of tutorial Sanity check (provided)

Sampling: Ogata’s algorithm Ogata’s method of thinning Drawing 1 sample with intensity Accepting it with prob

Careful about exploding Sampling: Achtung I Ogata’s method of thinning Careful about exploding intensities! Include a check on i. Ensure

Sampling: Achtung II Ogata’s method of thinning History up to but not including t.

Sampling: Achtung III Ogata’s method of thinning is not an actual sample!

Sampling: Evaluation python plot_hawkes.py 1.0 0.5 1.0 10 sampled-sequences.txt output-plot.png Theoretical value Empirical average Juxtaposed events from sequences

Sampling: Evaluation python plot_hawkes.py 1.0 0.5 1.0 10 sampled-sequences.txt output-plot.png Theoretical value Empirical average Poisson (incorrect) sampler Juxtaposed events from sequences

Sampling: Evaluation python plot_hawkes.py 1.0 0.5 1.0 10 sampled-sequences.txt output-plot.png Theoretical value Empirical average Deterministic (incorrect) sampler Juxtaposed events from sequences

Live coding Show baseline + sanity check Show Poisson + sanity check

Parameter fitting: Problem setting (Assignment 2) Sampler (Assignment 1) samples.txt samples-test.txt

Can we do it faster for exponential kernels? Parameter fitting: Method Maximum likelihood estimation: Can we do it faster for exponential kernels? Nested sum

Parameter fitting: Using cvxpy

Parameter fitting: Using cvxpy Change this to maximize the likelihood

Live coding Show baseline + desired output Evaluation via sampling

Happy coding and holidays! Questions? Drop me an e-mail at utkarshu@mpi-sws.org Skype: utkarsh.upadhyay