Rahul Sharma, Aditya V. Nori, Alex Aiken Stanford MSR India Stanford.

Slides:



Advertisements
Similar presentations
Machine Learning: Intro and Supervised Classification
Advertisements

Termination Proofs from Tests
Modular and Verified Automatic Program Repair Francesco Logozzo, Thomas Ball RiSE - Microsoft Research Redmond.
CHAPTER 2: Supervised Learning. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 2 Learning a Class from Examples.
Hypothesis testing Another judgment method of sampling data.
Lecture (11,12) Parameter Estimation of PDF and Fitting a Distribution Function.
© Anvesh Komuravelli Spacer Automatic Abstraction in SMT-Based Unbounded Software Model Checking Anvesh Komuravelli Carnegie Mellon University Joint work.
Rahul Sharma, Saurabh Gupta, Bharath Hariharan, Alex Aiken, and Aditya Nori (Stanford, UC Berkeley, Microsoft Research India) Verification as Learning.
Model Assessment, Selection and Averaging
Using Statically Computed Invariants Inside the Predicate Abstraction and Refinement Loop Himanshu Jain Franjo Ivančić Aarti Gupta Ilya Shlyakhter Chao.
ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Computational Learning Theory
4-1 Statistical Inference The field of statistical inference consists of those methods used to make decisions or draw conclusions about a population.
Experimental Evaluation
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, Lecture Slides for.
Stat 321 – Day 24 Point Estimation (cont.). Lab 6 comments Parameter  Population Sample Probability Statistics A statistic is an unbiased estimator for.
Part I: Classification and Bayesian Learning
Classification and Prediction: Regression Analysis
Chapter 9 Title and Outline 1 9 Tests of Hypotheses for a Single Sample 9-1 Hypothesis Testing Statistical Hypotheses Tests of Statistical.
1 Doing Statistics for Business Doing Statistics for Business Data, Inference, and Decision Making Marilyn K. Pelosi Theresa M. Sandifer Chapter 11 Regression.
1 © Lecture note 3 Hypothesis Testing MAKE HYPOTHESIS ©
4-1 Statistical Inference The field of statistical inference consists of those methods used to make decisions or draw conclusions about a population.
1 CS 475/575 Slide Set 6 M. Overstreet Spring 2005.
Classification. An Example (from Pattern Classification by Duda & Hart & Stork – Second Edition, 2001)
Chap 20-1 Statistics for Business and Economics, 6e © 2007 Pearson Education, Inc. Chapter 20 Sampling: Additional Topics in Sampling Statistics for Business.
PARAMETRIC STATISTICAL INFERENCE
Evaluating Hypotheses Reading: Coursepack: Learning From Examples, Section 4 (pp )
Reasoning about Information Leakage and Adversarial Inference Matt Fredrikson 1.
Rahul Sharma Joint work with Aditya Nori (MSR India) and Alex Aiken (Stanford)
Aditya V. Nori, Sriram K. Rajamani Microsoft Research India.
Rahul Sharma, Eric Schkufza, Berkeley Churchill, Alex Aiken.
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
Confidence intervals and hypothesis testing Petter Mostad
Scientific Method. My 4 out of 5 Rule If you make an observation.
CROSS-VALIDATION AND MODEL SELECTION Many Slides are from: Dr. Thomas Jensen -Expedia.com and Prof. Olga Veksler - CS Learning and Computer Vision.
ISCG8025 Machine Learning for Intelligent Data and Information Processing Week 3 Practical Notes Application Advice *Courtesy of Associate Professor Andrew.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Christoph Eick: Learning Models to Predict and Classify 1 Learning from Examples Example of Learning from Examples  Classification: Is car x a family.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN © The MIT Press, Lecture.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Review of fundamental 1 Data mining in 1D: curve fitting by LLS Approximation-generalization tradeoff First homework assignment.
Bias-Variance in Machine Learning. Bias-Variance: Outline Underfitting/overfitting: –Why are complex hypotheses bad? Simple example of bias/variance Error.
Bias and Variance of the Estimator PRML 3.2 Ethem Chp. 4.
Data Mining and Decision Support
Machine Learning for Program Language Research Yao Peisen Prism Group, HKUST
CS 2750: Machine Learning The Bias-Variance Tradeoff Prof. Adriana Kovashka University of Pittsburgh January 13, 2016.
CS357 Lecture 13: Symbolic model checking without BDDs Alex Aiken David Dill 1.
Regression Variance-Bias Trade-off. Regression We need a regression function h(x) We need a loss function L(h(x),y) We have a true distribution p(x,y)
MACHINE LEARNING 3. Supervised Learning. Learning a Class from Examples Based on E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Information criteria What function fits best? The more free parameters a model has the higher will be R 2. The more parsimonious a model is the lesser.
Rely: Verifying Quantitative Reliability for Programs that Execute on Unreliable Hardware Michael Carbin, Sasa Misailovic, and Martin Rinard MIT CSAIL.
© Anvesh Komuravelli Spacer Model Checking with Proofs and Counterexamples Anvesh Komuravelli Carnegie Mellon University Joint work with Arie Gurfinkel,
Abstraction and Abstract Interpretation. Abstraction (a simplified view) Abstraction is an effective tool in verification Given a transition system, we.
Rahul Sharma and Alex Aiken (Stanford University) 1.
Bias-Variance Analysis in Regression  True function is y = f(x) +  where  is normally distributed with zero mean and standard deviation .  Given a.
15 Inferential Statistics.
Presentation Title 2/4/2018 Software Verification using Predicate Abstraction and Iterative Refinement: Part Bug Catching: Automated Program Verification.
More on Inference.
Learning Invariants using Decision Trees and Implication Counterexamples Pranav Garg Amazon India.
Beyond Strong vs. Weak Updates Isil Dillig, Thomas Dillig, Alex Aiken
Chapter 11: Learning Introduction
More on Inference.
Inferring Simple Solutions to Recursion-free Horn Clauses via Sampling
9 Tests of Hypotheses for a Single Sample CHAPTER OUTLINE
Hyperparameters, bias-variance tradeoff, validation
Discrete Event Simulation - 4
Over-Approximating Boolean Programs with Unbounded Thread Creation
Overfitting and Underfitting
CHAPTER 9 Testing a Claim
Presentation transcript:

Rahul Sharma, Aditya V. Nori, Alex Aiken Stanford MSR India Stanford

int i = 1, j = 0; while (i<=5) { j = j+i ; i = i+1; } Increasing precision D. Monniaux and J. L. Guen. Stratified static analysis based on variable dependencies. Electr. Notes Theor. Comput. Sci. 2012

A. V. Nori and S. K. Rajamani. An empirical study of optimizations in YOGI. ICSE (1) 2010

 Increased precision is causing worse results  Programs have unbounded behaviors  Program analysis  Analyze all behaviors  Run for a finite time  In finite time, observe only finite behaviors  Need to generalize

 Generalization is ubiquitous  Abstract interpretation: widening  CEGAR: interpolants  Parameter tuning of tools  Lot of folk knowledge, heuristics, …

 “It’s all about generalization”  Learn a function from observations  Hope that the function generalizes  Work on formalization of generalization

 Model the generalization process  Probably Approximately Correct (PAC) model  Explain known observations by this model  Use this model to obtain better tools

INTERPOLANTSCLASSIFIERS + Rahul Sharma, Aditya V. Nori, Alex Aiken: Interpolants as Classifiers. CAV

c

c

H For any arbitrary labeling

Precision is low Underfitting Precision is high Overfitting Good fit Y X

 Generalization error is bounded by sum of  Bias: Empirical error of best available hypothesis  Variance: O (VC-d) Bias Variance Increase precision Generalization error Possible hypotheses

int i = 1, j = 0; while (i<=5) { j = j+i ; i = i+1; }

 What goes wrong with excess precision?  Fit polyhedra to program behaviors  Transfer functions, join, widening  Too many polyhedra, make a wrong choice

J. Henry, D. Monniaux, and M. Moy. Pagai: A path sensitive static analyser. Electr. Notes Theor. Comput. Sci

A. V. Nori and S. K. Rajamani. An empirical study of optimizations in YOGI. ICSE (1) 2010

 Parameter tuning of program analyses  Overfitting? Generalization on new tasks? P. Godefroid, A. V. Nori, S. K. Rajamani, and S. Tetali. Compositional may-must program analysis: unleashing the power of alternation. POPL Benchmark Set (2490 verification tasks) Train

 How to set the test length in Yogi Benchmark Set (2490 verification tasks) Training Set (1743) Test Set (747) Train Test

 On 2106 new verification tasks  40% performance improvement!  Yogi in production suffers from overfitting

 Keep separate training and test sets  Design of the tools governed by training set  Test set as a check  SVCOMP: all benchmarks are public  Test tools on some new benchmarks too

R. Jhala and K. L. McMillan. A practical and complete approach to predicate refinement. TACAS  Suggests incrementally increasing precision  Find a sweet spot where generalization error is low

 No generalization -> no bias-variance tradeoff  Certain classes of type inference  Abstract interpretation without widening  Loop-free and recursion-free programs  Verify a particular program (e.g., seL4)  Overfit on the one important program

 A model to understand generalization  Bias-Variance tradeoffs  These tradeoffs do occur in program analysis  Understand these tradeoffs for better tools