Test of Independence through Mutual Information

Slides:



Advertisements
Similar presentations
C1: Sigma Notation For Sequences Sigma is a Greek letter. Capital sigma looks like this: Σ In Maths this symbol is used to mean ‘sum of’: add together.
Advertisements

Unit 0, Pre-Course Math Review Session 0.2 More About Numbers
1 Multiple Regression Interpretation. 2 Correlation, Causation Think about a light switch and the light that is on the electrical circuit. If you and.
Copyright ©2006 Brooks/Cole, a division of Thomson Learning, Inc. More About Categorical Variables Chapter 15.
Fundamentals of Forensic DNA Typing Slides prepared by John M. Butler June 2009 Appendix 3 Probability and Statistics.
Binary number, Bits and Bytes and memory Sen Zhang.
Research Methods in MIS
Summation Notation. Terminology k –index of summation 1- lower limit n-upper limit.
Covariance and Correlation
Chapter 4 Probability Copyright © 2014 by The McGraw-Hill Companies, Inc. All rights reserved.McGraw-Hill/Irwin.
DATASET INTRODUCTION 1. Dataset: Urine 2 From Cleveland Clinic
McGraw-Hill/IrwinCopyright © 2009 by The McGraw-Hill Companies, Inc. All Rights Reserved. Chapter 9 Hypothesis Testing.
Chapter © 2015 Cengage Learning. All Rights Reserved. May not be scanned, copied or duplicated, or posted to a publicly accessible website, in whole or.
Chapter 15 Data Analysis: Testing for Significant Differences.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
CHAPTER 10 Sequences, Induction, & Probability Sequences & Summation Notation Objectives –Find particular terms of sequence from the general term.
CS161 Topic #16 1 Today in CS161 Lecture #16 Prepare for the Final Reviewing all Topics this term Variables If Statements Loops (do while, while, for)
Section Copyright © 2014, 2012, 2010 Pearson Education, Inc. Lecture Slides Elementary Statistics Twelfth Edition and the Triola Statistics Series.
Chapter 12 Introduction to Analysis of Variance PowerPoint Lecture Slides Essentials of Statistics for the Behavioral Sciences Eighth Edition by Frederick.
© 2013 Pearson Education, Inc. Active Learning Lecture Slides For use with Classroom Response Systems Introductory Statistics: Exploring the World through.
Lecture 6 Excess Notation. Excess 8 notation indicates that the value for zero is the bit pattern for 8, that is 1000 Excess 8 notation indicates that.
Probability Michael J. Watts
Copyright © 2011 by The McGraw-Hill Companies, Inc. All rights reserved. McGraw-Hill/Irwin Chapter 4 Probability.
Boolean Algebra and Computer Logic Mathematical Structures for Computer Science Chapter 7 Copyright © 2006 W.H. Freeman & Co.MSCS Slides Boolean Logic.
Digital logic COMP214  Lecture 2 Dr. Sarah M.Eljack Chapter 1 1.
Statistical Decision Making. Almost all problems in statistics can be formulated as a problem of making a decision. That is given some data observed from.
Chapter 12 Introduction to Analysis of Variance
Copyright © 2009 Pearson Education, Inc LEARNING GOAL Interpret and carry out hypothesis tests for independence of variables with data organized.
© 2012 Pearson Prentice Hall. All rights reserved. CHAPTER 3 Number Theory and the Real Number System.
Precedence Operators Error Types
Lecture #8 Thursday, September 15, 2016 Textbook: Section 4.4
The Effect of the 2016 Presidential Election on Humana Stock
Programming and Data Structure
Dr. Clincy Professor of CS
Digital Logic & Design Dr. Waseem Ikram Lecture 02.
Chapter 4 Probability.
Good morning! August 16, Good morning! August 16, 2017.
Number Theory and the Real Number System
A Level Computing Component 2
Scalars and Vectors.
Copyright 2012, 2008, 2004, 2000 Pearson Education, Inc.
Log Linear Modeling of Independence
Bayesian Networks HAP 823 George Mason University
Probability Calculus Farrokh Alemi Ph.D.
Chapter 9 Hypothesis Testing.
Number Theory and the Real Number System
Place Value, Names for Numbers, and Reading Tables
Calculating Product of Values in Same Column
Receiver Operating Curves
Stratification Matters: Analysis of 3 Variables
Saturday, August 06, 2016 Farrokh Alemi, PhD.
Test of Independence in 3 Variables
Digital Logic & Design Lecture 02.
Propagation Algorithm in Bayesian Networks
Linear Algebra Lecture 37.
Copyright © Cengage Learning. All rights reserved.
One-Way Analysis of Variance
Chapter 5: Control Structure
Wednesday, September 21, 2016 Farrokh Alemi, PhD.
Propagation Algorithm in Bayesian Networks
LESSON 5: PROBABILITY Outline Probability Events
Prepared by : Adeel Ahmad Updated by : Syed Ameen Quadri
William Stallings Computer Organization and Architecture 10th Edition
Xbar Chart By Farrokh Alemi Ph.D
Risk Adjusted P-chart Farrokh Alemi, Ph.D.
Type I and Type II Errors
Multiplying Fractions: Visual Fraction Models
Cost-Volume-Profit Relationships
§5.2, The Integers; Order of Operations
Lecture Slides Essentials of Statistics 5th Edition
Presentation transcript:

Test of Independence through Mutual Information Wednesday, September 21, 2016 Jee Vang Ph.D. This lecture focuses on possible ways variables can relate to each other. The lecture is based on slides prepared by Dr. Lin and modified by Dr. Alemi. It is also based on the treatment of the topic by Alan Agresti in the book on Categorical Data Analysis.

Something Else Mutual Information There are many different methods for assessing independence and mutual information is often used as a way to do so.

Mutual Information The mutual information between two variables, X and Y, is defined as weighted sum of the log of probability of both events divided by the product of the probability of each event.

Mutual Information Note how mutual information reduces from statements about probability to the counts of times events occur.

Mutual Information # times both x & y occur In this formula, N with index x and y indicates the number of times both x and y occur together.

Mutual Information # times both x & y occur # times y occurs N with index x indicates number of times x occurs. N with index y indicates number of times y occurs. # times x occurs

[0, +Infinity] Threshold ? Range For mutual information, a value closer to zero implies independence, and values farther away from zero implies dependence. The range of values for mutual information is from zero to plus infinity. If we are to use mutual information to determine independence or dependence, typically, a threshold is specified a priori as a cutoff to make such a determination. For example, we can say a mutual information smaller than 0.001 is interpreted as independence, and a value larger than 0.001 is interpreted as dependence. This threshold value must be tinkered with to gauge what is appropriate, and differs from dataset to dataset. Threshold ?

Notation for Example Notation 1 Notation 2 X = true, Y = true X = t, Y = t X = true, Y = false X = t, Y = f X = false, Y = true X = f, Y = t X = false, Y = false X = f, Y = f Now, imagine two variables, X and Y, are both binary variables, taking on the values { true, false }. Let’s represent these values as { t, f } to be more concise. There are four unique combinations of the values shown in the table below. Note the equivalent notations.

Components of Mutual Information If we compute the mutual information between X and Y, we need to compute these four values.

Calculate the Count Using SQL TRUE select count(*) as total from data where n1=TRUE and n3=TRUE 3791 FALSE 1276 6833 8100 - select count(*) as total from data 20000 select count(*) as total from data where n1=TRUE 5067 select count(*) as total from data where n1=FALSE 14933 select count(*) as total from data where n3=TRUE 10624 select count(*) as total from data where n3=FALSE 9376 If we store the data into a table called “data”, we can issue the following SQL queries to get the count we need to compute the quantities above.

Mutual Information I(N1;N3) = 0.0282 + -0.0172 + -0.0221 + 0.0257 = 0.0145 Now we can calculate the mutual information between N1 and N3. Keep in mind that higher numbers indicate more dependence.

Conditional Mutual Information The conditional mutual information between two variables, X and Y given Z, and denoted as I(X,Y|Z), is defined based on conditional probabilities .

Conditional Mutual Information As before, the conditional mutual information can be calculated from counts as well. In this calculation Z can be a set of events or variables.

Conditional Mutual Information [0, +Infinity] Like mutual information, the value of conditional mutual information is in the range 0 to +infinity, where values closer to zero implies conditional independence and values farther from zero implies conditional dependence. Again, we also have to specify a cutoff value a priori to separate conditional independence from conditional dependence. Threshold ?

Division by Zero Division by zero can happen. This situation can happen when number of cooccurence of z with x or y is zero. In these situations the inner sum is set to zero and contribution of this level of z and x and y are ignored.

Mutual Information This lecture has shown how mutual and conditional mutual information can be carried out to get insights into the relationship among variables.