Supplementary Text 2. If the relationship between two random variables X and Y is approximately illustrated in the following figure, then how can we discretize.

Slides:



Advertisements
Similar presentations
Original Figures for "Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring"
Advertisements

Fig. 4-1, p Fig. 4-2, p. 109 Fig. 4-3, p. 110.
Histogram Specification
Chapter 6 The Normal Distribution In this handout: Probability model for a continuous random variable Normal distribution.
Use of moment generating functions. Definition Let X denote a random variable with probability density function f(x) if continuous (probability mass function.
P.464. Table 13-1, p.465 Fig. 13-1, p.466 Fig. 13-2, p.467.
Fig. 11-1, p p. 360 Fig. 11-2, p. 361 Fig. 11-3, p. 361.
Probability & Statistical Inference Lecture 8 MSc in Computing (Data Analytics)
Table 6-1, p Fig. 6-1, p. 162 p. 163 Fig. 6-2, p. 164.
Variance and Standard Deviation The Expected Value of a random variable gives the average value of the distribution The Standard Deviation shows how spread.
CSE 221: Probabilistic Analysis of Computer Systems Topics covered: Continuous random variables Uniform and Normal distribution (Sec. 3.1, )
Continuous Random Variables and Probability Distributions
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Lecture II-2: Probability Review
Chapter 5: z-Scores. 5.1 Purpose of z-Scores Identify and describe location of every score in the distribution Take different distributions and make them.
Pairs of Random Variables Random Process. Introduction  In this lecture you will study:  Joint pmf, cdf, and pdf  Joint moments  The degree of “correlation”
Jeopardy Hypothesis Testing T-test Basics T for Indep. Samples Z-scores Probability $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
L7.1b Continuous Random Variables CONTINUOUS RANDOM VARIABLES NORMAL DISTRIBUTIONS AD PROBABILITY DISTRIBUTIONS.
Continuous Random Variables
Chapter 14 Monte Carlo Simulation Introduction Find several parameters Parameter follow the specific probability distribution Generate parameter.
Functions of Random Variables. Methods for determining the distribution of functions of Random Variables 1.Distribution function method 2.Moment generating.
Copyright © 2010, 2007, 2004 Pearson Education, Inc. All Rights Reserved. Section 6-1 Review and Preview.
Stats/Methods I JEOPARDY. Jeopardy CorrelationRegressionZ-ScoresProbabilitySurprise $100 $200$200 $300 $500 $400 $300 $400 $300 $400 $500 $400.
1 1.5 © 2016 Pearson Education, Inc. Linear Equations in Linear Algebra SOLUTION SETS OF LINEAR SYSTEMS.
Modular 11 Ch 7.1 to 7.2 Part I. Ch 7.1 Uniform and Normal Distribution Recall: Discrete random variable probability distribution For a continued random.
Basics of Functions.
5.3 Random Variables  Random Variable  Discrete Random Variables  Continuous Random Variables  Normal Distributions as Probability Distributions 1.
Scenario 6 Distinguishing different types of leukemia to target treatment.
P ROBABLITY S TATICS &. PROJECT. 1 Assuming that the error terms are distributed as: Please derive the maximum likelihood estimator for the simple linear.
Parametric Equations. You throw a ball from a height of 6 feet, with an initial velocity of 90 feet per second and at an angle of 40º with the horizontal.
1 Two Functions of Two Random Variables In the spirit of the previous lecture, let us look at an immediate generalization: Suppose X and Y are two random.
Chapter 7 Probability and Samples: The Distribution of Sample Means
BPS - 5th Ed. Chapter 31 The Normal Distributions.
The Purpose of Statistics (Data Analysis)
Discrete and Continuous Random Variables. Yesterday we calculated the mean number of goals for a randomly selected team in a randomly selected game.
Continuous Random Variables and Probability Distributions
Chapter 7 The Normal Probability Distribution 7.1 Properties of the Normal Distribution.
Tables 1 to 8 FM Run around!.
MATH Section 3.1.
Central Limit Theorem Let X 1, X 2, …, X n be n independent, identically distributed random variables with mean  and standard deviation . For large n:
INTRODUCTION TO z-SCORES  The purpose of z-scores, or standard scores, is to identify and describe the exact location of every score in a distribution.
PROBABILITY. OVERVIEW Relationships between samples and populations most often are described in terms of probability. Relationships between samples and.
Probability Distribution. Probability Distributions: Overview To understand probability distributions, it is important to understand variables and random.
Section 2 Standard Units and Areas under the Standard Normal Distribution.
Lesson 17: Sampling Variability Students use data from a random sample to estimate a population mean. Students understand the term “sampling variability”
Copyright © Cengage Learning. All rights reserved. 8 PROBABILITY DISTRIBUTIONS AND STATISTICS.
Analysis of Economic Data
5.3 The Central Limit Theorem
The Normal Distribution
Continuous Random Variables
Normal Distribution.
The Normal Probability Distribution
5.3 The Central Limit Theorem
Which One Doesn’t Belong?
Econometric Models The most basic econometric model consists of a relationship between two variables which is disturbed by a random error. We need to use.
Lecture Slides Elementary Statistics Twelfth Edition
External Text Features
Descriptive Statistics
Two-way analysis of variance (ANOVA)
Normal Probability Distributions
Learning Targets Students will be able to: Compare linear, quadratic, and exponential models and given a set of data, decide which type of function models.
Chapter 14 Monte Carlo Simulation
10.4 Parametric Equations.
5.3 The Central Limit Theorem
Normal Probability Distributions
Normal Distribution The Bell Curve.
Objectives Compare linear, quadratic, and exponential models.
Lesson #5: Probability Distributions
Discrete Random Variables: Joint PMFs, Conditioning and Independence
Continuous Random Variables: Basics
Presentation transcript:

Supplementary Text 2

If the relationship between two random variables X and Y is approximately illustrated in the following figure, then how can we discretize the sample values of them? μ1μ1 μ2μ2 N(μ1, σ1)N(μ1, σ1) N(μ2, σ2)N(μ2, σ2) t 1,1 t 1,2 t 1,3 t 1,4 t 2,1 t 2,2 t 2,3 t 2,4 Fig. A Fig. B

To uniformly describe the relation ship between X and Y, we can map them to normal distribution. And the corresponding relationship would be change into that illustrated in the figure below. 0 0 N(0, 1) t1t1 t2t2 t3t3 t4t4 t1t1 t2t2 t3t3 t4t4 Fig. C Fig. D

The data transformation between Figure A and C is implemented with the Equation (14) in the paper. And the data transformation between Figure B and D is also implemented with Equation (14) in the paper. Then our goal is to find the best common partitioning points t 1, t 2, …, t k. The Purpose of Equation (15) in the paper is to determine the best common partitioning points for both variable X and Y, while simultaneously making both of them as evenly partitioned as possible. The structures of the variables shown in Figure B and D reveal the essence of the relationship between the two random variables of the same kind. This is of more importance when the number of random variables becomes large. It makes the partition not “casual” just in order to adapt to the data, some of which are just noise.

Example 3 The gene expression data is partially listed below. Suppose that Gene #1 and #2 have been divided in the same group G 1, while Gene #3, #4, and #5 are in the same group G 2. Time 1Time 2Time 16 Gene No. Array1 (Low) Array2 (High) Array3 (Low) Array4 (High) ……Array31 (Low) Array32 (High) 1D 1,1,L D 1,1,H D 1,2,L D 1,2,H ……D 1,16,L D 1,16,H 2D 2,1,L D 2,1,H D 2,2,L D 2,2,H ……D 2,16,L D 2,16,H 3D 3,1,L D 3,1,H D 3,2,L D 3,2,H ……D 3,16,L D 3,16,H 4D 4,1,L D 4,1,H D 4,2,L D 4,2,H ……D 4,16,L D 4,16,H 5D 5,1,L D 5,1,H D 5,2,L D 5,2,H ……D 5,16,L D 5,16,H

For the data in the table above, we define 15 vectors as follows. Time 1Time 2Time 16 Gene No. Array1 (Low) Array2 (High) Array3 (Low) Array4 (High) ……Array31 (Low) Array32 (High) 1D 1,1,L D 1,1,H D 1,2,L D 1,2,H ……D 1,16,L D 1,16,H 2D 2,1,L D 2,1,H D 2,2,L D 2,2,H ……D 2,16,L D 2,16,H 3D 3,1,L D 3,1,H D 3,2,L D 3,2,H ……D 3,16,L D 3,16,H 4D 4,1,L D 4,1,H D 4,2,L D 4,2,H ……D 4,16,L D 4,16,H 5D 5,1,L D 5,1,H D 5,2,L D 5,2,H ……D 5,16,L D 5,16,H

To compute the mutual information between the two groups G 1 and G 2 about low pressure treatment described in Equation (17) in the paper, we can use the following detailed equation in stead of (17):

To compute the mutual information between the two groups G 1 and G 2 about high pressure treatment described in Equation (18) in the paper, we can use the following detailed equation in stead of (18):

To compute the mutual information between the two groups G 1 and G 2 about both high pressure and the high pressure treatment described in Equation (19) in the paper, we can use the following detailed equation in stead of (19):

To compute the mutual information between the treatment low pressure and the treatment high pressure within the group G 1 described in Equation (20) in the paper, we can use the following detailed equation in stead of (20):

To compute the mutual information between the treatment low pressure and the treatment high pressure within the group G 2 described in Equation (20) in the paper, we can use the following detailed equation in stead of (20):