Maximum Entropy and Fourier Transformation Nicole Rogers.

Slides:



Advertisements
Similar presentations
Chapter 17 Multivariable Calculus.
Advertisements

©Gioko ®2007 Uncertainties in calculated results.
Grand Canonical Ensemble and Criteria for Equilibrium
Modeling of Data. Basic Bayes theorem Bayes theorem relates the conditional probabilities of two events A, and B: A might be a hypothesis and B might.
The microcanonical ensemble Finding the probability distribution We consider an isolated system in the sense that the energy is a constant of motion. We.
AAEC 2305 Fundamentals of Ag Economics Chapter 2 Economics of Demand.
Differentiation.
Linear Inequalities and Linear Programming Chapter 5 Dr.Hayk Melikyan/ Department of Mathematics and CS/ Linear Programming in two dimensions:
CmpE 104 SOFTWARE STATISTICAL TOOLS & METHODS MEASURING & ESTIMATING SOFTWARE SIZE AND RESOURCE & SCHEDULE ESTIMATING.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Consumer Theory.

Visual Recognition Tutorial
1 Solver Finding maximum, minimum, or value by changing other cells Can add constraints Don’t need to “guess and check”
Constrained Maximization
1 Learning Entity Specific Models Stefan Niculescu Carnegie Mellon University November, 2003.
Entropy. Optimal Value Example The decimal number 563 costs 10  3 = 30 units. The binary number costs 2  10 = 20 units.  Same value as decimal.
Quality Control Procedures put into place to monitor the performance of a laboratory test with regard to accuracy and precision.
Economics 214 Lecture 37 Constrained Optimization.
In this chapter, look for the answers to these questions:
Lecture 38 Constrained Optimization
INTRODUCTORY MATHEMATICAL ANALYSIS For Business, Economics, and the Life and Social Sciences  2007 Pearson Education Asia Chapter 7 Linear Programming.
Partial Differentiation & Application
Nonlinear Filtering of Stochastic Volatility Aleksandar Zatezalo, University of Rijeka, Faculty of Philosophy in Rijeka, Rijeka, Croatia
The Statistical Interpretation of Entropy The aim of this lecture is to show that entropy can be interpreted in terms of the degree of randomness as originally.
Differential Equations 7. The Logistic Equation 7.5.
Probability (cont.). Assigning Probabilities A probability is a value between 0 and 1 and is written either as a fraction or as a proportion. For the.
THE MATHEMATICS OF OPTIMIZATION
Chapter 4 Systems of Linear Equations; Matrices
1 1 Slide LINEAR PROGRAMMING: THE GRAPHICAL METHOD n Linear Programming Problem n Properties of LPs n LP Solutions n Graphical Solution n Introduction.
AN INTRODUCTION TO MICROECONOMICS Dr. Mohammed Migdad.
Optimization Techniques Methods for maximizing or minimizing an objective function Examples –Consumers maximize utility by purchasing an optimal combination.
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
© 2010 Pearson Education Inc.Goldstein/Schneider/Lay/Asmar, CALCULUS AND ITS APPLICATIONS, 12e– Slide 1 of 62 Chapter 7 Functions of Several Variables.
1. Entropy as an Information Measure - Discrete variable definition Relationship to Code Length - Continuous Variable Differential Entropy 2. Maximum Entropy.
Introduction l Example: Suppose we measure the current (I) and resistance (R) of a resistor. u Ohm's law relates V and I: V = IR u If we know the uncertainties.
Determining Sample Size
Roman Keeney AGEC  In many situations, economic equations are not linear  We are usually relying on the fact that a linear equation.
Lecture # 2 Review Go over Homework Sets #1 & #2 Consumer Behavior APPLIED ECONOMICS FOR BUSINESS MANAGEMENT.
6.1 Chapter 7 – The Theory of Consumer Behavior  The Theory of Consumer behavior provides the theoretical basis for buyer decision- making and the foundation.
Lecture 12 Statistical Inference (Estimation) Point and Interval estimation By Aziza Munir.
ECONOMIC FOUNDATIONS OF FINANCE BASIC ELEMENTS OF THE THEORY OF CAPITAL.
Introduction A GENERAL MODEL OF SYSTEM OPTIMIZATION.
1 1 Slide Decision Theory Professor Ahmadi. 2 2 Slide Learning Objectives n Structuring the decision problem and decision trees n Types of decision making.
1 Lesson 8: Basic Monte Carlo integration We begin the 2 nd phase of our course: Study of general mathematics of MC We begin the 2 nd phase of our course:
Integrals  In Chapter 2, we used the tangent and velocity problems to introduce the derivative—the central idea in differential calculus.  In much the.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Managerial Economics Managerial Economics = economic theory + mathematical eco + statistical analysis.
Information theory (part II) From the form g(R) + g(S) = g(RS) one can expect that the function g( ) shall be a logarithm function. In a general format,
In Chapters 6 and 8, we will see how to use the integral to solve problems concerning:  Volumes  Lengths of curves  Population predictions  Cardiac.
Lecture 3: MLE, Bayes Learning, and Maximum Entropy
Economics 2301 Lecture 37 Constrained Optimization.
Copyright © Cengage Learning. All rights reserved. 9 Differential Equations.
The accuracy of averages We learned how to make inference from the sample to the population: Counting the percentages. Here we begin to learn how to make.
Money and Banking Lecture 11. Review of the Previous Lecture Application of Present Value Concept Internal Rate of Return Bond Pricing Real Vs Nominal.
This represents the most probable value of the measured variable. The more readings you take, the more accurate result you will get.
 This will explain how consumers allocate their income over many goods.  This looks at individual’s decision making when faced with limited income and.
WARM UP Solve: 1. 3x – 5 = (3x -5) = x – 3 + 4x = (2x – 4) = 6.
Bidding Strategies. Outline of Presentation Markup Expected Profit Cost of Construction Maximizing Expected Profit Case 1: Single Known Competitor Case.
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
Theoretical distributions: the Normal distribution.
Chapter 4 Systems of Linear Equations; Matrices
Lesson 8: Basic Monte Carlo integration
Chapter 4 Systems of Linear Equations; Matrices
Learning Objectives for Section 4.1
Chapter 4 Systems of Linear Equations; Matrices
Constrained Optimization
Constrained Optimization – Part 1
Chapter 7 Functions of Several Variables
Lecture 38 Constrained Optimization
Presentation transcript:

Maximum Entropy and Fourier Transformation Nicole Rogers

An Introduction to Entropy  Known as the ‘law of disorder.’  Entropy is a measurement of uncertainty associated with a random variable.  Measures the ‘multiplicity’ associated with the state of objects.

Thermodynamic Entropy  Thermodynamic entropy is related to Shannon entropy by normalizing it with a Boltzmann constant.

Shannon Entropy  Shannon Entropy measures how undetermined a state of uncertainty is.  The higher the Shannon Entropy, the more undetermined the system is.

Shannon Entropy Example RacersChance to Win (P) -log(P)-P log(P) Fido Ms Fluff Spike Woofers  Let’s use the example of a dog race.  Four dogs have various chances of winning the race. If we apply the entropy equation: H = ∑ P i log(P i )

Shannon Entropy Example (cont.) RacersChance to Win (P) -log(P)-P log(P) Fido Ms Fluff Spike Woofers H = ∑ P i log(P i ) H = The Shannon Entropy is 1.72

Things to Notice RacersChance to Win (P)-log(P)-P log(P) Fido Ms Fluff Spike Woofers  If you add the chance of each dog to win, the total will be one. This is because the chances are normalized and can me represented using a Gaussian curve.  The more uncertain a situation, the higher the Shannon entropy. This will be demonstrated in the next example.

Two Uncertain Examples RacersChance to Win (P) -log(P)-P log(P) Fido Ms Fluff Spike Woofers H = ∑ P i log(P i ) H = With every variable completely uncertain, the Shannon Entropy will be 2.0

Two Uncertain Examples RacersChance to Win (P) -log(P)-P log(P) Fido Ms Fluff Spike Woofers H = ∑ P i log(P i ) H = With the situation fairly certain, the Shannon Entropy will be 0.25.

Comparisons to Draw  High Uncertainty H = 2.00  Fair UncertaintyH = 1.72  Low UncertaintyH = 0.25  The more uncertain the situation, the higher the entropy, thus entropy is a measurement of chaos.

Maximum Entropy  The maximum entropy states that, subject to precisely stated prior data, which must be a proposition that expresses testable information, the probability distribution which best represents the current state of knowledge is one with the largest information theoretical entropy.  In most practical cases, the stated prior data or testable information is given by a set of conserved quantities associated with the probability distribution is question. We use Lagrange method to help us solve this.

Lagrange Multiplier  In mathematical optimization, the method of Lagrange multipliers provides a strategy for finding the local maxima and minima of a function subject to equality constraints.  Lagrange Method assumes maximum entropy. The first of these equations are a normalization constraint. All of the probabilities must equal 1. The second equation is a general constraint. We will see more of what this is in the next example.

Lagrange Multiplier  Since Lagrange Method assumes maximum entropy, we can say:  Maximizing L with respect to each of the p(A i ) is done by differentiating L with respect to one of the p(A i ) while keeping α, β, and all other p(A i ) constant. The result is:

Lagrange Multiplier  Rearranging the equation, we can get: Where f( β )=0 because. Using this method, we can solve equations with minimum constraints.

Fast Food Frenzy BurgerChickenFishTofu $1.00$2.00$3.00$8.00 A fast food restaurant sells four types of product. They find that the average amount of money made for each purchase is $2.50. The products are chosen by the consumer based on price alone, and not preference. What is the percentage of purchase for each of these four foods?

Fast Food Frenzy  We know that: Applying Lagrange Method:

Fast Food Frenzy Entropy is the largest, subject to the constraints, if Where

Fast Food Frenzy  A zero-finding program was used to find the variables in these equations. The results were: = This rounds to one, and therefore is normalized. Lagrange method and maximum entropy can determine probabilities using only a small set of constraints. This answer makes sense because the probabilities of each food being chosen are consistent with the price constraint given to them. FoodProbability of Purchase Burger Chicken Fish Tofu0.1011

Remarks  Only by assuming maximum entropy are we able to evaluate these equations.  Since this example is evaluated on price alone, then the burger would have been chosen with the most frequency because of the cheaper price. The probabilities are lower for the more expensive prices, as indicated by the results.  When the number of randomness increases, so does the entropy. Because we only had four variables, the entropy at maximum would have been lower than if there were five variables. BurgerChickenFishTofu $1.00$2.00$3.00$8.00

Fourier Transformation  Fourier transform is a mathematical operation with many applications in physics and engineering that expresses a mathematical function of time as a function of frequency. The frequency can be approximated with sine and cosine functions.  Fourier transforms and maximum entropy can both be utilized to find the specific frequencies of a sine/cosine wave.

Fourier vs. Max Entropy  Num=30  Num=90  Num=150 x(i)=dsin(twopi*2.d0*t) x(i)=x(i)+dsin(twopi*3.d0*t) x(i)=5.d0+x(i)+dsin(twopi*3.2d0*t)

Fourier vs. Max Entropy  Since we were looking for 2.0 π, 3.0 π, and 3.2 π in our sine and cosine waves, maximum entropy was consistently better at determining these numbers on the graphs  Maximum entropy works better than Fourier from the range of 30 to 150 data sets. This is because it calculates an average using a small amount of data. If the data were dramatically increased, Fourier Method would work better.

Sources    modynamics modynamics 