Lecture 19: Free Energies in Modern Computational Statistical Thermodynamics: WHAM and Related Methods Dr. Ronald M. Levy Statistical.

Slides:



Advertisements
Similar presentations
Pattern Recognition and Machine Learning
Advertisements

The Maximum Likelihood Method
The microcanonical ensemble Finding the probability distribution We consider an isolated system in the sense that the energy is a constant of motion. We.
Copyright (c) 2004 Brooks/Cole, a division of Thomson Learning, Inc. Chapter 6 Point Estimation.
Point Estimation Notes of STAT 6205 by Dr. Fan.
CHAPTER 8 More About Estimation. 8.1 Bayesian Estimation In this chapter we introduce the concepts related to estimation and begin this by considering.
ECE 8443 – Pattern Recognition LECTURE 05: MAXIMUM LIKELIHOOD ESTIMATION Objectives: Discrete Features Maximum Likelihood Resources: D.H.S: Chapter 3 (Part.
1 12. Principles of Parameter Estimation The purpose of this lecture is to illustrate the usefulness of the various concepts introduced and studied in.
Estimation  Samples are collected to estimate characteristics of the population of particular interest. Parameter – numerical characteristic of the population.
Fundamentals of Data Analysis Lecture 12 Methods of parametric estimation.
Chapter 3 Classical Statistics of Maxwell-Boltzmann
Modern Monte Carlo Methods: (2) Histogram Reweighting (3) Transition Matrix Monte Carlo Jian-Sheng Wang National University of Singapore.
Lecture 14: Advanced Conformational Sampling
Visual Recognition Tutorial
The Calculation of Enthalpy and Entropy Differences??? (Housekeeping Details for the Calculation of Free Energy Differences) first edition: p
Estimation A major purpose of statistics is to estimate some characteristics of a population. Take a sample from the population under study and Compute.
Maximum likelihood (ML) and likelihood ratio (LR) test
Resampling techniques Why resampling? Jacknife Cross-validation Bootstrap Examples of application of bootstrap.
Maximum likelihood Conditional distribution and likelihood Maximum likelihood estimations Information in the data and likelihood Observed and Fisher’s.
Statistical Inference Chapter 12/13. COMP 5340/6340 Statistical Inference2 Statistical Inference Given a sample of observations from a population, the.
Maximum likelihood (ML)
2. Point and interval estimation Introduction Properties of estimators Finite sample size Asymptotic properties Construction methods Method of moments.
Maximum likelihood (ML)
Principles of the Global Positioning System Lecture 10 Prof. Thomas Herring Room A;
Binary Variables (1) Coin flipping: heads=1, tails=0 Bernoulli Distribution.
ECE 8443 – Pattern Recognition LECTURE 06: MAXIMUM LIKELIHOOD AND BAYESIAN ESTIMATION Objectives: Bias in ML Estimates Bayesian Estimation Example Resources:
Prof. Dr. S. K. Bhattacharjee Department of Statistics University of Rajshahi.
Random Sampling, Point Estimation and Maximum Likelihood.
Lecture 2 : Canonical Ensemble and the Partition Function Dr. Ronald M. Levy Statistical Thermodynamics.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Deterministic vs. Random Maximum A Posteriori Maximum Likelihood Minimum.
1 CE 530 Molecular Simulation Lecture 6 David A. Kofke Department of Chemical Engineering SUNY Buffalo
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
The Ideal Monatomic Gas. Canonical ensemble: N, V, T 2.
SUPA Advanced Data Analysis Course, Jan 6th – 7th 2009 Advanced Data Analysis for the Physical Sciences Dr Martin Hendry Dept of Physics and Astronomy.
8. Selected Applications. Applications of Monte Carlo Method Structural and thermodynamic properties of matter [gas, liquid, solid, polymers, (bio)-macro-
Multiple Random Variables Two Discrete Random Variables –Joint pmf –Marginal pmf Two Continuous Random Variables –Joint Distribution (PDF) –Joint Density.
1 M.Sc. Project of Hanif Bayat Movahed The Phase Transitions of Semiflexible Hard Sphere Chain Liquids Supervisor: Prof. Don Sullivan.
Time-dependent Schrodinger Equation Numerical solution of the time-independent equation is straightforward constant energy solutions do not require us.
PROBABILITY AND STATISTICS FOR ENGINEERING Hossein Sameti Department of Computer Engineering Sharif University of Technology Principles of Parameter Estimation.
Chapter 3: Maximum-Likelihood Parameter Estimation l Introduction l Maximum-Likelihood Estimation l Multivariate Case: unknown , known  l Univariate.
Chapter 7 Point Estimation of Parameters. Learning Objectives Explain the general concepts of estimating Explain important properties of point estimators.
Molecular Modelling - Lecture 2 Techniques for Conformational Sampling Uses CHARMM force field Written in C++
Review of Probability. Important Topics 1 Random Variables and Probability Distributions 2 Expected Values, Mean, and Variance 3 Two Random Variables.
Interacting Molecules in a Dense Fluid
Lecture 9: Theory of Non-Covalent Binding Equilibria Dr. Ronald M. Levy Statistical Thermodynamics.
Lecture 20: Binding Free Energy Calculations Dr. Ronald M. Levy Statistical Thermodynamics.
Generalized van der Waals Partition Function
2006: Assoc. Prof. R. J. Reeves Gravitation 2.1 P113 Gravitation: Lecture 2 Gravitation near the earth Principle of Equivalence Gravitational Potential.
Statistics Sampling Distributions and Point Estimation of Parameters Contents, figures, and exercises come from the textbook: Applied Statistics and Probability.
1 CE 530 Molecular Simulation Lecture 21 Histogram Reweighting Methods David A. Kofke Department of Chemical Engineering SUNY Buffalo
ESTIMATION METHODS We know how to calculate confidence intervals for estimates of  and  2 Now, we need procedures to calculate  and  2, themselves.
G. Cowan Lectures on Statistical Data Analysis Lecture 9 page 1 Statistical Data Analysis: Lecture 9 1Probability, Bayes’ theorem 2Random variables and.
Statistical Mechanics and Multi-Scale Simulation Methods ChBE
Parameter Estimation. Statistics Probability specified inferred Steam engine pump “prediction” “estimation”
R. Kass/W03 P416 Lecture 5 l Suppose we are trying to measure the true value of some quantity (x T ). u We make repeated measurements of this quantity.
Lecture 14: Advanced Conformational Sampling Dr. Ronald M. Levy Statistical Thermodynamics.
Evaluating Hypotheses. Outline Empirically evaluating the accuracy of hypotheses is fundamental to machine learning – How well does this estimate its.
Conditional Expectation
Statistics 350 Lecture 2. Today Last Day: Section Today: Section 1.6 Homework #1: Chapter 1 Problems (page 33-38): 2, 5, 6, 7, 22, 26, 33, 34,
Fundamentals of Data Analysis Lecture 11 Methods of parametric estimation.
CWR 6536 Stochastic Subsurface Hydrology Optimal Estimation of Hydrologic Parameters.
12. Principles of Parameter Estimation
Overview of Molecular Dynamics Simulation Theory
8/7/2018 Statistical Thermodynamics
Lecture 17: Kinetics and Markov State Models
12. Reweighting Methods.
Lecture 17: Kinetics and Markov State Models
Lecture 7: Distribution functions in classical monatomic liquids
Computing and Statistical Data Analysis / Stat 7
12. Principles of Parameter Estimation
Presentation transcript:

Lecture 19: Free Energies in Modern Computational Statistical Thermodynamics: WHAM and Related Methods Dr. Ronald M. Levy Statistical Thermodynamics

Definitions Canonical ensemble: focus on free energy changes that do not involve changing the number of particles, or their velocities, nor volume and temperature. Can work with configurational partition function: : system parameters and/or constraints Or equival ently:

Example 1: Configurational Free Energy Changes controls a conformational constraint (distance between two particles, dihedral angle, etc.). Call the constrained property f(q): only conformations that satisfy the constraint f(q)= (within a tolerance of d are allowed: Now: So: integration of the constrained Z over all the possible constraints gives the unconstrained Z =1 Free energy work for imposing the constraint:

it's basically an average of a delta function over the unconstrained ensemble. The delta function is non-zero only exactly at. In actual numerical applications we would consider some finite interval  around a discrete set of i. Consider: Free energy for imposing constraint in interval  :

where: therefore: At the  ->0 continuous limit: probability density

The zero of the free energy is arbitrary: A ( ) is the potential of mean force along : discrete : continuous :

Lesson #1: configurational free energies can be computed by counting – that is by collecting histograms or probability densities over unrestricted trajectories. Probably most used method. Not necessarily the most efficient. Achievable range of free energies is limited: N samples, N-1 in bin 1 and one sample in bin 2 For N =10,000,  A max ~ 5 kcal/mol at room temperature But in practice needs many more samples than this minimum to achieve sufficient precision

Method #2: Biased Sampling “umbrella” potential Any thermodynamic observable can be “unbiased”: x

Example: Unbiased probability for being near x i : Works only if biased potential depends only on x

distribution with biasing potential 1 (histogram probabilities p 1i ) distribution with biasing potential 2 (histogram probabilities p 2i ) distribution with biasing potential 3 (histogram probabilities p 3i ) Unbiased distribution p° i w1w1 w2w2 w3w3 How do we “merge” data from multiple biased simulations? distribution with biasing potential 1 (histogram probabilities p 1i ) distribution with biasing potential 2 (histogram probabilities p 2i ) distribution with biasing potential 3 (histogram probabilities p 3i ) Unbiased distribution p° 1i w1w1 w2w2 w3w3 Unbiased distribution p° 2i Unbiased distribution p° 3i ? Any set of weights u si gives the right answer, but what is the best set of u si for a given finite sample size?

Multiple biased simulations to cover conformational space Biasing potentials w s (x) x

Weighted Histogram Analysis Method (WHAM): Optimal way of combining multiple sources of biased data unbiased probability bias of bin i in simulation s free energy factor Let's assume the unbiased probability is known, we can predict what would be the biased distribution with biased potential w s : biased probability Any set of weights u si gives the right answer, but what is the optimal (most accurate estimate) set of u si for a given finite sample size?

Likelihood of histogram at s given probabilities at s (multinomial distribution): In terms of unbiased probabilties: Joint likelihood of the histograms from all simulations:

Log likelihood: : Max likelihood principle: choose p i ° that maximize the likelihood of the observed histograms. Need to express f s in terms of p i ° :

Log likelihood: :

Thus (WHAM equations) : Solved by iteration until convergence. Compare with single simulation case derived earlier: WHAM gives both probabilities (PMFs) and state free energies Ferrenberg & Swendsen (1989) Kumar, Kollman et al. (1992) Bartels & Karplus (1997) Gallicchio, Levy et al. (2005)

What about those optimal combining weights we talked about? Substituting... Therefore... WHAM optimal combining weights A simulation makes a large contribution at bin i if: 1. It provides many samples (large N s ) 2. Its bias is small compared to its free energy relative to the unbiased state.

WHAM: getting unbiased averages Computing the average of an observable that depends only on x is straightforward: For a property that depends on some other coordinate y=y(q). Solve for p 0 (x,y) – no bias on y – and then integrate out x : where: From WHAM

Some WHAM Applications for PMFs Chekmarev, Ishida, & Levy (2004) J. Phys. Chem. B 108: PMF of Alanine Dipeptide

2D Potential. bias = temperature Gallicchio, Andrec, Felts & Levy (2005) J. Phys. Chem. B 109:

 -hairpin peptide. Bias = temperature Gallicchio, Andrec, Felts & Levy (2005) J. Phys. Chem. B 109:

The Concept of “WHAM weights” Consider the simple average of O(x) : Let's sum over individual samples rather than bins and set: Where: WHAM weight of sample k. Measures likelihood of encountering x k in unbiased simulation.

WHAM weights example: probability distribution Apply previous averaging formula to p° i That is the unbiased probability at x i is the sum of the WHAM weights of the sample belonging to that bin.

Latest development: No binning WHAM = MBAR WHAM equations: substitute Get: Sum over samples MBAR equation Solved iteratively to convergence to get the f 's Distributions obtained by binning the corresponding WHAM weights. Shirts & Chodera J. Chem. Phys. (2008). Tan, Gallicchio, Lapelosa, Levy JCTC (2012).