STATISTICAL COMPLEXITY ANALYSIS Dr. Dmitry Nerukh Giorgos Karvounis.

Slides:



Advertisements
Similar presentations
Time averages and ensemble averages
Advertisements

Lecture 2: Basic Information Theory TSBK01 Image Coding and Data Compression Jörgen Ahlberg Div. of Sensor Technology Swedish Defence Research Agency (FOI)
Hidden Markov Models (1)  Brief review of discrete time finite Markov Chain  Hidden Markov Model  Examples of HMM in Bioinformatics  Estimations Basic.
Supervisory Control of Hybrid Systems Written by X. D. Koutsoukos et al. Presented by Wu, Jian 04/16/2002.
Information Theory EE322 Al-Sanie.
Entropy in the Quantum World Panagiotis Aleiferis EECS 598, Fall 2001.
Complexity and “Quasi- Intermittency” of Electromagnetic Waves in Regular Time-Varying Medium Alexander Nerukh, Nataliya Ruzhytska, Dmitry Nerukh Kharkov.
Protein Threading Zhanggroup Overview Background protein structure protein folding and designability Protein threading Current limitations.
Matlab Simulations of Markov Models Yu Meng Department of Computer Science and Engineering Southern Methodist University.
1 Lecture 8: Genetic Algorithms Contents : Miming nature The steps of the algorithm –Coosing parents –Reproduction –Mutation Deeper in GA –Stochastic Universal.
Hidden Markov Models I Biology 162 Computational Genetics Todd Vision 14 Sep 2004.
R C Ball, Physics Theory Group and Centre for Complexity Science University of Warwick R S MacKay, Maths M Diakonova, Physics&Complexity Emergence in Quantitative.
Simulation.
Mossbauer Spectroscopy in Biological Systems: Proceedings of a meeting held at Allerton House, Monticello, Illinois. Editors: J. T. P. DeBrunner and E.
Simulation Models as a Research Method Professor Alexander Settles.
Causal-State Splitting Reconstruction Ziba Rostamian CS 590 – Winter 2008.
Molecular Information Theory Niru Chennagiri Probability and Statistics Fall 2004 Dr. Michael Partensky.
Behaviour of velocities in protein folding events Aldo Rampioni, University of Groningen Leipzig, 17th May 2007.
Dynamics of Learning & Distributed Adaptation PI: James P. Crutchfield, Santa Fe Institute Second PI Meeting, April 2001, SFe Dynamics of Learning:
Descriptive Modelling: Simulation “Simulation is the process of designing a model of a real system and conducting experiments with this model for the purpose.
The Role of Specialization in LDPC Codes Jeremy Thorpe Pizza Meeting Talk 2/12/03.
IE 594 : Research Methodology – Discrete Event Simulation David S. Kim Spring 2009.
Ch 9 pages Lecture 18 – Quantization of energy.
The Role of Artificial Life, Cellular Automata and Emergence in the study of Artificial Intelligence Ognen Spiroski CITY Liberal Studies 2005.
Do proteins fold ? Robert Glen Dmitry Nerukh Rimini 2005.
1 Performance Evaluation of Computer Networks: Part II Objectives r Simulation Modeling r Classification of Simulation Modeling r Discrete-Event Simulation.
Room 2032 China Canada Winnipeg Manitoba.
FUNCTIONAL COMPONENTS AS A BASIS FOR COMPLEX SYSTEM DESCRIPTION: SOME EXAMPLES AND DISCUSSION. D. C. Mikulecky Professor Emeritus and Senior Fellow  Center.
Summary: Isolated Systems, Temperature, Free Energy Zhiyan Wei ES 241: Advanced Elasticity 5/20/2009.
EME: Information Theoretic views of emergence and self-organisation Continuing the search for useful definitions of emergence and self organisation.
Minimization v.s. Dyanmics A dynamics calculation alters the atomic positions in a step-wise fashion, analogous to energy minimization. However, the steps.
Boltzmann Machine (BM) (§6.4) Hopfield model + hidden nodes + simulated annealing BM Architecture –a set of visible nodes: nodes can be accessed from outside.
1 CE 530 Molecular Simulation Lecture 6 David A. Kofke Department of Chemical Engineering SUNY Buffalo
Algorithmic Information Theory and the Emergence of Order Entropy and replication Sean Devine victoria management school.
Using Bayesian Networks to Analyze Whole-Genome Expression Data Nir Friedman Iftach Nachman Dana Pe’er Institute of Computer Science, The Hebrew University.
CSCI 6900/4900 Special Topics in Computer Science Automata and Formal Grammars for Bioinformatics Bioinformatics problems sequence comparison pattern/structure.
Introduction to Lattice Simulations. Cellular Automata What are Cellular Automata or CA? A cellular automata is a discrete model used to study a range.
Conformational Entropy Entropy is an essential component in ΔG and must be considered in order to model many chemical processes, including protein folding,
Computational Biology BS123A/MB223 UC-Irvine Ray Luo, MBB, BS.
Coding Theory Efficient and Reliable Transfer of Information
FAT TAILS REFERENCES CONCLUSIONS SHANNON ENTROPY AND ADJUSTMENT OF PARAMETERS AN ADAPTIVE STOCHASTIC MODEL FOR RETURNS An adaptive stochastic model is.
Lecture V Probability theory. Lecture questions Classical definition of probability Frequency probability Discrete variable and probability distribution.
Vaida Bartkutė, Leonidas Sakalauskas
1 CONTEXT DEPENDENT CLASSIFICATION  Remember: Bayes rule  Here: The class to which a feature vector belongs depends on:  Its own value  The values.
ECE-7000: Nonlinear Dynamical Systems Overfitting and model costs Overfitting  The more free parameters a model has, the better it can be adapted.
Molecular Modelling - Lecture 2 Techniques for Conformational Sampling Uses CHARMM force field Written in C++
Simple Lattice Model for Fluids. Introduction In this chapter we borrow the idea of lattice structures, characteristic of crystals, and apply it to develop.
Hidden Markovian Model. Some Definitions Finite automation is defined by a set of states, and a set of transitions between states that are taken based.
Bioinformatics: Practical Application of Simulation and Data Mining Markov Modeling I Prof. Corey O’Hern Department of Mechanical Engineering Department.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
Sharpening Occam’s razor with Quantum Mechanics SISSA Journal Club Matteo Marcuzzi 8th April, 2011.
Classical Control in Quantum Programs Dominique Unruh IAKS, Universität Karlsruhe Founded by the European Project ProSecCo IST
Stochastic Processes and Transition Probabilities D Nagesh Kumar, IISc Water Resources Planning and Management: M6L5 Stochastic Optimization.
Basic Concepts of Information Theory A measure of uncertainty. Entropy. 1.
DEPARTMENT/SEMESTER ME VII Sem COURSE NAME Operation Research Manav Rachna College of Engg.
Rate Distortion Theory. Introduction The description of an arbitrary real number requires an infinite number of bits, so a finite representation of a.
Statistical Mechanics and Multi-Scale Simulation Methods ChBE
Definition of the Hidden Markov Model A Seminar Speech Recognition presentation A Seminar Speech Recognition presentation October 24 th 2002 Pieter Bas.
Lecture 14: Advanced Conformational Sampling Dr. Ronald M. Levy Statistical Thermodynamics.
Introduction to Entropy. Entropy A formal attempt to quantify randomness by relating it to the amount of heat “wasted” by a thermodynamic process. The.
OPERATING SYSTEMS CS 3502 Fall 2017
Ch9: Decision Trees 9.1 Introduction A decision tree:
Modeling Arithmetic, Computation, and Languages
On calibration of micro-crack model of thermally induced cracks through inverse analysis Dr Vladimir Buljak University of Belgrade, Faculty of Mechanical.
Quantum mechanics from classical statistics
Dynamics of Learning & Distributed Adaptation James P
CONTEXT DEPENDENT CLASSIFICATION
Boltzmann Machine (BM) (§6.4)
A measurable definition of Emergence in quantitative systems
Intelligent Systems (AI-2) Computer Science cpsc422, Lecture 3
Presentation transcript:

STATISTICAL COMPLEXITY ANALYSIS Dr. Dmitry Nerukh Giorgos Karvounis

What is Complexity? Many different definitions. A natural system, is converted into a formal system that our mind can manipulate and we have a model. “Complexity is the property of a real world system that is manifest in the inability of any one formalism being adequate to capture all its properties…”

WHY USE COMPLEXITY ANALYSIS (1) When a new state of matter emerges from a phase transition, …certain “pattern formation” takes on the this “newness” with respect to other structures. This process is defined as “intrinsic” emergence. There is an increase in intrinsic computational capability which can be capitalised and measured.

WHY USE COMPLEXITY ANALYSIS (2) Contemporary physics can measure order (e.g. temperature) or ideal randomness (e.g. entropy, thermodynamics). No tools to address problems of innovation, or the discovery of patterns Measuring the computational capabilities of the system is the only way to address such questions: …discovering and quantifying emergence, pattern, information process and memory in quantitative units. The term intrinsic computation defines the way the system stores information with respect to time, transmits it between internal degrees of freedom and makes use of it in order to produce future behaviour.

METHODOLOGY(1) Complexity estimates how sophisticated are the dynamical laws governing the time evolution of the system. We adopted the approach by Crutchfield et. al. termed “computational mechanics”. We implement ideas from both Shannon entropy and KC algorithmic complexity theories, measuring the size of the informational description of the process. This is a direct approach to reveal the symmetries possessed by any spatial patterns and to estimate the “minimum amount of memory required to reproduce any configuration ensemble”.

…BUT WE ARE MODELLERS… HOW DO WE DEAL WITH MODELS…?? …HOW DO WE DEAL WITH MODELS…??

METHODOLOGY(2) We can reconstruct an algorithmic machine (termed as “ε- machine”) that provides the means to build the statistically minimal “optimally predictive model ”. In order to build this machine, we need the smallest possible set of predictive states, the “causal states”. We can state that two predictive states are equivalent (~) if and only if they give rise to same future values in terms of conditional probabilities:

COMPUTATIONAL IMPLEMENTATION(1) The algorithm is based on the use of symbolic dynamics generated from symbols assigned to discrete time steps. The crucial part in the implementation of the methodology is converting a continuous real signal into a sequence of symbols i.e “ signal symbolization” of the molecular trajectory. The one – dimensional case is shown below:

CAUSAL STATE Consider the following sequence: bla.bla.bla.lab.lba.bla.bla.lab.bal.bla.alb.alb.bla.bla… Bla. BlaPASTFUTURE Alb. BlaPASTFUTURE Bla Alb Bla CAUSALstate S causal

E-machine An  -machine, the set of causal states and the probabilities of the transitions between them, provides a direct description of the patterns present in the system’s internal degrees of freedom.

FINITE STATISTICAL COMPLEXITY Finite Statistical Complexity can be defined as “the minimum amount of evolutionary information (or hidden memory) required to statistically reproduce a process.” It expresses the informational size of the distribution of the causal states as measured by the Shannon Entropy: Statistical Complexity is based on the assumption that randomness is statistically simple: an ideal random process has zero statistical complexity. Equally, simple periodic processes give low complexity values as well. Complex process is the one that lies between these extremes and is an amalgam of predictable and stochastic mechanisms.

STATISTICAL COMPLEXITY OF A ZWITTERION: a folding event We measured the statistical complexity of the dynamical trajectories of four significant atoms within a zwitterion. Attain insights regarding complexity and how this can be a useful tool to characterise or capture the folding event. Depending on the temperature of the simulation, the zwitterion adopts a stable folded conformation. Statistical Complexity Analysis of various atoms’ trajectories at the unfolded configuration and compare their values at the folded state.

COMPLEXITY ANALYSIS OF THE EXTENDED STATE At the extended state, there is no significant change on the complexity value, as the zwitterion remains as an extended chain, following basically the same pattern throughout the process.

COMPLEXITY ANALYSIS OF THE FOLDED STATE In the folding event, there is a considerable drop in the complexity value, assigned to the transitional stage. Afterwards, there is a sudden rise in the complexity, until all atoms reach the same value, assigned to the pattern of the folded state.

COMPLEXITY ANALYSIS(1)

COMPLEXITY ANALYSIS (2) The essentiality of complexity measurements is that we can distinguish those patterns in quantitative terms. Better insight to the mechanisms that underlie the formation of this structure and separate the more “ordered” regularities to those that are more “random”:

FUTURE WORK Further development of the algorithm in order to achieve a better representation of the  -machine. Apply Statistical Complexity Analysis to a larger system such as protein folding and polymers’ phase transitions.

ACKNOWLEDGMENTS For this work we are grateful to: Prof. R. Glen The Newton Trust and UNILEVER for their financial support.