Discovery of Conservation Laws via Matrix Search Oliver Schulte and Mark S. Drew School of Computing Science Simon Fraser University Vancouver, Canada.

Slides:



Advertisements
Similar presentations
Principal Component Analysis Based on L1-Norm Maximization Nojun Kwak IEEE Transactions on Pattern Analysis and Machine Intelligence, 2008.
Advertisements

Support Vector Machines
Machine learning continued Image source:
Computer vision: models, learning and inference Chapter 13 Image preprocessing and feature extraction.
Section 4.6 (Rank).
Patch to the Future: Unsupervised Visual Prediction
Ziming Zhang*, Ze-Nian Li, Mark Drew School of Computing Science Simon Fraser University Vancouver, Canada {zza27, li, AdaMKL: A Novel.
10/11/2001Random walks and spectral segmentation1 CSE 291 Fall 2001 Marina Meila and Jianbo Shi: Learning Segmentation by Random Walks/A Random Walks View.
Automated Search for Conserved Quantities in Particle Reactions Oliver Schulte School of Computing Science Simon Fraser University
Learning Conservation Principles in Particle Physics Oliver Schulte School of Computing Science Simon Fraser University
Energy Balance Analysis Reference Paper: Beard et al. Energy Balance for Analysis of Complex Metabolic Networks. Biophysical Journal 83, (2002) Presented.
Sampling algorithms for l 2 regression and applications Michael W. Mahoney Yahoo Research (Joint work with P. Drineas.
Principal Component Analysis
The value of kernel function represents the inner product of two training points in feature space Kernel functions merge two steps 1. map input data from.
Reduced Support Vector Machine
Gradient Methods May Preview Background Steepest Descent Conjugate Gradient.
Independent Component Analysis (ICA) and Factor Analysis (FA)
Support Vector Machines and Kernel Methods
Clustering In Large Graphs And Matrices Petros Drineas, Alan Frieze, Ravi Kannan, Santosh Vempala, V. Vinay Presented by Eric Anderson.
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Adapted by Doug Downey from Machine Learning EECS 349, Bryan Pardo Machine Learning Clustering.
Lecture outline Support vector machines. Support Vector Machines Find a linear hyperplane (decision boundary) that will separate the data.
Tutorial 10 Iterative Methods and Matrix Norms. 2 In an iterative process, the k+1 step is defined via: Iterative processes Eigenvector decomposition.
An Introduction to Support Vector Machines Martin Law.
Overview of Kernel Methods Prof. Bennett Math Model of Learning and Discovery 2/27/05 Based on Chapter 2 of Shawe-Taylor and Cristianini.
Efficient Model Selection for Support Vector Machines
Algorithms for a large sparse nonlinear eigenvalue problem Yusaku Yamamoto Dept. of Computational Science & Engineering Nagoya University.
Cs: compressed sensing
CHEMISTRY: STRUCTURE OF MATTER. THE STRUCTURE OF MATTER What is matter? – Matter is anything that takes up space and has mass All matter is made up of.
Universit at Dortmund, LS VIII
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
1 Lower Bounds Lower bound: an estimate on a minimum amount of work needed to solve a given problem Examples: b number of comparisons needed to find the.
An Introduction to Support Vector Machines (M. Law)
Regret Minimizing Equilibria of Games with Strict Type Uncertainty Stony Brook Conference on Game Theory Nathanaël Hyafil and Craig Boutilier Department.
Mind Change Optimal Learning Of Bayes Net Structure Oliver Schulte School of Computing Science Simon Fraser University Vancouver, Canada
4.6: Rank. Definition: Let A be an mxn matrix. Then each row of A has n entries and can therefore be associated with a vector in The set of all linear.
Algorithms 2005 Ramesh Hariharan. Algebraic Methods.
Simplicity, Induction, and Scientific Discovery School of Computing Science Simon Fraser University Vancouver, Canada.
Optimal Component Analysis Optimal Linear Representations of Images for Object Recognition X. Liu, A. Srivastava, and Kyle Gallivan, “Optimal linear representations.
Information Theory Linear Block Codes Jalal Al Roumy.
Optimal Dimensionality of Metric Space for kNN Classification Wei Zhang, Xiangyang Xue, Zichen Sun Yuefei Guo, and Hong Lu Dept. of Computer Science &
Survey of Kernel Methods by Jinsan Yang. (c) 2003 SNU Biointelligence Lab. Introduction Support Vector Machines Formulation of SVM Optimization Theorem.
ES 240: Scientific and Engineering Computation. Chapter 8 Chapter 8: Linear Algebraic Equations and Matrices Uchechukwu Ofoegbu Temple University.
1  The Problem: Consider a two class task with ω 1, ω 2   LINEAR CLASSIFIERS.
Support vector machine LING 572 Fei Xia Week 8: 2/23/2010 TexPoint fonts used in EMF. Read the TexPoint manual before you delete this box.: A 1.
Review of Linear Algebra Optimization 1/16/08 Recitation Joseph Bradley.
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
State Observer (Estimator)
Arab Open University Faculty of Computer Studies M132: Linear Algebra
Learning Bayes Nets Based on Conditional Dependencies Oliver Schulte Department of Philosophy and School of Computing Science Simon Fraser University Vancouver,
A New Method for Quick Solving Quadratic Assignment Problem CO 769. Paper presentation. by Zhongzhen Zhang.
Feature Selction for SVMs J. Weston et al., NIPS 2000 오장민 (2000/01/04) Second reference : Mark A. Holl, Correlation-based Feature Selection for Machine.
STATIC ANALYSIS OF UNCERTAIN STRUCTURES USING INTERVAL EIGENVALUE DECOMPOSITION Mehdi Modares Tufts University Robert L. Mullen Case Western Reserve University.
ATOMIC THEORY Philosophical Idea to Scientific Theory Chapter 3 Section 1.
Department of Computer Science and Engineering On Computing handles and tunnels for surfaces Tamal K. DeyKuiyu LiJian Sun.
Compression for Fixed-Width Memories Ori Rottenstriech, Amit Berman, Yuval Cassuto and Isaac Keslassy Technion, Israel.
How Particle Physics Cuts Nature At Its Joints Oliver Schulte Department of Philosophy Simon Fraser University `
Spectral Algorithms for Learning HMMs and Tree HMMs for Epigenetics Data Kevin C. Chen Rutgers University joint work with Jimin Song (Rutgers/Palentir),
EE611 Deterministic Systems Controllability and Observability Discrete Systems Kevin D. Donohue Electrical and Computer Engineering University of Kentucky.
Deep Feedforward Networks
Extreme Learning Machine
Glenn Fung, Murat Dundar, Bharat Rao and Jinbo Bi
Chemistry: Structure of Matter
How Particle Physics Cuts Nature At Its Joints
Evolutionary Stability in Bayesian Routing Games
Analysis and design of algorithm
Combinatorial Auctions (Bidding and Allocation)
4.6: Rank.
Chapter 11 Limitations of Algorithm Power
Generally Discriminant Analysis
Presentation transcript:

Discovery of Conservation Laws via Matrix Search Oliver Schulte and Mark S. Drew School of Computing Science Simon Fraser University Vancouver, Canada

DS Discovering Conservation Laws Via Matrix Search 2 /17 Outline Problem Definition: Scientific Model Discovery as Matrix Search. Algorithm for discovering maximally simple maximally strict conservation laws. Comparison with:  Particle physics Standard Model (quark model).  Molecular structure model.

DS Discovering Conservation Laws Via Matrix Search 3 /17 The Matrix Search Model Scientific principle: possible changes in the world are those that conserve relevant quantities. (e.g. conserve energy, electric charge). (Valdes, Zytkow, Simon AAAI 1993) Modelling reactions in chemistry, physics, engineering. n entities participating in m reactions. Input: Reaction integer matrix R mxn. Output: Integer matrix Q nxq s.t. RQ = 0. Q represents hidden features or conserved quantities. These classify reactions as “possible” or “impossible”.

DS Discovering Conservation Laws Via Matrix Search 4 /17 Example: Particle Physics Reactions and Quantities represented as Vectors (Aris 69; Valdés-Pérez 94, 96) i = 1,…n entities r(i) = # of entity i among reagents - # of entity i among products. A quantity is conserved in a reaction if and only if the corresponding vectors are orthogonal. A reaction is “possible” iff it conserves all quantities.

DS Discovering Conservation Laws Via Matrix Search 5 /17 Conserved Quantities in the Standard Model Standard Model based on Gell- Mann’s quark model (1964). Full set of particles: n = 193. Quantity Particle Family (Cluster).

The Learning Task (Toy Example) DS Discovering Conservation Laws Via Matrix Search 6 /17 Given: 1.fixed list of known detectable particles. 2.Input reactions Reactions Reaction Matrix R Output Quantity Matrix Q Learning Cols in Q are conserved, so RQ = 0. Not Given: 1.# of quantities 2.Interpretation of quantities.

Chemistry Example 7 /17 Langley et al Reactions among Chemical Substances Interpretation: #element atoms in each substance molecule. #element atoms conserved in each reaction!

DS Discovering Conservation Laws Via Matrix Search 8 /17 The Totalitarian Principle There are many matrices Q satisfying RQ = 0---how to select? Apply classic “maximally specific criterion” from version spaces (Mitchell 1990). Same general intuition used by physicists:  “everything that can happen without violating a conservation law does happen.” Ford  “anything which is not prohibited is compulsory”. Gell- Mann 1960s. Learning-theoretically optimal (Schulte and Luo COLT 2005)

9 /17 Maximal Strictness (Schulte IJCAI 09) R the smallest generalization of observed reactions R = linear span of R larger generalization of observed reactions R Definition. Q is maximally strict for R if Q allows a minimal superset of R. Proposition. Q is maximally strict for R iff the columns of Q are a basis for the nullspace of R. nullspace of R = null(R) = {v: Rv = 0} Unobserved allowed reactions

DS Discovering Conservation Laws Via Matrix Search 10 /17 Maximally Simple Maximally Strict Matrices (MSMS) L1-norm |M| of matrix M = sum of absolute values of entries. Definition. Conservation matrix Q is an MSMS matrix for reaction matrix R iff Q minimizes |Q| among maximally strict matrices.

DS Discovering Conservation Laws Via Matrix Search 11 /17 Minimization Algorithm Problem. Minimize L1-norm |Q|, subject to nonlinear constraint: Q columns are basis for nullspace of R. Key Ideas. 1.Preprocess to find a basis V of null(R). Search space = {X s.t. Q = VX}. X is small continuous change-of-basis matrix. 2.Discretize after convergence. 1.Set small values to 0. 2.Multiply by lcd to obtain integers.

Example: Chemistry DS Discovering Conservation Laws Via Matrix Search 12 /17 Input Data MSMS Matrix Minimization Program multiply by lcd

Pseudo Code DS Discovering Conservation Laws Via Matrix Search 13 /17

DS Discovering Conservation Laws Via Matrix Search 14 /17 Comparison with Standard Model Implementation in Matlab, use built-in null function. Code available on-line. Dataset  complete set of 193 particles (antiparticles listed separately).  included most probable decay for each unstable particle  182 reactions.  Some others from textbooks for total of 205 reactions.

DS Discovering Conservation Laws Via Matrix Search 15 /17 Results S = Standard Model Laws. Q = output of minimization. Ex2: charge given as input.

DS Discovering Conservation Laws Via Matrix Search 16 /17 Conclusion Conservation Matrix Search problem: for input reactions R, solve RQ=0. New model selection criterion: choose maximally simple maximally strict Q. Efficient local search optimization. Comparison with Standard quark Model  Predictively equivalent.  (Re)discovers particle families-predicted by theorem. Cogsci perspective: MSMS criterion formalizes scientists’ objective.

DS Discovering Conservation Laws Via Matrix Search 17 /17 Thank You Any questions? Oliver at

18 /17 Simplicity and Particle Families Theorem (Schulte and Drew 2006). Let R be a reaction data matrix. If there is a maximally strict conservation matrix Q with disjoint entity clusters, then The clusters (families) are uniquely determined. There is a unique MSMS matrix Q*. pn -- 00 e-e- e --  --  Baryon#Electron#Muon#Tau# Quantity#1Quantity#2Quantity#3Quantity#4 Any alternative set of 4 Q#s with disjoint carriers