Biological Data Mining A comparison of Neural Network and Symbolic Techniques

Slides:



Advertisements
Similar presentations
Artificial Intelligence 12. Two Layer ANNs
Advertisements

Rule Extraction from trained Neural Networks Brian Hudson, Centre for Molecular Design, University of Portsmouth.
Ignas Budvytis*, Tae-Kyun Kim*, Roberto Cipolla * - indicates equal contribution Making a Shallow Network Deep: Growing a Tree from Decision Regions of.
Rule extraction in neural networks. A survey. Krzysztof Mossakowski Faculty of Mathematics and Information Science Warsaw University of Technology.
Decision Tree Approach in Data Mining
Data Mining Classification: Basic Concepts, Decision Trees, and Model Evaluation Lecture Notes for Chapter 4 Part I Introduction to Data Mining by Tan,
Feature selection and transduction for prediction of molecular bioactivity for drug design Reporter: Yu Lun Kuo (D )
Data Mining Classification: Alternative Techniques
Deriving rules from data Decision Trees a.j.m.m (ton) weijters.
Comparison of Data Mining Algorithms on Bioinformatics Dataset Melissa K. Carroll Advisor: Sung-Hyuk Cha March 4, 2003.
Application of Stacked Generalization to a Protein Localization Prediction Task Melissa K. Carroll, M.S. and Sung-Hyuk Cha, Ph.D. Pace University, School.
Decision Tree Rong Jin. Determine Milage Per Gallon.
Rule Extraction From Trained Neural Networks Brian Hudson University of Portsmouth, UK.
Biological Data Mining A comparison of Neural Network and Symbolic Techniques
1 Classification with Decision Trees I Instructor: Qiang Yang Hong Kong University of Science and Technology Thanks: Eibe Frank and Jiawei.
Classification Continued
Classification.
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
1 Decision Trees Greg Grudic (Notes borrowed from Thomas G. Dietterich and Tom Mitchell) Modified by Longin Jan Latecki Some slides by Piyush Rai Intro.
Statistical Learning: Pattern Classification, Prediction, and Control Peter Bartlett August 2002, UC Berkeley CIS.
Data Mining – Intro.
Comp 540 Chapter 9: Additive Models, Trees, and Related Methods
Machine Learning Márk Horváth Morgan Stanley FID Institutional Securities.
Computer Science Universiteit Maastricht Institute for Knowledge and Agent Technology Data mining and the knowledge discovery process Summer Course 2005.
GUHA method in Data Mining Esko Turunen Tampere University of Technology Tampere, Finland.
Enterprise systems infrastructure and architecture DT211 4
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Inductive Logic Programming Includes slides by Luis Tari CS7741L16ILP.
DATA MINING : CLASSIFICATION. Classification : Definition  Classification is a supervised learning.  Uses training sets which has correct answers (class.
Copyright R. Weber Machine Learning, Data Mining ISYS370 Dr. R. Weber.
Evolutionary Algorithms BIOL/CMSC 361: Emergence Lecture 4/03/08.
Dr. Russell Anderson Dr. Musa Jafar West Texas A&M University.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Predicting the Cellular Localization Sites of Proteins Using Decision Tree and Neural Networks Yetian Chen
Bayesian networks Classification, segmentation, time series prediction and more. Website: Twitter:
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Kernel Methods A B M Shawkat Ali 1 2 Data Mining ¤ DM or KDD (Knowledge Discovery in Databases) Extracting previously unknown, valid, and actionable.
Chapter 6: Techniques for Predictive Modeling
Business Intelligence and Decision Modeling Week 9 Customer Profiling Decision Trees (Part 2) CHAID CRT.
Data Mining – Intro. Course Overview Spatial Databases Temporal and Spatio-Temporal Databases Multimedia Databases Data Mining.
Computational Intelligence: Methods and Applications Lecture 20 SSV & other trees Włodzisław Duch Dept. of Informatics, UMK Google: W Duch.
Module networks Sushmita Roy BMI/CS 576 Nov 18 th & 20th, 2014.
CS690L Data Mining: Classification
1 Data Mining dr Iwona Schab Decision Trees. 2 Method of classification Recursive procedure which (progressively) divides sets of n units into groups.
Protein motif extraction with neuro-fuzzy optimization Bill C. H. Chang and Author : Bill C. H. Chang and Saman K. Halgamuge Saman K. Halgamuge Adviser.
Data Mining and Decision Trees 1.Data Mining and Biological Information 2.Data Mining and Machine Learning Techniques 3.Decision trees and C5 4.Applications.
Digital Camera and Computer Vision Laboratory Department of Computer Science and Information Engineering National Taiwan University, Taipei, Taiwan, R.O.C.
Decision Trees Binary output – easily extendible to multiple output classes. Takes a set of attributes for a given situation or object and outputs a yes/no.
Classification (slides adapted from Rob Schapire) Eran Segal Weizmann Institute.
Support Vector Machines. Notation Assume a binary classification problem. –Instances are represented by vector x   n. –Training examples: x = (x 1,
An Introduction Student Name: Riaz Ahmad Program: MSIT( ) Subject: Data warehouse & Data Mining.
Biological Data Mining A comparison of Neural Network and Symbolic Techniques
Chong Ho Yu.  Data mining (DM) is a cluster of techniques, including decision trees, artificial neural networks, and clustering, which has been employed.
CSC 8520 Spring Paula Matuszek DecisionTreeFirstDraft Paula Matuszek Spring,
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
An Interval Classifier for Database Mining Applications Rakes Agrawal, Sakti Ghosh, Tomasz Imielinski, Bala Iyer, Arun Swami Proceedings of the 18 th VLDB.
1 Using Graph Theory to Analyze Gene Network Coherence José A. Lagares Jesús S. Aguilar Norberto Díaz-Díaz Francisco A. Gómez-Vela
Eco 6380 Predictive Analytics For Economists Spring 2016 Professor Tom Fomby Department of Economics SMU.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Classification Tree Interaction Detection. Use of decision trees Segmentation Stratification Prediction Data reduction and variable screening Interaction.
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
Spectral Algorithms for Learning HMMs and Tree HMMs for Epigenetics Data Kevin C. Chen Rutgers University joint work with Jimin Song (Rutgers/Palentir),
RESEARCH APPROACH.
NBA Draft Prediction BIT 5534 May 2nd 2018
Decision Trees Greg Grudic
Virtual Screening.
Prepared by: Mahmoud Rafeek Al-Farra
Artificial Intelligence 12. Two Layer ANNs
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Presentation transcript:

Biological Data Mining A comparison of Neural Network and Symbolic Techniques

1. Objectives The project aims: –to develop and validate techniques for extracting explicit information from bioinformatic data –to express this information as logical rules and decision trees –to apply these new procedures to a range of scientific problems related to bioinformatics and cheminformatics

2. Extracting information Artificial neural networks can be trained to reproduce the non-linear relationships underlying bioinformatic data with good predictive accuracy –but it is often hard to comprehend those relationships from the internal structure of the network –with the result that networks are often regarded as ‘black boxes’. Decision trees using symbolic rules are easier to interpret – leading to a greater likelihood of understanding the relationships in the data –allowing the behaviour of individual cases to be explained.

3. Extracting Decision Trees The Trepan procedure (Craven,1996) extracts decision trees from a neural network and a set of training cases by recursively partitioning the input space. The decision tree is built in a best-first manner, expanding the tree at nodes where there is greatest potential for increasing the fidelity of the tree to the network.

4. Splitting Tests The splitting tests at the nodes are m-of-n expressions, e.g. 2-of-{x 1, ¬x 2, x 3 }, where the x i are Boolean conditions. Start with a set of candidate tests –binary tests on each value for nominal features –binary tests on thresholds for real-valued features Use a beam search with a beam width of two. Initialize the beam with the candidate test that maximizes the information gain.

5. Splitting Tests (II) To each m-of-n test in the beam and each candidate test, apply two operators: m-of-n+1e.g. 2-of-{x 1, x 2 } => 2-of-{x 1, x 2, x 3 } m+1-of-n+1e.g. 2-of-{x 1, x 2 } => 3-of-{x 1, x 2, x 3 } Admit new tests to the beam if they increase the information gain and are significantly different (chi-squared) from existing tests.

6. Example: Substance P Binding to NK1 Receptors Substance P is a neuropeptide with the sequence: H-Arg-Pro-Lys-Pro-Gln-Gln-Phe-Phe-Gly-Leu-Met-NH 2 Wang et al. used the multipin technique to synthesize 512 = 2 9 stereoisomers generated by systematic replacement of L- by D-amino acids at 9 positions The aim was to measure binding potencies to NK1 receptors & identify the positions at which stereo- chemistry affects binding strength.

7. Application of Trepan A series of networks with 9:9:1 architectures were trained using 90% of the data as a training set. For each network a decision tree was grown using Trepan. The trees showed high fidelity with the networks on a 10% test set.

8. Results Binding activity was determined by five positions, viz. –H-Arg-Pro-Lys-Pro-Gln-Gln-Phe-Phe-Gly-Leu-Met- NH 2 The positions identified agree with the FIRM (Formal Inference-based Recursive Modelling) analysis of Young and Hawkins –Young S & Hawkins D.M. (2000) Analysis of a large, high-throughput screening data using recursive partitioning. Molecular Modelling & Prediction of Bioactivity (ed. Gundertofte & J Ø rgensen).

9. A Typical Trepan Tree

10. Test set confusion matrix: tree versus network

11. Test set confusion matrix: tree versus observed

12. Future Work Complete the implementation of the Trepan algorithm. –model the distribution of the input data and generate a set of query instances to be classified by the network & used as additional training cases during tree extraction. Extend the algorithm to enable the extraction of regression trees. Provide a Bayesian formulation for the decision tree extraction algorithm.

13. Future Applications Apply Trepan to ligand-receptor binding problems. –compare the performance of these algorithms with existing symbolic data mining techniques (ID3/C5).

14. References Wang J-X et al. (1993) Study of stereo-requirements of substance P binding to NK1 receptors using analogues with systematic D-amino acid replacements. Biorganic & Medicinal Chemistry Letters, 3, Young S & Hawkins D.M. (2000) Analysis of a large, high-throughput screening data using recursive partitioning. Molecular Modelling & Prediction of Bioactivity (ed. Gundertofte & JØrgensen).

Grantholder Professor Martyn Ford Centre for Molecular Design University of Portsmouth Dr Shuang Cang Mar - Sept 2000 Dr Abul Azad Jan Research Fellows

Collaborators Dr Antony Browne School of Computing, Information Systems and Mathematics, London Guildhall University. Professor Philip Picton School of Technology and Design, University College Northampton. Dr David Whitley Centre for Molecular Design, University of Portsmouth.