Detailed q2/Q2 results for 100 bootstraps for final runs with (38 + dummy features)

Slides:



Advertisements
Similar presentations
Pseudo Inverse Heisenberg Uncertainty for Data Mining Explicit Principal Components Implicit Principal Components NIPALS Algorithm for Eigenvalues and.
Advertisements

A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
NEURAL NETWORKS Backpropagation Algorithm
Random Forest Predrag Radenković 3237/10
Ensemble Learning Reading: R. Schapire, A brief introduction to boosting.
Multilayer Perceptrons 1. Overview  Recap of neural network theory  The multi-layered perceptron  Back-propagation  Introduction to training  Uses.
Ensemble Methods An ensemble method constructs a set of base classifiers from the training data Ensemble or Classifier Combination Predict class label.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
1 Part I Artificial Neural Networks Sofia Nikitaki.
Sparse vs. Ensemble Approaches to Supervised Learning
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
Neural Networks. R & G Chapter Feed-Forward Neural Networks otherwise known as The Multi-layer Perceptron or The Back-Propagation Neural Network.
Ensemble Learning: An Introduction
Prénom Nom Document Analysis: Artificial Neural Networks Prof. Rolf Ingold, University of Fribourg Master course, spring semester 2008.
1 Simulation Modeling and Analysis Verification and Validation.
Classification of Music According to Genres Using Neural Networks, Genetic Algorithms and Fuzzy Systems.
Bagging LING 572 Fei Xia 1/24/06. Ensemble methods So far, we have covered several learning methods: FSA, HMM, DT, DL, TBL. Question: how to improve results?
Sparse vs. Ensemble Approaches to Supervised Learning
Saturation, Flat-spotting Shift up Derivative Weight Decay No derivative on output nodes.
Self-organizing Learning Array based Value System — SOLAR-V Yinyin Liu EE690 Ohio University Spring 2005.
Traffic Sign Recognition Using Artificial Neural Network Radi Bekker
Genetic Programming System for Music Generation With Automated Fitness Raters.
Predicting Income from Census Data using Multiple Classifiers Presented By: Arghya Kusum Das Arnab Ganguly Manohar Karki Saikat Basu Subhajit Sidhanta.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
1 st Neural Network: AND function Threshold(Y) = 2 X1 Y X Y.
Cascade Correlation Architecture and Learning Algorithm for Neural Networks.
Artificial Neural Networks
Forecasting Revenue: An Example of Regression Model Building Setting: Possibly a large set of predictor variables used to predict future quarterly revenues.
Appendix B: An Example of Back-propagation algorithm
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
Chapter 18 Four Multivariate Techniques Angela Gillis & Winston Jackson Nursing Research: Methods & Interpretation.
NEURAL NETWORKS FOR DATA MINING
 Diagram of a Neuron  The Simple Perceptron  Multilayer Neural Network  What is Hidden Layer?  Why do we Need a Hidden Layer?  How do Multilayer.
Modelling Language Evolution Lecture 1: Introduction to Learning Simon Kirby University of Edinburgh Language Evolution & Computation Research Unit.
Dünaamiliste süsteemide modelleerimine Identification for control in a non- linear system world Eduard Petlenkov, Automaatikainstituut, 2013.
Ensemble Methods: Bagging and Boosting
Ensembles. Ensemble Methods l Construct a set of classifiers from training data l Predict class label of previously unseen records by aggregating predictions.
Multi-Layer Perceptron
1 Lecture 6 Neural Network Training. 2 Neural Network Training Network training is basic to establishing the functional relationship between the inputs.
ROOT ROOT.PAT ROOT.TES (ROOT.WGT) (ROOT.FWT) (ROOT.DBD) MetaNeural ROOT.XXX ROOT.TTT ROOT.TRN (ROOT.DBD) ROOT.WGT ROOT.FWT Use Analyze root –34 for easy.
Procedure for Training a Child to Identify a Cat using 10,000 Example Cats For Cat_index  1 to Show cat and describe catlike features (Cat_index)
Introduction to Neural Networks Introduction to Neural Networks Applied to OCR and Speech Recognition An actual neuron A crude model of a neuron Computational.
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
Classification and Prediction: Ensemble Methods Bamshad Mobasher DePaul University Bamshad Mobasher DePaul University.
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
COMPUTATIONAL INTELLIGENCE
Artificial Neural Networks for Data Mining. Copyright © 2011 Pearson Education, Inc. Publishing as Prentice Hall 6-2 Learning Objectives Understand the.
Neural networks – Hands on
Bab 5 Classification: Alternative Techniques Part 4 Artificial Neural Networks Based Classifer.
Computer Programming 12 Lesson 6 – Loop structure By: Dan Lunney.
Neural networks (2) Reminder Avoiding overfitting Deep neural network Brief summary of supervised learning methods.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Ensemble Classifiers.
Neural Networks.
Deep Feedforward Networks
Artificial Neural Networks I
LEAST – SQUARES REGRESSION
Artificial Neural Networks
A Taylor Rule with Monthly Data
Chapter 13 – Ensembles and Uplift
Photorealistic Image Colourization with Generative Adversarial Nets
Java Implementation of Optimal Brain Surgeon
A “Holy Grail” of Machine Learing
Generalization ..
Prof. Carolina Ruiz Department of Computer Science
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Prof. Carolina Ruiz Department of Computer Science
Presentation transcript:

Detailed q2/Q2 results for 100 bootstraps for final runs with (38 + dummy features)

Ordered by correlation coefficient

Example last pass feature Selection from 80 sensitivity bags Random dummy variable Bagged relative sensitivity from 80 bootstraps for the random dummy variable: descriptors with lower sensitivities will be eliminated in the next iteration Descriptors that will be eliminated in the next iteration

STRIPMINER OPERATION MODE Mode #6: feature selection with sensitivity analysis (~ 1000 neural nets) (Q2 = 0.46, all molecules) Bootstraps with sensitivity analysis with a dummy var for descriptor selection (480  39 descriptors) Mode #0: train neural nets 300 bootstrap ANNs (300 neural nets trained) Ensemble bagging for selected descriptors Note: all ANNs 39x13x11x1 trained to error of pats in validation set Mode #4: predict for test set using bagging weights (100x30/300 bags) (3000 ANNs in user mode) Bag prediction on test set Note: ensemble results weighted by Q2 calculated in Mode#0

Stripminer Neural Network Sensitivity Analysis With Dummy Feature REPEAT REPEAT 100x Do neural network bootstrap And calculate Q2 for validation set There is one random dummy feature There is a validation set for bagging Prepare file for sensitivity analysis (can be up to 30 MB) Run neural net in user mode for sensitivity analysis MetaNeural Calculate sensitivity results for 13 levels and tally results in sen#.txt SENSIT Reduce features by dropping feats with lower sensitivity than dummy CONTINUE Bag sensitivities TEST (repeat until the dummy variable is the least sensitive feature) Bagging and feature selection

   Molecular weight H-bonding Boiling Point Hydrofobicity Biological response Electrostatic interactions   w 11 w 34   w 23 w 11 h h Neural Network Molecular Descriptor Observable Projection Neural Network Sensitivity Analysis RENSSELAER Keep all inputs frozen at median values Turn one input at a time from 0 to 1 Monitor vaqiation in outputs Outputs with largest variation are most sensitive  more important

Correleation biased Sum of 30 best correlated variables seems to have spurious correlation