Pooja Pun, Avdesh Mishra, Simon Lailvaux, Md Tamjidul Hoque

Slides:



Advertisements
Similar presentations
ECG Signal processing (2)
Advertisements

Integrated Instance- and Class- based Generative Modeling for Text Classification Antti PuurulaUniversity of Waikato Sung-Hyon MyaengKAIST 5/12/2013 Australasian.
Image classification Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing them?
Combining Classification and Model Trees for Handling Ordinal Problems D. Anyfantis, M. Karagiannopoulos S. B. Kotsiantis, P. E. Pintelas Educational Software.
Indian Statistical Institute Kolkata
High Throughput Computing and Protein Structure Stephen E. Hamby.
Sparse vs. Ensemble Approaches to Supervised Learning
These slides are based on Tom Mitchell’s book “Machine Learning” Lazy learning vs. eager learning Processing is delayed until a new instance must be classified.
Sparse vs. Ensemble Approaches to Supervised Learning
Computational Analysis of USA Swimming Data Junfu Xu School of Computer Engineering and Science, Shanghai University.
For Better Accuracy Eick: Ensemble Learning
Attention Deficit Hyperactivity Disorder (ADHD) Student Classification Using Genetic Algorithm and Artificial Neural Network S. Yenaeng 1, S. Saelee 2.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Prediction model building and feature selection with SVM in breast cancer diagnosis Cheng-Lung Huang, Hung-Chang Liao, Mu- Chen Chen Expert Systems with.
LOGO Ensemble Learning Lecturer: Dr. Bo Yuan
The Broad Institute of MIT and Harvard Classification / Prediction.
Transfer Learning with Applications to Text Classification Jing Peng Computer Science Department.
11/12/2012ISC471 / HCI571 Isabelle Bichindaritz 1 Prediction.
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
Gang WangDerek HoiemDavid Forsyth. INTRODUCTION APROACH (implement detail) EXPERIMENTS CONCLUSION.
A NOVEL METHOD FOR COLOR FACE RECOGNITION USING KNN CLASSIFIER
Random Forests Ujjwol Subedi. Introduction What is Random Tree? ◦ Is a tree constructed randomly from a set of possible trees having K random features.
Feature Extraction Artificial Intelligence Research Laboratory Bioinformatics and Computational Biology Program Computational Intelligence, Learning, and.
Classification Ensemble Methods 1
A Brief Introduction and Issues on the Classification Problem Jin Mao Postdoc, School of Information, University of Arizona Sept 18, 2015.
Competition II: Springleaf Sha Li (Team leader) Xiaoyan Chong, Minglu Ma, Yue Wang CAMCOS Fall 2015 San Jose State University.
Musical Genre Categorization Using Support Vector Machines Shu Wang.
SUPPORT VECTOR MACHINES Presented by: Naman Fatehpuria Sumana Venkatesh.
We propose an accurate potential which combines useful features HP, HH and PP interactions among the amino acids Sequence based accessibility obtained.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Avdesh Mishra, Md Tamjidul Hoque {amishra2,
Results for all features Results for the reduced set of features
Bag-of-Visual-Words Based Feature Extraction
CLASSIFICATION OF TUMOR HISTOPATHOLOGY VIA SPARSE FEATURE LEARNING Nandita M. Nayak1, Hang Chang1, Alexander Borowsky2, Paul Spellman3 and Bahram Parvin1.
Evaluating Techniques for Image Classification
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
CSE 4705 Artificial Intelligence
Reading: Pedro Domingos: A Few Useful Things to Know about Machine Learning source: /cacm12.pdf reading.
Natural Language Processing of Knee MRI Reports
Evaluating classifiers for disease gene discovery
Feature Extraction Introduction Features Algorithms Methods
Avdesh Mishra, Manisha Panta, Md Tamjidul Hoque, Joel Atallah
Introduction Feature Extraction Discussions Conclusions Results
Brain Hemorrhage Detection and Classification Steps
Prediction of RNA Binding Protein Using Machine Learning Technique
Machine Learning Week 1.
Extra Tree Classifier-WS3 Bagging Classifier-WS3
Bird-species Recognition Using Convolutional Neural Network
Support Vector Machine (SVM)
TED Talks – A Predictive Analysis Using Classification Algorithms
Prediction of Wine Grade
RECOMMENDER SYSTEMS WITH SOCIAL REGULARIZATION
Abdur Rahman Department of Statistics
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
MACHINE LEARNING TECHNIQUES IN IMAGE PROCESSING
Concave Minimization for Support Vector Machine Classifiers
Machine learning overview
Reecha Khanal Mentor: Avdesh Mishra Supervisor: Dr. Md Tamjidul Hoque
Detection of Sand Boils using Machine Learning Approaches
Machine Learning with Clinical Data
Topological Signatures For Fast Mobility Analysis
Machine Learning – a Probabilistic Perspective
Physics-guided machine learning for milling stability:
MAS 622J Course Project Classification of Affective States - GP Semi-Supervised Learning, SVM and kNN Hyungil Ahn
Manisha Panta, Avdesh Mishra, Md Tamjidul Hoque, Joel Atallah
Results Motivation Introduction Methods Conclusions Acknowledgements
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Outlines Introduction & Objectives Methodology & Workflow
Machine Learning for Cyber
Presentation transcript:

Pooja Pun, Avdesh Mishra, Simon Lailvaux, Md Tamjidul Hoque A Machine Learning Approach to Functional Morphology and Performance Prediction Pooja Pun, Avdesh Mishra, Simon Lailvaux, Md Tamjidul Hoque Email: ppun@uno.edu, amishra2@uno.edu, slailvau@uno.edu, thoque@uno.edu Department of Computer Science, Department of Biological Sciences, University of New Orleans, New Orleans, LA, USA Introduction Functional morphology involves the study of the relationship between the physical structures and the functions of the various parts of an organism. Morphology-to-performance relationships are well explained for individual performance traits, particularly in species that are highly specialized for conducting a specific biological task. But the same relationship is ambiguous and controversial for extinct species, mainly because these species have no living analogues, which makes it difficult to validate the predicted relationship obtained by the extrapolation from the form and function relationships of other living animals, to which they bear little resemblance. Classification Table 1: Accuracy and MCC for classification Feature selection before base layers Overall Accuracy Matthews Correlation Coefficient Support Vector Machines (SVM) 0.94537 0.93264 Stacking without feature selection 0.93191 0.916 SVM after feature selection 0.94695 0.93455 Stacking after feature selection 0.94062 0.92672 Stacking with feature selection before meta layer 0.94141 0.92773 Base Layers: SVC, LogReg, XGBC Feature Selection before meta layer Meta Layer: SVC Fig 1: Flowchart for classification Regression The dataset was divided into ten test and train datasets which were used to train and test different regressors. The best performing regressors were combined in a stacking manner for more robust results. Objectives Proper understanding of the morphology-to-performance relationships of extinct animals, mainly Australian Marsupials Rigorously validate the predictions made by using machine learning tools Results Methods Training Datasets Dataset containing functional and morphological data of 31 lizard species with a total of 1263 samples was provided by the Department of Biological Sciences. Missing values in the dataset were replaced by two-step k-Nearest Neighbor (kNN) based approach. In the first step, the missing value of a target sample, belonging to a target species, was replaced by the average value of five samples closest to the target sample determined by computing the euclidean distance between the target sample and all other samples present in the target species. In the second step, the same process was applied but to the whole dataset. Machine Learning Methods Feature Selection is done using an evolutionary algorithm.. In stacking framework, classifiers are stacked in two layers: base layer and meta layer. The results from the base layer is a basis for the dataset feed into the meta layer. Fig 3: PCC and MAE for different regression methods for Bite power Table 2: PCC and MAE for regression of Bite power Pearson Correlation Coefficient (PCC) Mean Absolute Error (MAE) Gradient Boosting Regressor (without using Test and Train Datasets) 0.73044 0.85863 XGBoost Regressor (using Test and Train Datasets) 0.95016 1.10059 Stacking 0.97637 0.72926 Fig 2: Accuracy and MCC for Classification Discussions kNN played a key role in obtaining high accuracy in both classification and regression. Stacking with feature selection elevated the accuracy for both classification and regression. Conclusions Our results suggest that machine learning is a promising approach for making accurate predictions of performance abilities from morphology alone. Future applications of these methods to the morphology of extinct organisms may allow us to reduce the uncertainty in performance prediction whilst making fewer assumptions. Acknowledgements We gratefully acknowledge the Louisiana Board of Regents through the Board of Regents Support Fund, LEQSF (2016-19)-RD-B-07. We also gratefully acknowledge the University of New Orleans for the Internal FY19 IGD Fund, Award #: CON000000002946.