An Investigation into the Relationship between Semantic and Content Based Similarity Using LIDC Grace Dasovich Robert Kim Midterm Presentation August 21.

Slides:



Advertisements
Similar presentations
Integration of Radiologists Feedback into Computer-Aided Diagnosis Systems Sarah A. Jabon a Daniela S. Raicu b Jacob D. Furst b a Rose-Hulman Institute.
Advertisements

Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
My name is Dustin Boswell and I will be presenting: Ensemble Methods in Machine Learning by Thomas G. Dietterich Oregon State University, Corvallis, Oregon.
Data Mining Classification: Alternative Techniques
FUZZ-IEEE 2009 Methods of Interpretation of a Non-stationary Fuzzy System for the Treatment of Breast Cancer Xiao-Ying Wang 1, Jonathan M. Garibaldi 1,
Instance Based Learning
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Chapter 17 Overview of Multivariate Analysis Methods
Outline Introduction Anotation Segmentation Detection.
Neural Network Based Approach for Short-Term Load Forecasting
RBF Neural Networks x x1 Examples inside circles 1 and 2 are of class +, examples outside both circles are of class – What NN does.
A Computer Aided Detection System For Digital Mammograms Based on Radial Basis Functions and Feature Extraction Techniques By Mohammed Jirari Shanghai,
Three-dimensional co-occurrence matrices & Gabor filters: Current progress Gray-level co-occurrence matrices Carl Philips Gabor filters Daniel Li Supervisor:
Slide 1 EE3J2 Data Mining EE3J2 Data Mining Lecture 15: Introduction to Artificial Neural Networks Martin Russell.
Artificial Neural Networks Artificial Neural Networks are (among other things) another technique for supervised learning k-Nearest Neighbor Decision Tree.
Mutual Information as a Measure for Image Quality of Temporally Subtracted Chest Radiographs Samantha Passen Samuel G. Armato III, Ph.D.
NSF REU Program in Medical Informatics 1 D. Raicu, 1 J. Furst, 2 D. Channin, 3 S. Armato, and 3 K. Suzuki 1 DePaul University, 2 Northwestern University,
A study on the effect of imaging acquisition parameters on lung nodule image interpretation Presenters: Shirley Yu (University of Southern California)
Lazy Learning k-Nearest Neighbour Motivation: availability of large amounts of processing power improves our ability to tune k-NN classifiers.
NSF MedIX REU Program Medical Imaging DePaul CDM Daniela S. Raicu, PhD Associate Professor Lab URL:
Liver Segmentation Using Active Learning Ankur Bakshi Allison Petrosino Advisor: Dr. Jacob Furst August 21, 2008.
Variability of LIDC Panel Segmentations and Soft Segmentation of Lung Nodules Presented by Stephen Siena and Olga Zinoveva.
1 Automated Feature Abstraction of the fMRI Signal using Neural Network Clustering Techniques Stefan Niculescu and Tom Mitchell Siemens Medical Solutions,
Classification and Prediction by Yen-Hsien Lee Department of Information Management College of Management National Sun Yat-Sen University March 4, 2003.
October 28, 2010Neural Networks Lecture 13: Adaptive Networks 1 Adaptive Networks As you know, there is no equation that would tell you the ideal number.
CS Instance Based Learning1 Instance Based Learning.
Oral Defense by Sunny Tang 15 Aug 2003
Discriminant Analysis Testing latent variables as predictors of groups.
197 Case Study: Predicting Breast Cancer Invasion with Artificial Neural Networks on the Basis of Mammographic Features MEDINFO 2004, T02: Machine Learning.
ENDA MOLLOY, ELECTRONIC ENG. FINAL PRESENTATION, 31/03/09. Automated Image Analysis Techniques for Screening of Mammography Images.
Classification of multiple cancer types by multicategory support vector machines using gene expression data.
Hurieh Khalajzadeh Mohammad Mansouri Mohammad Teshnehlab
Outline What Neural Networks are and why they are desirable Historical background Applications Strengths neural networks and advantages Status N.N and.
NEURAL NETWORKS FOR DATA MINING
Chapter 7 Neural Networks in Data Mining Automatic Model Building (Machine Learning) Artificial Intelligence.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Chapter 6: Techniques for Predictive Modeling
Filter + Support Vector Machine for NIPS 2003 Challenge Jiwen Li University of Zurich Department of Informatics The NIPS 2003 challenge was organized to.
Relevance Feedback: New Trends Derive global optimization methods: More computationally robust Consider the correlation between different attributes Incorporate.
Ensemble Methods: Bagging and Boosting
BAGGING ALGORITHM, ONLINE BOOSTING AND VISION Se – Hoon Park.
EE459 Neural Networks Examples of using Neural Networks Kasin Prakobwaitayakit Department of Electrical Engineering Chiangmai University.
Evolutionary Algorithms for Finding Optimal Gene Sets in Micro array Prediction. J. M. Deutsch Presented by: Shruti Sharma.
Non-Bayes classifiers. Linear discriminants, neural networks.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Intro. ANN & Fuzzy Systems Lecture 14. MLP (VI): Model Selection.
Reservoir Uncertainty Assessment Using Machine Learning Techniques Authors: Jincong He Department of Energy Resources Engineering AbstractIntroduction.
Alex Stabile. Research Questions: Could a computer learn to distinguish between different composers? Why does music by different composers even sound.
Date of download: 5/28/2016 Copyright © 2016 SPIE. All rights reserved. Flowchart of the computer-aided diagnosis (CAD) tool. (a) Segmentation: The region.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Automatic Lung Cancer Diagnosis from CT Scans (Week 1)
Debesh Jha and Kwon Goo-Rak
Artificial Neural Networks
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Restricted Boltzmann Machines for Classification
Implementing Boosting and Convolutional Neural Networks For Particle Identification (PID) Khalid Teli .
Final Year Project Presentation --- Magic Paint Face
CSSE463: Image Recognition Day 17
CAMCOS Report Day December 9th, 2015 San Jose State University
Collaborative Filtering Nearest Neighbor Approach
Advanced Artificial Intelligence Classification
CSSE463: Image Recognition Day 17
Somi Jacob and Christian Bach
Areas under the receiver operating characteristic (ROC) curves for both the training and testing data sets based on a number of hidden-layer perceptrons.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
What is Artificial Intelligence?
Presentation transcript:

An Investigation into the Relationship between Semantic and Content Based Similarity Using LIDC Grace Dasovich Robert Kim Midterm Presentation August

Outline Related Work Data Modeling Approach and Results –Similarity Measures –Artificial Neural Network –Multivariate Linear Regression Conclusions Future Work

Computer-Aided Diagnosis (CADx) based on low-level image features –Armato et al. developed a linear discriminant classifier using features of lung nodules –Need to find the relationship between the image features and radiologists’ ratings Related Work

Image features and the semantic ratings –Lung Interpretations Barb et al. developed Evolutionary System for Semantic Exchange of Information in Collaborative Environments (ESSENCE) Raicu et al. used ensemble classifiers and decision trees to predict semantic ratings Samala et al. used several combinations of image features and the radiologists’ ratings to classify nodules Related Work

–Similarity Li et al. investigated four different methods to compute similarity measures for lung nodules –Feature-based –Pixel-value-difference –Cross correlation –ANN Related Work

Materials LIDC Dataset 149 Unique Nodules –One slice per nodule, largest nodule area 9 Semantic Characteristics –Calcification and Internal Structure had little variation, thus were not used 64 Content Features –Shape, size, intensity, and texture 6 Data

Related Work Data Modeling Approach and Results –Similarity Measures –Artificial Neural Network –Multivariate Linear Regression Conclusions Future Work Outline

Cosine Similarity Jeffrey Divergence Euclidean Distance Similarity Measures

Computed feature distance measures Similarity Measures

Outline Related Work Data Modeling Approach and Results –Similarity Measures –Artificial Neural Network –Multivariate Linear Regression Conclusions Future Work

Two three-layer ANNs –Input (64 neurons), hidden layer (5 neurons), output (1) –Input (64 neurons), hidden layer (5 neurons), output (7) Input = 64 feature distances Output = Semantic similarity or difference in semantic ratings Hyperbolic tangent function, backpropagation algorithm, 200 iterations Methods

ANN with a single output –640 random pairs from all 109 nodules –231 pairs from nodules with malignancy > 3 –496 pairs from nodules with area > 122 mm 2 Methods

ANN with seven outputs –640 random pairs from all 109 nodules

Leave-one-out method –Cosine similarity or Jeffrey divergence or difference in Semantic ratings used as teaching data –An ANN trained with entire dataset minus one image pair –The pair left out used for testing –Correlation between calculated radiologists’ similarity and ANN output calculated Methods

ANN with a single output –640 random pairs from all 109 nodules –231 pairs from nodules with malignancy > 3 –496 pairs from nodules with area > 122 mm 2 ANN with seven outputs –640 random pairs from all 109 nodules Methods

ANN using 640 random pairs Results

ANN using 231 pairs with malignancy rating > 3 Results

ANN using 496 pairs with area > 122 mm 2 Results

ANN output vs. target values using Jeffrey divergence for the 640 pairs (r = 0.438) Results

ANN using random 640 pairs and the Jeffrey divergence with seven semantic ratings Results

Outline Related Work Data Modeling Approach and Results –Similarity Measures –Artificial Neural Network –Multivariate Linear Regression Conclusions Future Work

Methods Normalization of Features –Min-Max Technique –Z-Score Technique Pair Selection –Looked for matches between k number of most similar images based on semantic and content 24 Methods

Multivariate Regression Analysis –Select features with highest correlation coefficients –Feature distance measures 25 Methods

Nodule Analysis –Determine differences between selected and non-selected nodules –Define requirements for our model Methods

Results 27 Results

d(i, j)d 2 (i, j)exp(d(i, j)) Cosine Jeffrey

Results Correlation CoefficientFeature Equivalent Diameter Energy (Haralick) Gabor Mean 135_ Convex Area Gabor STD 135_ Min Intensity BG Markov Variance (Haralick) Gabor STD 45_ SD Intensity R 2 = Results

30 Results

31 Results

32 Results

A. Equivalent Diameter, B. Standard Deviation of Intensity, C. Malignancy, D. Subtlety

Preliminary Issues The ANN also is not yet sufficient to predict semantic similarity from content –Best correlation –Malignancy correlation –Jeffrey performed better unlike linear model A semantic gap still exists Conclusions

Our linear model applies to a specific type of nodule –Characteristics: High malignancy, high texture, low lobulation, and low spiculation –Features: Larger diameter, greater intensity Linear models are not sufficient for determination of similarities –R 2 of with chosen nodules 35 Conclusions

Future Work Reduce variability among radiologists –Use only nodules with radiologists’ agreement Find best combination of content features –64 may be too many –Currently only using 2D Future Work

Different semantic distance measures –Some ratings are ordinal, Jeffery is for categorical Different methods of machine learning –Incorporate radiologists’ feedback into training –Ensemble of classifiers Future Work

Thanks for Listening Any Questions? 38 Thanks for Listening