Chemometrics for Analysis of NIR Spectra

Slides:



Advertisements
Similar presentations
Things to do in Lecture 1 Outline basic concepts of causality
Advertisements

PCA for analysis of complex multivariate data. Interpretation of large data tables by PCA In industry, research and finance the amount of data is often.
Regression analysis Relating two data matrices/tables to each other Purpose: prediction and interpretation Y-data X-data.
Supervised Learning Recap
10/11/ ENGINEERING RESEARCH CENTER FOR S TRUCTURED O RGANIC P ARTICULATE S YSTEMS RUTGERS UNIVERSITY PURDUE UNIVERSITY NEW JERSEY INSTITUTE OF TECHNOLOGY.
Division of Pharmaceutical Analysis Research in support of the Critical Path Dimensions Ensuring Safety Demonstrating Medical Utility Industrialization.
Sparse vs. Ensemble Approaches to Supervised Learning
Bayesian Framework EE 645 ZHAO XIN. A Brief Introduction to Bayesian Framework The Bayesian Philosophy Bayesian Neural Network Some Discussion on Priors.
Lasso regression. The Goals of Model Selection Model selection: Choosing the approximate best model by estimating the performance of various models Goals.
Data mining and statistical learning - lecture 11 Neural networks - a model class providing a joint framework for prediction and classification  Relationship.
Customizing the Content: The placeholders in this poster are formatted for you. Type in the placeholders to add text, or click an icon to add a table,
Chapter 5 Data mining : A Closer Look.
Midterm Review. 1-Intro Data Mining vs. Statistics –Predictive v. experimental; hypotheses vs data-driven Different types of data Data Mining pitfalls.
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Model Building III – Remedial Measures KNNL – Chapter 11.
Introduction to machine learning and data mining 1 iCSC2014, Juan López González, University of Oviedo Introduction to machine learning Juan López González.
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression.
CS Statistical Machine learning Lecture 18 Yuan (Alan) Qi Purdue CS Oct
Jeff Howbert Introduction to Machine Learning Winter Regression Linear Regression Regression Trees.
Feature (Gene) Selection MethodsSample Classification Methods Gene filtering: Variance (SD/Mean) Principal Component Analysis Regression using variable.
CpSc 881: Machine Learning
Neural Networks Demystified by Louise Francis Francis Analytics and Actuarial Data Mining, Inc.
From OLS to Generalized Regression Chong Ho Yu (I am regressing)
QUANTITATIVE ANALYSIS OF POLYMORPHIC MIXTURES USING INFRARED SPECTROSCOPY IR Spectroscopy Calibration –Homogeneous Solid-State Mixtures –Multivariate Calibration.
Data analysis tools Subrata Mitra and Jason Rahman.
Logistic Regression & Elastic Net
1 Traffic accident analysis using machine learning paradigms Miao Chong, Ajith Abraham, Mercin Paprzycki Informatica 29, P89, 2005 Report: Hsin-Chan Tsai.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
Regularized Least-Squares and Convex Optimization.
Multivariate statistical methods. Multivariate methods multivariate dataset – group of n objects, m variables (as a rule n>m, if possible). confirmation.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Machine Learning Supervised Learning Classification and Regression K-Nearest Neighbor Classification Fisher’s Criteria & Linear Discriminant Analysis Perceptron:
Data Science Practical Machine Learning Tools and Techniques 6.8: Clustering Rodney Nielsen Many / most of these slides were adapted from: I. H. Witten,
PREDICT 422: Practical Machine Learning Module 4: Linear Model Selection and Regularization Lecturer: Nathan Bastian, Section: XXX.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
CSE 4705 Artificial Intelligence
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
JMP Discovery Summit 2016 Janet Alvarado
Course Review Questions will not be all on one topic, i.e. questions may have parts covering more than one area.
An Artificial Intelligence Approach to Precision Oncology
Boosting and Additive Trees (2)
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
Matteo Reggente Giulia Ruggeri Adele Kuzmiakova Satoshi Takahama
CSE 4705 Artificial Intelligence
10701 / Machine Learning.
Mixture of SVMs for Face Class Modeling
Estimating Link Signatures with Machine Learning Algorithms
Basic machine learning background with Python scikit-learn
USPACOR: Universal Sparsity-Controlling Outlier Rejection
Data Mining Practical Machine Learning Tools and Techniques
Design of Hierarchical Classifiers for Efficient and Accurate Pattern Classification M N S S K Pavan Kumar Advisor : Dr. C. V. Jawahar.
What is Regression Analysis?
Presenter: Georgi Nalbantov
Linear Model Selection and Regularization
Creating Data Representations
Linear Model Selection and regularization
The Bias Variance Tradeoff and Regularization
Deep Learning for Non-Linear Control
Regression Usman Roshan.
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Support Vector Machine I
Data Mining, Machine Learning, Data Analysis, etc. scikit-learn
Chapter 7: Transformations
Microarray Data Set The microarray data set we are dealing with is represented as a 2d numerical array.
Junheng, Shengming, Yunsheng 11/09/2018
FOUNDATIONS OF BUSINESS ANALYTICS Introduction to Machine Learning
Support Vector Machines 2
About Data Analysis.
Advisor: Dr.vahidipour Zahra salimian Shaghayegh jalali Dec 2017
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

Chemometrics for Analysis of NIR Spectra on Pharmaceutical Oral Dosages William Welsh, Sastry Isukapalli, Rodolfo Romañach, Bozena Kohn-Michniak, Alberto Cuitino, Fernando Muzzio

Destructive vs Non-Destructive Testing Tablets – the most common drug delivery vehicle Dissolution tests: Key requirement for the development, registration, approval, and quality control of these tablets. Disadvantages of dissolution: Destructive, time consuming, expensive, and tedious. Need a fast, non-destructive and easy technique for tablet characterization. Near IR (NIR) spectroscopy serves this purpose.  Shearing Stress = a force that causes layers or parts of a substance to slide upon each other in opposite directions

Diffuse NIR Spectrometry Detector for Reflectance NIR spectra tablet Page 22, Handbook of Near Infrared Analysis: Light that is scattered from the sample toward a detector on the opposite side of the sample is said to be detected in transmission. Light that is scattered from the sample toward a detector on the same side of the sample is said to be detected in reflection. If the angle of reflection is equal to the angle of incidence, the reflection is said to be specular. Radiation reflected at all other angles is diffuse. The sum of the specularly and diffusely reflected radiation is the remitted radiation. For samples with a matte finish, especially powdered samples, the specularly reflected radiation is generally of low intensity. Hence measurement of the radiation from this type of sample is frequently known as diffuse reflection spectrometry. Detector for Transmission Use Chemometrics to Correlate NIR spectral features to sample properties

NIR Dataset Reflectance and Transmission Spectra Chemometric Models 47 samples: API (acetaminophen); lactose; MgStearate (1%) Two dependent variables: %API (10-30%), and Compaction Force (7 - 20 kN) Output data: Reflectance (R), Transmittance (T); Pooled R & T data, 1st and 2nd derivatives Chemometric Models Correlate %API and CF with NIR spectral data Standard Models: HCA, Regression Trees (CART), PLS Approaches for improved predictive ability: LASSO Regression, Ridge regression, Elastic Nets, Bayesian models

Chemometrics – Two General Approaches Unsupervised Principal Component Analysis (PCA) Hierarchical Cluster Analysis (HCA) Supervised Partial Least Squares (PLS) regression Classification and Regression Tree (CART) Support Vector Machine (SVM) Artificial Neural Network (ANN) LASSO, Ridge Regression, and Elastic Nets clustering group 1 group 2 active drug inactive classification Predicted Value Actual Value regression 5 5

Chemometrics Many Common Methods Many Uses Data Exploration & Clustering Classification & Discrimination Quantitative Prediction & Correlation Many Uses PCA HCA kNN ANN SVR Cooman’s plot OUTLIERS CLASS 1 CLASS 2 CLASS 1&2 ? Many Common Methods

Clustering with Pooled Reflectance and Transmission Data Unsupervised clustering Distinct clusters for low CF cases and for low %API cases Single sub-cluster for high CF-%API cases

Hierarchical Cluster Analysis Iterative agglomeration of clusters through “distance” similarity measures To estimate control variables from experimental conditions (CF, %API) Cluster Nodes Clustering of samples based on their spectra only moderately correlated to CF-%API groupings “Distance”

LASSO, Ridge, and Elastic Net Regressions Ordinary ‘least squares’ regression models Y ≈ ƒ(Xi) tend to overfit the data, leading to poor predictive ability. The problem is over-determined: many more variables (spectral data) than solutions (property values; samples). A process called Regularization can be introduced to prevent overfitting and to provide models that are predictive (low bias) & robust (low variance). Examples of this approach are LASSO (Least Absolute Shrinkage and Selection Operator) Ridge regression Elastic Nets

LASSO, Ridge, and Elastic Net Regressions Methods like LASSO penalize over-complex models, thereby leading to models with fewer terms (Occam’s Razor). Occam’s Razor: All other things being equal, simpler solutions are preferred over complex ones. Simpler models discern “hidden” structure, and may thus have better predictive performance. LASSO models are more easily interpretable; fewer variables.

All Data Pooled (including derivatives) Transmission Data only CART Regression Trees All Data Pooled (including derivatives) Transmission Data only Reflectance Data only % Active Ingredient % Active Ingredient % Active Ingredient predicted actual actual actual Compaction Force Compaction Force predicted

All Data Pooled (including derivatives) Transmission Data only LASSO Regression All Data Pooled (including derivatives) Transmission Data only Reflectance Data only LASSO Regression [Baseline T] 10 15 20 25 30 Actual % Active Ingredient Predicted % Active Ingredient LASSO Regression [SNV R] 10 15 20 25 30 Actual % Active Ingredient Predicted % Active Ingredient LASSO Regression [B/SNV/G1/G2: R + T combined] % A.I. 30 25 predicted Predicted % Active Ingredient 20 15 10 10 15 actual 20 30 25 Actual % Active Ingredient actual actual 8 10 12 14 16 18 Actual Compaction Force (kN) Predicted Compaction Force (kN) LASSO Regression [SNV R] 8 10 12 14 16 18 Actual Compaction Force (kN) Predicted Compaction Force (kN) LASSO Regression [B/SNV/G1/G2: R + T combined] 8 10 12 14 16 18 Actual Compaction Force (kN) Predicted Compaction Force (kN) LASSO Regression [Baseline T] Compaction Force predicted

LASSO Regression Simultaneous prediction of %API and Compaction Force 20 LASSO Regression [B/SNV/G1/G2: R only] o actual o predicted Actual [cross-validation] Predicted R only Reflectance Data 18 Simultaneous prediction of %API and Compaction Force Improvement when both reflectance and transmission data are used 16 14 Compaction Force Compaction Force (kN) 12 10 8 6 5 10 15 %API 20 25 30 35 5 10 15 20 25 30 35 % Active Ingredient 6 8 12 14 16 18 Compaction Force (kN) LASSO Regression [B/SNV/G1/G2: T only] Actual [cross-validation] % Active Ingredient 20 LASSO Regression [B/SNV/G1/G2: R + T combined] o actual o predicted Predicted T only o actual o predicted Actual [cross-validation] Predicted Pooled Data R and T Transmission Data 18 16 Compaction Force Compaction Force 14 Compaction Force (kN) 12 10 8 6 5 10 15 20 25 30 %API %API % Active Ingredient

Concluding Remarks Novel chemometrics methods can build relationships between non-destructive test data and biorelevant properties of tablets, including dissolution Pooling reflectance & transmission data advantageous LASSO regression delivers substantial model improvements, and identifies subset of information-rich variables (NIR features) Methods used in the analysis are easily scalable Thousands of dimensions/millions of rows Open source tool kits

Thank You Acknowledgments People Funding: NSF-AIR Rutgers: Bozena Michniak-Kohn, Alberto Cuitino, Fernando Muzzio UPR-Mayaguez: Rodolfo J. Romañach Team at Rutgers’ ERC-Structured Organic Particulate Systems http://www.ercforsops.org/ Snowdon: Sastry Isukapalli Funding: NSF-AIR Thank You