M Machine Learning F# and Accord.net. Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher.

Slides:



Advertisements
Similar presentations
Neural Networks: Learning
Advertisements

A Brief Overview of Neural Networks By Rohit Dua, Samuel A. Mulder, Steve E. Watkins, and Donald C. Wunsch.
Slides from: Doug Gray, David Poole
Neural networks Introduction Fitting neural networks
Linear Regression.
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Neural Networks: Backpropagation algorithm Data Mining and Semantic Web University of Belgrade School of Electrical Engineering Chair of Computer Engineering.
Logistic Regression.
Supervised Learning Recap
Lecture 13 – Perceptrons Machine Learning March 16, 2010.
Classification and Prediction: Regression Via Gradient Descent Optimization Bamshad Mobasher DePaul University.
Machine Learning Neural Networks
Lecture 14 – Neural Networks
Supervised learning 1.Early learning algorithms 2.First order gradient methods 3.Second order gradient methods.
Lecture 29: Optimization and Neural Nets CS4670/5670: Computer Vision Kavita Bala Slides from Andrej Karpathy and Fei-Fei Li
Radial Basis Functions
Data mining and statistical learning - lecture 11 Neural networks - a model class providing a joint framework for prediction and classification  Relationship.
Review Rong Jin. Comparison of Different Classification Models  The goal of all classifiers Predicating class label y for an input x Estimate p(y|x)
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
Review of Lecture Two Linear Regression Normal Equation
Collaborative Filtering Matrix Factorization Approach
Machine Learning1 Machine Learning: Summary Greg Grudic CSCI-4830.
Backpropagation An efficient way to compute the gradient Hung-yi Lee.
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
CSC 4510 – Machine Learning Dr. Mary-Angela Papalaskari Department of Computing Sciences Villanova University Course website:
Model representation Linear regression with one variable
Andrew Ng Linear regression with one variable Model representation Machine Learning.
CS 478 – Tools for Machine Learning and Data Mining Backpropagation.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 9: Ways of speeding up the learning and preventing overfitting Geoffrey Hinton.
CS 782 – Machine Learning Lecture 4 Linear Models for Classification  Probabilistic generative models  Probabilistic discriminative models.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Back-Propagation Algorithm AN INTRODUCTION TO LEARNING INTERNAL REPRESENTATIONS BY ERROR PROPAGATION Presented by: Kunal Parmar UHID:
An Artificial Neural Network Approach to Surface Waviness Prediction in Surface Finishing Process by Chi Ngo ECE/ME 539 Class Project.
M Machine Learning F# and Accord.net.
COSC 4426 AJ Boulay Julia Johnson Artificial Neural Networks: Introduction to Soft Computing (Textbook)
Neural Networks The Elements of Statistical Learning, Chapter 12 Presented by Nick Rizzolo.
Bump Hunting The objective PRIM algorithm Beam search References: Feelders, A.J. (2002). Rule induction by bump hunting. In J. Meij (Ed.), Dealing with.
WEEK 2 SOFT COMPUTING & MACHINE LEARNING YOSI KRISTIAN Gradient Descent for Linear Regression.
Chapter 11 – Neural Nets © Galit Shmueli and Peter Bruce 2010 Data Mining for Business Intelligence Shmueli, Patel & Bruce.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
CMPS 142/242 Review Section Fall 2011 Adapted from Lecture Slides.
Multinomial Regression and the Softmax Activation Function Gary Cottrell.
Learning with Neural Networks Artificial Intelligence CMSC February 19, 2002.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Machine Learning Supervised Learning Classification and Regression
Big data classification using neural network
Fall 2004 Backpropagation CS478 - Machine Learning.
Chapter 7. Classification and Prediction
Deep Feedforward Networks
Artificial Neural Networks
Machine Learning & Deep Learning
Learning with Perceptrons and Neural Networks
Table 1. Advantages and Disadvantages of Traditional DM/ML Methods
A Simple Artificial Neuron
CSE 473 Introduction to Artificial Intelligence Neural Networks
Announcements HW4 due today (11:59pm) HW5 out today (due 11/17 11:59pm)
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
CSE P573 Applications of Artificial Intelligence Neural Networks
Collaborative Filtering Matrix Factorization Approach
Neural Networks Geoff Hulten.
Neural networks (1) Traditional multi-layer perceptrons
Artificial Intelligence 10. Neural Networks
Machine Learning Algorithms – An Overview
Backpropagation David Kauchak CS159 – Fall 2019.
Linear regression with one variable
PYTHON Deep Learning Prof. Muhammad Saeed.
Patterson: Chap 1 A Review of Machine Learning
Presentation transcript:

m Machine Learning F# and Accord.net

Alena Dzenisenka Software architect at Luxoft Poland Member of F# Software Foundation Board of Trustees Researcher in the field of mathematical theoretical possible in modern programming concepts Speaker and active software engineering community lenadroid

Machine Learning

Why machine learning? What is the data? How? Questions

Data Questions.

Data reality :\

Path to grasping machine learning and data science…

Contents Multiple Linear  Regression Logistic Regression  Classification Neural Networks  Classification K Means  Clustering What’s next?

F# for machine learning and data science!

Linear Regression

MATH

How to predict? 1.Make a guess. 2.Measure how wrong the guess is. 3.Fix the error.

Make a guess!

Make a guess? What does it mean?... Hypothesis /guess : weights

Find out our mistake… Cost function/ Mistake function: … and minimize it:

Mistake function looks like… Global minimums

How to reduce the mistake? Update each slope parameter until Mistake Function minimum is reached: Simultaneously Alpha Learning rate Derivative Direction of moving

Fix the error

Multiple Linear Regression X [ ] – Predictors: Statistical data about bike rentals for previous years or months. Y – Output: Amount of bike rentals we should expect today or some other day in the future. * Y is not nominal, here it’s numerical continuous range.

Make a guess!

Fix the error

Multiple linear regression: Bike rentals demand “Talk is cheap. Show me the code.”

What to remember? 1.Simplest regression algorithm 2.Very fast, runs in constant time 3.Good at numerical data with lots of features 4. Output from numerical continuous range 5. Linear hypothesis 6. Uses gradient descent Linear Regression

Logistic Regression

Hypothesis function Estimated probability that Y = 1 on input X

Mistake function Mistake function is the cost for a single training data example h(x)

Full mistake function 1. Uses the principle of maximum likelihood estimation. 2. We minimize it same way as with Linear Regression

“Talk is cheap. Show me the code.” Logistic Regression Classification Example

What to remember? Classification algorithm Classification algorithm Output is the binary value, either 1 or 0 Output is the binary value, either 1 or 0 Relatively small number of predictors Relatively small number of predictors Uses logistics function for hypothesis Uses logistics function for hypothesis Has the cost function that is convex Has the cost function that is convex Uses gradient descent for correcting the mistake Uses gradient descent for correcting the mistake Logistic Regression

At this point…

Machine Learning What society thinks I do…What other programmers think I do…

What I really do is…

Neural Networks

Neuron

Neural Networks

Layer transitions! InputWeightsActivation function Activation unit (value which is computed and outputted by the node)

Activation function? Step activation function Used in binary classification Sigmoid activation function Uses logistic function to perform the activation

Forward propagation x x x x a a a a a a a X = X X = a 1 X = a 2 Hypothetic Outputs! Activation Computational Nodes! Predictors

Multi class? OR x x x a a a a a x x x a a a a a

Cost function? x x x Looks difficult? No, it’s really not so complicated… 1.Outputs the vector with size == number of output classes 2.Similar to Logistic Regression cost function 3.Finds the mistake of our hypothesis and real outputs

Minimize the error! x x x x Error value And then… recalculate weights!

“Talk is cheap. Show me the code.” Neural Networks Classification Example.

What to remember? Extremely powerful Extremely powerful Slow learning / Fast prediction Slow learning / Fast prediction Big number of predictors Big number of predictors Complex hierarchical relationships Complex hierarchical relationships No need to understand the data No need to understand the data Wide applications Wide applications Neural Networks

K-Means

Clustering

Birth-death percentage rates

… with clustering K = 2

… with clustering K = 3

… with clustering K = 4

What’s next?

I’m

Thank you!