Rating Systems Vs Machine Learning on the context of sports George Kyriakides, Kyriacos Talattinis, George Stefanides Department of Applied Informatics,

Slides:



Advertisements
Similar presentations
Florida International University COP 4770 Introduction of Weka.
Advertisements

Slides from: Doug Gray, David Poole
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
Ch. Eick: More on Machine Learning & Neural Networks Different Forms of Learning: –Learning agent receives feedback with respect to its actions (e.g. using.
also known as the “Perceptron”
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 15 Nov, 1, 2011 Slide credit: C. Conati, S.
2002/11/15Game Programming Workshop1 A Neural Network for Evaluating King Danger in Shogi Reijer Grimbergen Department of Information Science Saga University.
Alberto Trindade Tavares ECE/CS/ME Introduction to Artificial Neural Network and Fuzzy Systems.
A (very) brief introduction to multivoxel analysis “stuff” Jo Etzel, Social Brain Lab
By Andrew Finley. Research Question Is it possible to predict a football player’s professional based on collegiate performance? That is, is it possible.
Decision Tree Rong Jin. Determine Milage Per Gallon.
The back-propagation training algorithm
Data Mining Techniques Outline
© Prentice Hall1 DATA MINING Introductory and Advanced Topics Part II Margaret H. Dunham Department of Computer Science and Engineering Southern Methodist.
Evaluation of Results (classifiers, and beyond) Biplav Srivastava Sources: [Witten&Frank00] Witten, I.H. and Frank, E. Data Mining - Practical Machine.
Chapter 5 Data mining : A Closer Look.
CSCI 347 / CS 4206: Data Mining Module 04: Algorithms Topic 06: Regression.
An Exercise in Machine Learning
 The Weka The Weka is an well known bird of New Zealand..  W(aikato) E(nvironment) for K(nowlegde) A(nalysis)  Developed by the University of Waikato.
Artificial neural networks:
Presentation on Neural Networks.. Basics Of Neural Networks Neural networks refers to a connectionist model that simulates the biophysical information.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Artificial Neural Networks
© N. Kasabov Foundations of Neural Networks, Fuzzy Systems, and Knowledge Engineering, MIT Press, 1996 INFO331 Machine learning. Neural networks. Supervised.
Chapter 9 Neural Network.
ANNs (Artificial Neural Networks). THE PERCEPTRON.
Appendix B: An Example of Back-propagation algorithm
Artificial Neural Network Supervised Learning دكترمحسن كاهاني
NEURAL NETWORKS FOR DATA MINING
Chapter 7 Neural Networks in Data Mining Automatic Model Building (Machine Learning) Artificial Intelligence.
Classification / Regression Neural Networks 2
LINEAR CLASSIFICATION. Biological inspirations  Some numbers…  The human brain contains about 10 billion nerve cells ( neurons )  Each neuron is connected.
Data Mining: Classification & Predication Hosam Al-Samarraie, PhD. Centre for Instructional Technology & Multimedia Universiti Sains Malaysia.
Data Mining Practical Machine Learning Tools and Techniques Chapter 4: Algorithms: The Basic Methods Section 4.6: Linear Models Rodney Nielsen Many of.
Computer Go : A Go player Rohit Gurjar CS365 Project Presentation, IIT Kanpur Guided By – Prof. Amitabha Mukerjee.
Today Ensemble Methods. Recap of the course. Classifier Fusion
Artificial Neural Network Building Using WEKA Software
Jennifer Lewis Priestley Presentation of “Assessment of Evaluation Methods for Prediction and Classification of Consumer Risk in the Credit Industry” co-authored.
Multi-Layer Perceptron
What is Data Mining? process of finding correlations or patterns among dozens of fields in large relational databases process of finding correlations or.
Chapter 11 Statistical Techniques. Data Warehouse and Data Mining Chapter 11 2 Chapter Objectives  Understand when linear regression is an appropriate.
Neural Network Implementation of Poker AI
CSE 5331/7331 F'07© Prentice Hall1 CSE 5331/7331 Fall 2007 Machine Learning Margaret H. Dunham Department of Computer Science and Engineering Southern.
Project 2: Classification Using Genetic Programming Kim, MinHyeok Biointelligence laboratory Artificial.
Fundamentals of Artificial Neural Networks Chapter 7 in amlbook.com.
1 Statistics & R, TiP, 2011/12 Neural Networks  Technique for discrimination & regression problems  More mathematical theoretical foundation  Works.
Assignments CS fall Assignment 1 due Generate the in silico data set of 2sin(1.5x)+ N (0,1) with 100 random values of x between.
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
***Classification Model*** Hosam Al-Samarraie, PhD. CITM-USM.
Artificial Neural Networks (ANN). Artificial Neural Networks First proposed in 1940s as an attempt to simulate the human brain’s cognitive learning processes.
Neural Networks References: “Artificial Intelligence for Games” "Artificial Intelligence: A new Synthesis"
Announcements 1. Textbook will be on reserve at library 2. Topic schedule change; modified reading assignment: This week: Linear discrimination, evaluating.
Supervised Learning – Network is presented with the input and the desired output. – Uses a set of inputs for which the desired outputs results / classes.
An Evolutionary Algorithm for Neural Network Learning using Direct Encoding Paul Batchis Department of Computer Science Rutgers University.
Data Mining: Concepts and Techniques1 Prediction Prediction vs. classification Classification predicts categorical class label Prediction predicts continuous-valued.
Machine Learning Usman Roshan Dept. of Computer Science NJIT.
Machine Learning Supervised Learning Classification and Regression
Chapter 7. Classification and Prediction
Artificial neural networks:
Classification with Perceptrons Reading:
with Daniel L. Silver, Ph.D. Christian Frey, BBA April 11-12, 2017
NBA Draft Prediction BIT 5534 May 2nd 2018
Collaborative Filtering Matrix Factorization Approach
Discriminative Frequent Pattern Analysis for Effective Classification
Aleysha Becker Ece 539, Fall 2018
Machine Learning: Lecture 4
ECE/CS/ME 539 Artificial Neural Networks Final Project
Somi Jacob and Christian Bach
Machine Learning: UNIT-2 CHAPTER-1
Identifying Severe Weather Radar Characteristics
Presentation transcript:

Rating Systems Vs Machine Learning on the context of sports George Kyriakides, Kyriacos Talattinis, George Stefanides Department of Applied Informatics, University Of Macedonia

Aim of the paper Study the performance of linear algebra rating systems and machine learning methods. Evaluate the accuracy of each method. Evaluate the quality of the predictions.

Structure of the presentation Clarify what is rating and ranking. Explain linear algebra rating systems used in this paper. Explain machine learning methods used in this paper. Compare the rating systems and machine learning methods by predicting soccer games (English Premier League). Conclusions. Possible future work.

Related research Nivard van Wijk uses a Pseudo Least-Squares Estimator to predict soccer matches. Paul Kvam and Joel S. Sokol use logistic regression and markov chains to predict basketball matches. Keeneth Massey uses linear algebra to rank basketball and football teams. Search engines use machine learning to rank search results. No comparison between machine learning and linear algebra has been made. Studies on accuracy have been conducted for machine learning and linear algebra independently, but never for profitability of the methods.

Rating The evaluation of an object, based on some desirable criteria. For example, a car may have a 1 to 5 stars NCAP safety rating.

Ranking A relationship between a set of objects, such that for any two items, one is ranked higher than the other and is consequently better, assuming that the object ranked highest is the best. Rating can be used to rank a set of objects. Sorting a vector containing ratings we are effectively ranking the elements.

Ranking

Rating Systems Many systems have been proposed: Massey Colley Markov Keener Google Page Rank mHITS Elo

Methods used in this paper Massey Colley mHITS

Massey Method Proposed by Kenneth Massey in 1997 to rank NCAA (National Collegiate Athletic Association) teams. It uses a linear least squares regression to solve a system of linear equations. A rating vector is calculated, where each entry corresponds to the rating of the team.

Massey Method (2) First, the matrix M is generated, where M ij is the number of games teams i and j played, multiplied by -1 and M ii is the total number of games team i has played. The second step is to calculate the vector p, where p i is the total number of points scored by team i, minus the points that were scored against the team. The final step is to solve the system of equations: Mr= p where r is the vector of the ratings, so if r i > r j team i is better than team j.

Colley Method Proposed by astrophysicist Dr.Wesley Colley in Variation of a simple method used to rank teams, which calculated the win ratio of each team (wins divided by total games). A system of linear equations is solved in order to find a rating vector.

Colley Method (2)

mHITS(Offence-Defense model)

mHITS(Offence-Defense model) (2)

Machine Learning

Methods used in this paper Decision Trees Artificial Neural Networks Random Forests

Artificial Neural Networks Networks of nodes which accept inputs and produce an output based on an activation function. Nodes’ connections are weighted. Usually organized in layers. Each layer is a group of nodes not connected to any node of the same group. Input data is presented to the input layer, “hidden” layers process the data and the presentation layer outputs the results.

Artificial Neural Networks Multilayer Perceptron, which uses back propagation was used in the paper. Learning rate:0.3 Momentum:0.2 Hidden layers:3 Epochs: Fold Cross-Validation

Decision Tree Learning Uses a decision tree to classify/predict. Also known as classification/regression trees. Leaves represent class labels. Branches split the data into appropriate sets. Different algorithms use different criteria to split the data.

C4.5 Uses Information Gain to measure the quality of the split. Confidence factor: 0.7 Unpruned 10-Fold Cross-Validation

Random Forest A multitude of decision trees is generated Each tree is trained independently, using a different subset of the data Input data is presented to all trees The class that the highest percentage of trees produce is the output

Soccer

Ternary Result

Predictions Hindsight Foresight Betting

Implementation Weka Open-source GNU GPL Data mining software Implemented in Java Association rules, Classification, Clustering Custom Java code mHits Colley Massey Jamma Numerical linear algebra library

Hindsight Accuracy Method Season NeuralTreesForestMHITSColleyMassey 2008 / %71.32%97.11%51.37%53.62%38.90% 2009 / %72.11%94.74%55.20%56.54%48.67% 2010 / %60.79%94.74%45.07%47.05%42.18% 2011 / %58.42%96.32%54.95%55.46%47.68% 2012 / %55.00%95.79%50.88%51.88%42.95%

Hindsight (2)

Confusion Matrices Confusion matrices allow the visualization of the performance of an algorithm It indicates the quality of the predictions 95% Accuracy: Class Actual AB A955 B 0 0

Confusion Matrices (Hindsight) Class Actual LossWinDraw Loss8722 Win51862 Draw1887 Random Forest Decision TreeANN 94.74% 72.11% 56.32% Class Actual LossWinDraw Loss65215 Win Draw Class Actual LossWinDraw Loss Win Draw115629

Foresight Accuracy Method Season NeuralTreesForestMHITSColleyMassey 2008 / %49.47%50.00%56.97%48.78%36.25% 2009 / %52.63%38.42%53.05%48.17%42.10% 2010 / % 41.58%46.63%42.78%41.06% 2011 / % 37.89%53.35%46.90%45.83% 2012 / %48.74%48.42%52.40%47.70%40.54%

Foresight (2)

Confusion Matrices (Foresight) Random Forest Decision Tree ANN 37.89% 46.84% 46.84% Class Actual LossDrawWin Loss Draw Win Class Actual LossDrawWin Loss Draw13624 Win Class Actual LossDrawWin Loss16433 Draw19024 Win21073

Making a profit Quantity VS Quality

Making a profit Quantity. 90% Accuracy – 1.05 average booking odds In 100 games, betting 1 unit each time: Win 90*0.05 = 4.5 Lose 10 Net profit = = -5.5

Making a profit Quality 60% Accuracy – 2.5 average booking odds In 100 games, betting 1 unit each time: Win 60*1.5 = 90 Lose 40 Net profit = = 50

Betting The average odds of the 5 biggest online booking companies money units starting capital. 50 units betted each time. Each method chose the outcome it thought would occur.

Foresight Betting Money Sum Method Season NeuralTreesForestMHITSColleyMassey 2008 / / / / /

Foresight Betting Net Profits sum

Conclusions Machine learning proved to be superior in hindsight predictions and prediction quality. Draws are the most difficult to predict. mHITS is the best in foresight prediction accuracy, but not in quality. Neural Networks are the most profitable of all. Random Forests constructed the best hindsight models.

Future Work Other Machine Learning Methods Portfolio management for betting section Other sports, where a draw is much less likely to be the outcome

Thank you very much for your attention!