GECCO 2013 Industrial Competition Computer Engineering Lab, School of Electrical and IT Engineering Rommel Vergara.

Slides:



Advertisements
Similar presentations
Time series modelling and statistical trends
Advertisements

GECCO 2013 Industrial Competition Computer Engineering Lab, School of Electrical and IT Engineering Farzad Noorian.
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
My name is Dustin Boswell and I will be presenting: Ensemble Methods in Machine Learning by Thomas G. Dietterich Oregon State University, Corvallis, Oregon.
Pattern Recognition and Machine Learning: Kernel Methods.
Hazırlayan NEURAL NETWORKS Least Squares Estimation PROF. DR. YUSUF OYSAL.
Vision Based Control Motion Matt Baker Kevin VanDyke.
A neural network based several-hour- ahead electric load forecasting using similar days approach Paras Mandal, Tomonobu Senjyu, Naomitsu Urasaki, Toshihisa.
Overview of Adaptive Multi-Rate Narrow Band (AMR-NB) Speech Codec
PROVIDING DISTRIBUTED FORECASTS OF PRECIPITATION USING A STATISTICAL NOWCAST SCHEME Neil I. Fox and Chris K. Wikle University of Missouri- Columbia.
Class 15: Tuesday, Nov. 2 Multiple Regression (Chapter 11, Moore and McCabe).
Paper Title Your Name CMSC 838 Presentation. CMSC 838T – Presentation Motivation u Problem paper is trying to solve  Characteristics of problem  … u.
Ensemble Learning: An Introduction
Induction of Decision Trees
Bioinformatics Challenge  Learning in very high dimensions with very few samples  Acute leukemia dataset: 7129 # of gene vs. 72 samples  Colon cancer.
Elaine Martin Centre for Process Analytics and Control Technology University of Newcastle, England The Conjunction of Process and.
Face Processing System Presented by: Harvest Jang Group meeting Fall 2002.
1 Statistical Tools for Multivariate Six Sigma Dr. Neil W. Polhemus CTO & Director of Development StatPoint, Inc.
© Janice Regan, CMPT 128, Feb CMPT 128: Introduction to Computing Science for Engineering Students Running Time Big O Notation.
ES 128 Computational Solid and Structural Mechanics
Gabriel Kronberger and Michael Kommenda GECCO 2013, Industrial Challenge, 2nd place Contact: Gabriel Kronberger Heuristic and Evolutionary Algorithms Lab.
Introduction to variable selection I Qi Yu. 2 Problems due to poor variable selection: Input dimension is too large; the curse of dimensionality problem.
M ULTIFRAME P OINT C ORRESPONDENCE By Naseem Mahajna & Muhammad Zoabi.
Facial animation retargeting framework using radial basis functions Tamás Umenhoffer, Balázs Tóth Introduction Realistic facial animation16 is a challenging.
22C:082:001 Human-Computer Interaction. Fall Copyright © 2013 Juan Pablo Hourcade. 1 Group Project Phase 1.
Kalman Filter (Thu) Joon Shik Kim Computational Models of Intelligence.
Time Series Data Analysis - I Yaji Sripada. Dept. of Computing Science, University of Aberdeen2 In this lecture you learn What are Time Series? How to.
Department of Electrical Engineering, Southern Taiwan University Robotic Interaction Learning Lab 1 The optimization of the application of fuzzy ant colony.
Lecture 10: 8/6/1435 Machine Learning Lecturer/ Kawther Abas 363CS – Artificial Intelligence.
Performance Prediction for Random Write Reductions: A Case Study in Modelling Shared Memory Programs Ruoming Jin Gagan Agrawal Department of Computer and.
Introduction to Machine Learning Supervised Learning 姓名 : 李政軒.
MA3264 Mathematical Modelling Lecture 3 Model Fitting.
Concept learning, Regression Adapted from slides from Alpaydin’s book and slides by Professor Doina Precup, Mcgill University.
CSC321 Introduction to Neural Networks and Machine Learning Lecture 3: Learning in multi-layer networks Geoffrey Hinton.
Time Series Analysis and Forecasting. Introduction to Time Series Analysis A time-series is a set of observations on a quantitative variable collected.
Introduction to Operations Research. MATH Mathematical Modeling 2 Introduction to Operations Research Operations research/management science –Winston:
MODEL FITTING jiangyushan. Introduction The goal of model fitting is to choose values for the parameters in a function to best describe a set of data.
Prediction of Protein Binding Sites in Protein Structures Using Hidden Markov Support Vector Machine.
1 Academic Writing II Spring 2013 Meet twice a week Monday, Tuesday Time: Monday 8:00-9:50 (3-106) Tuesday 2:00-3:50 (3- 422)
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Nonlinear State Estimation
CS Statistical Machine learning Lecture 12 Yuan (Alan) Qi Purdue CS Oct
Reinforcement Learning AI – Week 22 Sub-symbolic AI Two: An Introduction to Reinforcement Learning Lee McCluskey, room 3/10
Machine Learning in CSC 196K
MITM613 Wednesday [ 6:00 – 9:00 ] am 1 st week. Good evening …. Every body.
Monday October 5, 2015 WARM-UP: QUESTION #1 WEEK OF OCT 5 QUESTION - DESCRIBE HOW THE RESPIRATORY SYSTEM AND THE CIRCULATORY SYSTEM WORK TOGETHER. Unit.
State-Space Recursive Least Squares with Adaptive Memory College of Electrical & Mechanical Engineering National University of Sciences & Technology (NUST)
Introduction It had its early roots in World War II and is flourishing in business and industry with the aid of computer.
LOAD FORECASTING. - ELECTRICAL LOAD FORECASTING IS THE ESTIMATION FOR FUTURE LOAD BY AN INDUSTRY OR UTILITY COMPANY - IT HAS MANY APPLICATIONS INCLUDING.
UNIT PLAN: FROM ATOMS TO POLYMERS Father Judge High School Grade 9 Physical Science Mr. A. Gutzler.
2011 Data Mining Industrial & Information Systems Engineering Pilsung Kang Industrial & Information Systems Engineering Seoul National University of Science.
Machine Learning Homework Gaining familiarity with Weka, ML tools and algorithms.
CPSC 121: Models of Computation REVIEW. Course Learning Outcomes You should be able to: – model important problems so that they are easier to discuss,
1 Dongheng Sun 04/26/2011 Learning with Matrix Factorizations By Nathan Srebro.
Fuzzy Systems Simulation Session 5
11-1 Empirical Models Many problems in engineering and science involve exploring the relationships between two or more variables. Regression analysis.
Action-Grounded Push Affordance Bootstrapping of Unknown Objects
R. E. Wyllys Copyright 2003 by R. E. Wyllys Last revised 2003 Jan 15
CSE 4705 Artificial Intelligence
Modeling Bike Availability in a Bike-Sharing System Using Machine Learning Hesham Rakha, Ph.D., P.Eng. Samuel Reynolds Pritchard Professor of Engineering,
Machine Learning Week 1.
1) Project Outline/ Assignment Sheet 2) Project Grading Rubric
Unfolding Problem: A Machine Learning Approach
Introduction to Weather
The Science of Predicting Outcome
Analysis for Predicting the Selling Price of Apartments Pratik Nikte
Artificial Intelligence 10. Neural Networks
Aug. 27th -Monday—BELLWORK Week 4
Machine Learning in Business John C. Hull
Credit Card Fraudulent Transaction Detection
Presentation transcript:

GECCO 2013 Industrial Competition Computer Engineering Lab, School of Electrical and IT Engineering Rommel Vergara

Introduction ›Machine Learning Algorithm used: Kernel Recursive Least Squares -Used an open-source C++ library, dlib ( which has a full implementation of the KRLS algorithm. ›Challenges: -Aperiodic and missing data samples -Selection of the feature set to describe the Temperature and Humidity values. 2

Data Preprocessing 3 ›All data sets contained missing data and were aperiodic in nature. ›SOLUTION: -Missing Data: Missing data occurrences were linearly approximated between their adjacent available data points. -Aperiodic Nature: Data was standardized into 10 minutes intervals (as required by the format for the competition)

Data Preprocessing 4 ›Data set was also narrowed to contain the target weekdays only: Tuesday, Wednesday and Thursday. ›Able to capture more accurately minor and specific changes that occur on these particular weekdays.

Data Preprocessing 5 Tue Wed Thu

Feature Set Selection ›The ‘Feature Set’ is a selection of inputs that contribute and explain the output ›Important to choose the feature set to describe the outputs we are predicting ›There are many ways to represent the feature set to describe an output. ›Each feature set is represented as a column vector and is fed into the KRLS algorithm. 6

Temperature Feature Set ›The temperature feature set chosen contained 145 values -Current weather value -144 temperature values 10 minute rolling window of the previous and current weekday lagged by 1 week. ›REASONING: -Remove noise that exists in other weekdays -Allow KRLS to concentrate and focus more on the specific weekdays that needed to be judged in the competition. 7

Temperature Feature Set ›Example: -To predict the following data point: -20/02/ :00 (Wednesday) -The following temperature values were used: -12/02/ :00 (Tuesday) to -12/02/ :50 (Tuesday) -To predict the next data point: -20/02/ :10 (Wednesday) -The following temperature values were used (10 minute rolling window): -12/02/ :10 (Tuesday) to -13/02/ :00 (Wednesday) 8

Temperature Results 9

Humidity Feature Set ›The humidity feature set chosen contained 2 values -Current weather value -Predicted KRLS temperature value ›This proved to be ineffective, providing an RMSE of 0.12 in the competition. ›CHALLENGE: Humidity data set was aperiodic and observed to have discrete-like behaviour. ›IMPROVEMENT: Given a more continuous data set, I would have chosen the same technique as the Temperature Feature Set, which is to take a rolling window of previous and current weekday humidity values lagged by 1 week. 10

Humidity Results 11

References ›“The Kernel Recursive Least Squares Algorithm” (2003), Yaakov Engel, Shie Mannor, Ron Meir. ›dlib library: ›Contact: -Rommel Vergara (University Of Sydney) 12