LABEL CORRECTION AND EVENT DETECTION FOR ELECTRICITY DISAGGREGATION

Slides:



Advertisements
Similar presentations
Clustering. How are we doing on the pass sequence? Pretty good! We can now automatically learn the features needed to track both people But, it sucks.
Advertisements

Applications of one-class classification
WATERSENSE: WATER FLOW DISAGGREGATION USING MOTION SENSORS Vijay Srinivasan, John Stankovic, Kamin Whitehouse Department of Computer Science University.
An Approach to ECG Delineation using Wavelet Analysis and Hidden Markov Models Maarten Vaessen (FdAW/Master Operations Research) Iwan de Jong (IDEE/MI)
Some problems... Lens distortion  Uncalibrated structure and motion recovery assumes pinhole cameras  Real cameras have real lenses  How can we.
Data Mining Methodology 1. Why have a Methodology  Don’t want to learn things that aren’t true May not represent any underlying reality ○ Spurious correlation.
FixtureFinder: Discovering the Existence of Electrical and Water Fixtures Vijay Srinivasan*, John Stankovic, Kamin Whitehouse University of Virginia *(Currently.
An Overview of Machine Learning
Doorjamb: Unobtrusive Room-level Tracking of People in Homes using Doorway Sensors Timothy W. Hnat, Erin Griffiths, Ray Dawson, Kamin Whitehouse U of Virginia.
Low Complexity Keypoint Recognition and Pose Estimation Vincent Lepetit.
Hidden Markov Model based 2D Shape Classification Ninad Thakoor 1 and Jean Gao 2 1 Electrical Engineering, University of Texas at Arlington, TX-76013,
Robust Moving Object Detection & Categorization using self- improving classifiers Omar Javed, Saad Ali & Mubarak Shah.
Locally Constraint Support Vector Clustering
Statistics for the Social Sciences Psychology 340 Fall 2006 Review For Exam 1.
Robust Real Time Pattern Matching using Bayesian Sequential Hypothesis Testing Ofir PeleMichael Werman The Hebrew University of Jerusalem TexPoint fonts.
Distinctive image features from scale-invariant keypoints. David G. Lowe, Int. Journal of Computer Vision, 60, 2 (2004), pp Presented by: Shalomi.
Unsupervised Training and Clustering Alexandros Potamianos Dept of ECE, Tech. Univ. of Crete Fall
Student: Hsu-Yung Cheng Advisor: Jenq-Neng Hwang, Professor
Robust Real-Time Object Detection Paul Viola & Michael Jones.
Learning Hidden Markov Model Structure for Information Extraction Kristie Seymour, Andrew McCullum, & Ronald Rosenfeld.
Handwritten Character Recognition using Hidden Markov Models Quantifying the marginal benefit of exploiting correlations between adjacent characters and.
(ACM KDD 09’) Prem Melville, Wojciech Gryc, Richard D. Lawrence
INTRODUCTION  Sibilant speech is aperiodic.  the fricatives /s/, / ʃ /, /z/ and / Ʒ / and the affricatives /t ʃ / and /d Ʒ /  we present a sibilant.
Machine Learning CUNY Graduate Center Lecture 1: Introduction.
Zero Resource Spoken Term Detection on STD 06 dataset Justin Chiu Carnegie Mellon University 07/24/2012, JHU.
EADS DS / SDC LTIS Page 1 7 th CNES/DLR Workshop on Information Extraction and Scene Understanding for Meter Resolution Image – 29/03/07 - Oberpfaffenhofen.
Virtual Mirror for Fashion Retailing
EE 492 ENGINEERING PROJECT LIP TRACKING Yusuf Ziya Işık & Ashat Turlibayev Yusuf Ziya Işık & Ashat Turlibayev Advisor: Prof. Dr. Bülent Sankur Advisor:
Annealing Paths for the Evaluation of Topic Models James Foulds Padhraic Smyth Department of Computer Science University of California, Irvine* *James.
CPSC 502, Lecture 15Slide 1 Introduction to Artificial Intelligence (AI) Computer Science cpsc502, Lecture 16 Nov, 3, 2011 Slide credit: C. Conati, S.
Detection, Classification and Tracking in a Distributed Wireless Sensor Network Presenter: Hui Cao.
A Novel Local Patch Framework for Fixing Supervised Learning Models Yilei Wang 1, Bingzheng Wei 2, Jun Yan 2, Yang Hu 2, Zhi-Hong Deng 1, Zheng Chen 2.
ECE 8443 – Pattern Recognition ECE 8423 – Adaptive Signal Processing Objectives: Supervised Learning Resources: AG: Conditional Maximum Likelihood DP:
Computational Approaches for Biomarker Discovery SubbaLakshmiswetha Patchamatla.
Genotype Calling Matt Schuerman. Biological Problem How do we know an individual’s SNP values (genotype)? Each SNP can have two values (A/B) Each individual.
Unsupervised Mining of Statistical Temporal Structures in Video Liu ze yuan May 15,2011.
Robust Kernel Density Estimation by Scaling and Projection in Hilbert Space Presented by: Nacer Khalil.
Dimensionality Reduction in Unsupervised Learning of Conditional Gaussian Networks Authors: Pegna, J.M., Lozano, J.A., Larragnaga, P., and Inza, I. In.
Ensemble Methods Construct a set of classifiers from the training data Predict class label of previously unseen records by aggregating predictions made.
Machine Learning ICS 178 Instructor: Max Welling Supervised Learning.
1 Chapter 8: Model Inference and Averaging Presented by Hui Fang.
Face Recognition and Tracking for Human-Robot Interaction using PeCoFiSH Alex Eisner This material is based upon work supported by the National Science.
IEEE AI - BASED POWER SYSTEM TRANSIENT SECURITY ASSESSMENT Dr. Hossam Talaat Dept. of Electrical Power & Machines Faculty of Engineering - Ain Shams.
SUPERVISED AND UNSUPERVISED LEARNING Presentation by Ege Saygıner CENG 784.
CSCI 631 – Foundations of Computer Vision March 15, 2016 Ashwini Imran Image Stitching.
Tree and Forest Classification and Regression Tree Bagging of trees Boosting trees Random Forest.
Gaussian Mixture Model classification of Multi-Color Fluorescence In Situ Hybridization (M-FISH) Images Amin Fazel 2006 Department of Computer Science.
Voice Activity Detection Based on Sequential Gaussian Mixture Model Zhan Shen, Jianguo Wei, Wenhuan Lu, Jianwu Dang Tianjin Key Laboratory of Cognitive.
Machine Learning with Spark MLlib
ECG data classification with deep learning tools
A Personal Tour of Machine Learning and Its Applications
Data Mining, Neural Network and Genetic Programming
Traffic Sign Recognition Using Discriminative Local Features Andrzej Ruta, Yongmin Li, Xiaohui Liu School of Information Systems, Computing and Mathematics.
A Forest of Sensors: Using adaptive tracking to classify and monitor activities in a site Eric Grimson AI Lab, Massachusetts Institute of Technology
Mark Valovage1† and Maria Gini2
Outlier Processing via L1-Principal Subspaces
Motion Detection And Analysis
Shay Ben-Elazar, Gal Lavee, Noam Koenigstein,
Vijay Srinivasan Thomas Phan
Incremental Boosting Incremental Learning of Boosted Face Detector ICCV 2007 Unsupervised Incremental Learning for Improved Object Detection in a Video.
Fast and Robust Object Tracking with Adaptive Detection
Baselining PMU Data to Find Patterns and Anomalies
P-value Approach for Test Conclusion
SLOPE: A MATLAB Revival
Ensembles.
Supervised vs. unsupervised Learning
Deep Robust Unsupervised Multi-Modal Network
Finding Periodic Discrete Events in Noisy Streams
EE 492 ENGINEERING PROJECT
Sign Language Recognition With Unsupervised Feature Learning
Presentation transcript:

LABEL CORRECTION AND EVENT DETECTION FOR ELECTRICITY DISAGGREGATION Mark Valovage1 and Maria Gini2 1Ph.D. Candidate, University of Minnesota 2Professor of Computer Science, University of Minnesota

OVERVIEW Background Contributions Conclusions Electricity Disaggregation Contributions Training Sample Label Correction (Supervised Learning) Parameter-free Event Detection (Unsupervised Learning) Conclusions

ELECTRICITY DISAGGREGATION Annual electrical waste (United States) $70 billion/year. 1.1 gigatons pollution. Aggregate data “You used X KWh.” “Pay us $Y.” Disaggregated data Can reduce power consumption by 15%.

LABEL CORRECTION Background Contributions Conclusions Electricity Disaggregation Contributions Training Sample Label Correction (Supervised Learning) Parameter-free Event Detection (Unsupervised Learning) Conclusions

ELECTRICITY DISAGGREGATION LABELING ERRORS Limiting Assumptions Correctly labeled samples Isolated appliances Labels error sources Offsets Raw user errors Truncated shutdowns Contaminations/Not isolated

LABEL ERRORS IMPACT ON CLASSIFICATION Back Porch Lights Samples Mean Power Increase (W) Uncorrected Corrected Sample 1 229.7 331.8 Sample 2 145.04 333.15 Sample 3 144 330.74 Sample 4 143.11 N/A Mean 165.46 331.90 Std. Dev. 37.09 0.99 Simple Mean Classification (+/- 10%) True Positives 0/4 4/4 False Positives 3 8 False Positive Appliances (Power Phase 2) Dining Room Lights 344.84 Master Bathroom Lights 314.52 Bedroom 1 Lights 290.19 Bedroom 1 LCD TV 152.04 Simple Mean Classification (+/-2 std) True Positives 0/4 4/4 False Positives 8

BAYESIAN CHANGE DETECTION Detects abrupt changes in 1-dimensional data Probability based Tracks active sequence (highest probability) Identifies changepoint when active run sequence changes Advantages: Robust to noise Does not require parameter tuning BCD is a tool that detects abrupt changes in 1-dimensional data.

LABEL CORRECTION ALGORITHMS (1/2) Naïve Proximal (NaïveProx) Run Bayesian Change Detection on training sample Use closest changepoints to user on/off labels Does not account for domain-level properties

LABEL CORRECTION ALGORITHMS (2/2) Power-constrained Proximal (PCProx) Can be online or offline (OnlinePCProx/OfflinePCProx) Run Bayesian Change Detection on training sample Use closest changepoints to user on/off labels Ensure new on/off labels satisfy real-power constraints

LABEL CORRECTION DATASET Kaggle Belkin dataset 4 houses Real power sampled at 5 Hz 147 appliances 462 training samples Label errors have not been removed

LABEL CORRECTION RESULTS

EVENT DETECTION Background Contributions Conclusions Electricity Disaggregation Contributions Training Sample Label Correction (Supervised Learning) Parameter-free Event Detection (Unsupervised Learning) Conclusions

UNSUPERVISED LEARNING EVENT DETECTION No training samples Appliances not isolated Limitations of existing event detection methods Require pre-processing to reduce noise Parameters must be tuned

EVENT DETECTION COMPETING ALGORITHMS Modified Generalized Likelihood Ratio (mGLR) Change Detection Requires parameter tuning Dirichlet Process Gaussian Mixture Models (DPGMM) Identifies steady-state power levels through iterative clustering Marks events as changes between steady-states Reduces noise through median window filtering

EVENT DETECTION DATASETS BLUED (Building-Level fUlly-labeled dataset for Electricity Disaggregation) 1 house Power sampled at 60 Hz for 1 week REDD (Reference Electricity Disaggregation Dataset) 6 houses Power sampled at ~0.25 Hz

EVENT DETECTION PERFORMANCE METRICS Unweighted F-measure Power-weighted F-measure

EVENT DETECTION RESULTS (1/2) UNWEIGHTED POWER-WEIGHTED

EVENT DETECTION RESULTS (2/2) UNWEIGHTED POWER-WEIGHTED

CONCLUSIONS Introduced application of Bayesian change detection Label Correction (Supervised Learning) Event Detection (Unsupervised Learning) Advantages Improves accuracy of samples for supervised learning Eliminate the need for parameter tuning Perform competitively against existing methods Can be run in real time on inexpensive hardware

FUTURE WORK Label Correction Event Detection Improve correction accuracy Recover information from contaminated samples Event Detection Automatically compute baseline noise threshold, Improve appliance reconstruction algorithms

QUESTIONS LABEL CORRECTION AND EVENT DETECTION FOR ELECTRICITY DISAGGREGATION Additional materials available at www.cs.umn.edu/~valovage/AAMAS-2017 Mark Valovage Ph.D. Candidate, University of Minnesota Research Intern, SIFT Maria Gini Professor, Computer Science University of Minnesota That concludes my presentation. I will now take your questions.