Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University.

Slides:



Advertisements
Similar presentations
WATERSENSE: WATER FLOW DISAGGREGATION USING MOTION SENSORS Vijay Srinivasan, John Stankovic, Kamin Whitehouse Department of Computer Science University.
Advertisements

Imbalanced data David Kauchak CS 451 – Fall 2013.
Sensor-Based Abnormal Human-Activity Detection Authors: Jie Yin, Qiang Yang, and Jeffrey Junfeng Pan Presenter: Raghu Rangan.
Yasuhiro Fujiwara (NTT Cyber Space Labs)
© Tan,Steinbach, Kumar Introduction to Data Mining 4/18/ Other Classification Techniques 1.Nearest Neighbor Classifiers 2.Support Vector Machines.
FixtureFinder: Discovering the Existence of Electrical and Water Fixtures Vijay Srinivasan*, John Stankovic, Kamin Whitehouse University of Virginia *(Currently.
1 Prediction-based Strategies for Energy Saving in Object Tracking Sensor Networks Yingqi Xu, Wang-Chien Lee Proceedings of the 2004 IEEE International.
1 Learning to Detect Objects in Images via a Sparse, Part-Based Representation S. Agarwal, A. Awan and D. Roth IEEE Transactions on Pattern Analysis and.
Distributed Microsystems Laboratory: Developing Microsystems that Make Sense Sensor Validation Techniques Sponsoring Agency: Center for Process Analytical.
Rodent Behavior Analysis Tom Henderson Vision Based Behavior Analysis Universitaet Karlsruhe (TH) 12 November /9.
UNIVERSITY OF SOUTH CAROLINA Department of Computer Science and Engineering On-line Alert Systems for Production Plants A Conflict Based Approach.
Ensemble Learning: An Introduction
An Experimental Evaluation on Reliability Features of N-Version Programming Xia Cai, Michael R. Lyu and Mladen A. Vouk ISSRE’2005.
Soft. Eng. II, Spr. 2002Dr Driss Kettani, from I. Sommerville1 CSC-3325: Chapter 9 Title : Reliability Reading: I. Sommerville, Chap. 16, 17 and 18.
A Cross Layer Approach for Power Heterogeneous Ad hoc Networks Vasudev Shah and Srikanth Krishnamurthy ICDCS 2005.
An Intelligent Cache System with Hardware Prefetching for High Performance Jung-Hoon Lee; Seh-woong Jeong; Shin-Dug Kim; Weems, C.C. IEEE Transactions.
Neural Networks (NN) Ahmad Rawashdieh Sa’ad Haddad.
05/06/2005CSIS © M. Gibbons On Evaluating Open Biometric Identification Systems Spring 2005 Michael Gibbons School of Computer Science & Information Systems.
seminar on Intrusion detection system
PSYC512: Research Methods PSYC512: Research Methods Lecture 14 Brian P. Dyre University of Idaho.
On Comparing Classifiers: Pitfalls to Avoid and Recommended Approach Published by Steven L. Salzberg Presented by Prakash Tilwani MACS 598 April 25 th.
Non-functional requirements
Selective Sampling on Probabilistic Labels Peng Peng, Raymond Chi-Wing Wong CSE, HKUST 1.
EE392J Final Project, March 20, Multiple Camera Object Tracking Helmy Eltoukhy and Khaled Salama.
Real-Time Odor Classification Through Sequential Bayesian Filtering Javier G. Monroy Javier Gonzalez-Jimenez
Evaluating Classifiers
Software Reliability Categorising and specifying the reliability of software systems.
©Ian Sommerville 2004Software Engineering, 7th edition. Chapter 24 Slide 1 Critical Systems Validation 1.
A Multivariate Biomarker for Parkinson’s Disease M. Coakley, G. Crocetti, P. Dressner, W. Kellum, T. Lamin The Michael L. Gargano 12 th Annual Research.
Extracting Places and Activities from GPS Traces Using Hierarchical Conditional Random Fields Yong-Joong Kim Dept. of Computer Science Yonsei.
Alert Correlation for Extracting Attack Strategies Authors: B. Zhu and A. A. Ghorbani Source: IJNS review paper Reporter: Chun-Ta Li ( 李俊達 )
Design of Experiments Chapter 21.
Snooping based privacy attacks based on transmission timing and wireless fingerprinting Master’s project presentation Vijay Srinivasan University of Virginia.
Leader Election Algorithms for Mobile Ad Hoc Networks Presented by: Joseph Gunawan.
Data Analysis 1 Mark Stamp. Topics  Experimental design o Training set, test set, n-fold cross validation, thresholding, imbalance, etc.  Accuracy o.
Mining and Analysis of Control Structure Variant Clones Guo Qiao.
Evaluating What’s Been Learned. Cross-Validation Foundation is a simple idea – “ holdout ” – holds out a certain amount for testing and uses rest for.
Experiments in Machine Learning COMP24111 lecture 5 Accuracy (%) A BC D Learning algorithm.
One-class Training for Masquerade Detection Ke Wang, Sal Stolfo Columbia University Computer Science IDS Lab.
1 Webcam Mouse Using Face and Eye Tracking in Various Illumination Environments Yuan-Pin Lin et al. Proceedings of the 2005 IEEE Y.S. Lee.
Data Mining Algorithms for Large-Scale Distributed Systems Presenter: Ran Wolff Joint work with Assaf Schuster 2003.
Exploiting Context Analysis for Combining Multiple Entity Resolution Systems -Ramu Bandaru Zhaoqi Chen Dmitri V.kalashnikov Sharad Mehrotra.
Accuracy Based Generation of Thermodynamic Properties for Light Water in RELAP5-3D 2010 IRUG Meeting Cliff Davis.
1 Context-dependent Product Line Practice for Constructing Reliable Embedded Systems Naoyasu UbayashiKyushu University, Japan Shin NakajimaNational Institute.
©2009 Mladen Kezunovic. Improving Relay Performance By Off-line and On-line Evaluation Mladen Kezunovic Jinfeng Ren, Chengzong Pang Texas A&M University,
1 Test Selection for Result Inspection via Mining Predicate Rules Wujie Zheng
Stable Multi-Target Tracking in Real-Time Surveillance Video
K. Kolomvatsos 1, C. Anagnostopoulos 2, and S. Hadjiefthymiades 1 An Efficient Environmental Monitoring System adopting Data Fusion, Prediction & Fuzzy.
A Content-Based Approach to Collaborative Filtering Brandon Douthit-Wood CS 470 – Final Presentation.
Tell Me What You See and I will Show You Where It Is Jia Xu 1 Alexander G. Schwing 2 Raquel Urtasun 2,3 1 University of Wisconsin-Madison 2 University.
Online Multiple Kernel Classification Steven C.H. Hoi, Rong Jin, Peilin Zhao, Tianbao Yang Machine Learning (2013) Presented by Audrey Cheong Electrical.
KAIS T SIGF : A Family of Configurable, Secure Routing Protocols for WSNs Sep. 20, 2007 Presented by Kim, Chano Brian Blum, Tian He, Sang Son, Jack Stankovic.
Efficient Data Compression in Location Based Services Yuni Xia, Yicheng Tu, Mikhail Atallah, Sunil Prabhakar.
An introduction to Fault Detection in Logic Circuits By Dr. Amin Danial Asham.
COT6930 Course Project. Outline Gene Selection Sequence Alignment.
Improved Video Categorization from Text Metadata and User Comments ACM SIGIR 2011:Research and development in Information Retrieval - Katja Filippova -
Using Temporal Logic and Model Checking in Automated Recognition of Human Activities for Ambient- Assisted Living Authors : Tommaso Magherini, Alessandro.
Identifying “Best Bet” Web Search Results by Mining Past User Behavior Author: Eugene Agichtein, Zijian Zheng (Microsoft Research) Source: KDD2006 Reporter:
Toward Reliable and Efficient Reporting in Wireless Sensor Networks Authors: Fatma Bouabdallah Nizar Bouabdallah Raouf Boutaba.
Lighting facts.  Lighting can be a big energy consumer in offices and production areas and experience shows that energy savings may be achieved - often.
Experience Report: System Log Analysis for Anomaly Detection
Software Metrics and Reliability
Hardware & Software Reliability
Predicting Interface Failures For Better Traffic Management.
WalkSense: Classifying Home Occupancy States Using Walkway Sensing
Critical Systems Validation
RAID Redundant Array of Inexpensive (Independent) Disks
Palanivel Kodeswaran, Ravi Kokku, Sayandeep Sen, Mudhakar Srivatsa
A handbook on validation methodology. Metrics.
Presentation transcript:

Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University of Virginia, DGIST, Dept of Information and Communication Engineering Study Group Junction

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Introduction  Smart home applications:  Home automation, energy efficiency, home security  Commercial Sensors:  Inexpensive, wireless, battery-powered  Reducing hardware and installation costs

Problems  Do-it-yourself, low-cost sensors  Homes with hundreds of sensors had one sensor failure per day on average  Suffer from many type of faults:  Break down, lose power  Non-fail-stop failure  Sensor does not completely fail. It continues to report values, but the meaning of the value changes or becomes invalid.  e.g. be dislodged, fall off and be re-mounted, covered by objects or blocked by an open door or re-arranged furniture. The maintenance cost of fixing all such failures is prohibitive, and may negate any cost advantage of inexpensive hardware and installation

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Existing solution  Detect and report fail-stop hardware failures  Repeatedly querying the nodes or checking for lost data  Not non-fail-stop (still response)  Detect non-fail-stop failures  Exploiting correlations between neighboring sensors  bottom up approach  Use patterns in the raw sensor data  (O) Homogeneous, periodic and continuous-valued sensors  (X) Heterogeneous, binary, and event-triggered sensors

Proposed solution  Simultaneous Multi-classifier Activity Recognition Technique (SMART)  Use top down application-level semantics to detect, assess and adapt to sensor failures  Detect non-fail-stop node failures:  Getting stuck at a value, node displacement, or node relocation  Runtime failure detection using multiple classifier instances  That are trained to recognize the same set of activities based on different subsets of sensors  One fails -> affect a subset of the classifiers -> change the ratio of activity detection among the classifier

Proposed solution  Once failure is detected  Adapts to the failure  excluding the failed node  creating a new classifier ensemble based on the remaining subset of nodes  Use data replay analysis to assess  whether the failure would have affected activity recognition in the past had the new classifier ensemble been used.  Y: dispatches a maintenance person to repair the failure  N: no maintenance is necessary

Comparison with bottom up method  How accurately correlation-based techniques can detect non-fail-stop failures in event-driven application?  Activity recognition  43-day-long dataset from a two-resident home  Failure -> movement failure, where one of the kitchen motion sensor is accidentally moved to point in a different direction

Correlation-based results  Cannot achieve failure detection accuracy higher than 80%  The accuracy decreases as the number of consecutive testing days increases.  Reason:  Temporal correlation between nodes => when two nodes fire together  But not looks at which activity is being performed  Application-level feature  Top-down approach !

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

SMART Approach Train classifier instances for all possible combinations of node failures by holding those nodes out of the training set. Analyze the effect of sensor failures on the classifiers’ performance Update the classifier ensemble to contain classifiers that are trained for the failure. If the new classifier ensemble can maintain the detection accuracy of the application above the specified severity threshold TH S. 1 The training and use of the classifier ensemble 2 The detection of non-fail-stop failures 3 The node failure severity analysis, which allows to decrease the number of maintenance dispatches. 4 How SMART maintains high detection accuracy under failures

Using multiple simultaneous classifiers  Single-node failures  |S|+1 classifier is used  Detect the occurrence of failure  Identify which nodes have failed by monitoring the relative behavior of the original classifier instance and that of the other |S| instances  Maintain high detection accuracy in the presence of failures

Failure detection  By analyzing the relative performance of the classifiers that had that node in its training set versus those classifiers that did not.  e.g. sensor s fails, a change in the relative behavior of  C ( S, S-s ) and C ( S-s, S-s )  Calculate F-score of each of these classifiers with respect to the original classifier C 0 to measure the similarity between their outputs.  e.g. the F-score for a classifier C i ( S-s i, S )

Definition actual class (observation) predicted class (expectation) tp Correct result fp Unexpected result fn Missing result tn Correct absence of result Precision = Recall =

Failure detection  Each of the |S| classifiers has an F-score associated with it, forming F-score vector:  Characterize the behavior of the system when there are no failures  If a failure occurs, based on the severity of the failure, it might affect some or even all of the values in the F-score vector.

Failure detection  Since C A, C B, and C 0 have all been trained with node C, their relative ratios will remain similar  They are affected in a similar way.  C C was trained by holding node C out and will therefore change in behavior relative to classifier C 0.  SMART can thus infer a failure has occurred and identify the cause of that failure.

Failure Recognition  Step 1: Is there a failure or not?  FDC (failure detection classifier)  Trained to distinguish between non-failure F-score vectors and failure F-score vectors.  Method: use historical data to generate both failure and non-failure F- score and train the FDC.  Run-time:  The system calculates the relative F-scores for the |S| classifiers and builds a F-score vector.  FDC determine if the F-score vector represents a system that has a failure or not. => determine the failure detection accuracy Artificially introducing failures in the historical data through modifying the readings of the “failed” nodes.

Failure Recognition  Step 2: Which node has failed?  FIC (failure identification classifier)  Determines which of the nodes has failed.  Is trained to distinguish between different node failures.  By modifying the readings of the “failed” nodes in the historical data.  Evaluate the failure identification accuracy.

Reaction to small fluctuations  Small fluctuations between F-score vectors might occur even without failures.  If residents of a home alter their normal activity patterns  Non-sever failures might cause very small changes to the behavior of the classifiers.  FDC -> false positive and false negative  To improve the accuracy of the failure detection:  Increase the size of the historical data used for training  Increase the failure detection latency

Node failure severity analysis  node failures are going to impact the application differently based on the available level of redundancy in the system.  e.g. cooking activity  2 sensors in the kitchen  the star one close to the sink  Different degree of impaction  More sensors?

Node failure severity analysis  Severity measurement when failure detected  Assume sensor s fails  To determine the effect of the failure on the appliction  If it decreases the detection accuracy below the severity threshold TH S, and maintenance is dispatched every time a severe failure is detected.  TH S, can be specified by per-activity severity

Maintaining detection accuracy under failure  SMART uses a variety of classifier types  Naïve Bayesian  Hidden Markov Model  Switch to the classifier instance among all types and training sets that performed best on the training data

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Experimental Setup  Evaluate SMART on 3 houses in 2 publicly available activity recognition datasets.  1) CASAS datasets:  40 days of labeled data from over 60 sensors, 2-resident home  2) two single-resident homes: House A and House B  House A: 25 days from 14 sensors  House B: 13 days from 27 sensors

Experimental Setup  Complex activities detection (more than 1 sensor)  Severity threshold TH S = 0.3  The datasets do not contain failure information  All node failures in the experiments were simulated by modifying the values reported by the “failed” node.  “stuck at” failure: set the value of the failed node to 1  “misplacement” failure: replaced the data of the failed sensor s with data from a sensor located at s’ new position

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Results  Detecting sensor node failures  Kitchen -> living room

Results  Node failure severity assessment  WSU: 1/8, HA: 3/8, HB: 3/7 significant sensors

Results  Evaluate SMART’s impact on the MTTF of the application  MTTF: number of time units after which the detection accuracy falls below TH S  Unlike the baseline, our approach determines that the application has filed not when the first node fails, but when the first severe node failure occurs.

Results  Detailed view of the MTTF for prepare breakfast from the WSU house.

Results  Maintaining high activity recognition accuracy under failures  Compare to a classifier trained on all nodes in the system, our approach achieves higher activity recognition accuracy in the presence of node failures

Results

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Discussion  Effect of node use on the failure detection accuracy  The failure detection accuracy of a node v.s.  How frequently this node is used for a particular activity  node usage ratio per activity  The percentage of instances of that activity where node n is was used A positive correlation between the importance of that node for the activity and how accurately we can detect the node’s failure

Discussion  Effect of node use on the failure detection accuracy  The accuracy of detecting a “stuck at” failure for the important nodes for the kitchen activities in House A.  When only the important nodes are considered, the failure detection accuracy increases dramatically

Discussion  Limitations and future work  SMART cannot accurately detect the failures of sensors that are not frequently used in any of the activities.  SMART can be combined with state of the art health- monitoring systems, which can accurately detect fail-stop failures experienced by the rarely used nodes.  Assumption made: single-node failures not multiple failures  Plan to analyze how SMART’s failure detection accuracy is affected by multiple-node failures.

Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

Conclusions  SMART: a general failure detection, assessment, and adaptation approach for smart home applications  Decreases the number of maintenance dispatches by 55%  Triples the MTTF of the application on average  Maintains sufficient activity recognition accuracy in the presence of failures by dynamically updating the classifiers at runtime with over 85% accuracy  Improves the activity recognition accuracy under node failures by more than 15% on average.