Presentation is loading. Please wait.

Presentation is loading. Please wait.

Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University.

Similar presentations


Presentation on theme: "Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University."— Presentation transcript:

1 Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University of Virginia, DGIST, Dept of Information and Communication Engineering 2013.02.25 Study Group Junction

2 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

3 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

4 Introduction  Smart home applications:  Home automation, energy efficiency, home security  Commercial Sensors:  Inexpensive, wireless, battery-powered  Reducing hardware and installation costs

5 Problems  Do-it-yourself, low-cost sensors  Homes with hundreds of sensors had one sensor failure per day on average  Suffer from many type of faults:  Break down, lose power  Non-fail-stop failure  Sensor does not completely fail. It continues to report values, but the meaning of the value changes or becomes invalid.  e.g. be dislodged, fall off and be re-mounted, covered by objects or blocked by an open door or re-arranged furniture. The maintenance cost of fixing all such failures is prohibitive, and may negate any cost advantage of inexpensive hardware and installation

6 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

7 Existing solution  Detect and report fail-stop hardware failures  Repeatedly querying the nodes or checking for lost data  Not non-fail-stop (still response)  Detect non-fail-stop failures  Exploiting correlations between neighboring sensors  bottom up approach  Use patterns in the raw sensor data  (O) Homogeneous, periodic and continuous-valued sensors  (X) Heterogeneous, binary, and event-triggered sensors

8 Proposed solution  Simultaneous Multi-classifier Activity Recognition Technique (SMART)  Use top down application-level semantics to detect, assess and adapt to sensor failures  Detect non-fail-stop node failures:  Getting stuck at a value, node displacement, or node relocation  Runtime failure detection using multiple classifier instances  That are trained to recognize the same set of activities based on different subsets of sensors  One fails -> affect a subset of the classifiers -> change the ratio of activity detection among the classifier

9 Proposed solution  Once failure is detected  Adapts to the failure  excluding the failed node  creating a new classifier ensemble based on the remaining subset of nodes  Use data replay analysis to assess  whether the failure would have affected activity recognition in the past had the new classifier ensemble been used.  Y: dispatches a maintenance person to repair the failure  N: no maintenance is necessary

10 Comparison with bottom up method  How accurately correlation-based techniques can detect non-fail-stop failures in event-driven application?  Activity recognition  43-day-long dataset from a two-resident home  Failure -> movement failure, where one of the kitchen motion sensor is accidentally moved to point in a different direction

11 Correlation-based results  Cannot achieve failure detection accuracy higher than 80%  The accuracy decreases as the number of consecutive testing days increases.  Reason:  Temporal correlation between nodes => when two nodes fire together  But not looks at which activity is being performed  Application-level feature  Top-down approach !

12 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

13 SMART Approach Train classifier instances for all possible combinations of node failures by holding those nodes out of the training set. Analyze the effect of sensor failures on the classifiers’ performance Update the classifier ensemble to contain classifiers that are trained for the failure. If the new classifier ensemble can maintain the detection accuracy of the application above the specified severity threshold TH S. 1 The training and use of the classifier ensemble 2 The detection of non-fail-stop failures 3 The node failure severity analysis, which allows to decrease the number of maintenance dispatches. 4 How SMART maintains high detection accuracy under failures

14 Using multiple simultaneous classifiers  Single-node failures  |S|+1 classifier is used  Detect the occurrence of failure  Identify which nodes have failed by monitoring the relative behavior of the original classifier instance and that of the other |S| instances  Maintain high detection accuracy in the presence of failures

15 Failure detection  By analyzing the relative performance of the classifiers that had that node in its training set versus those classifiers that did not.  e.g. sensor s fails, a change in the relative behavior of  C ( S, S-s ) and C ( S-s, S-s )  Calculate F-score of each of these classifiers with respect to the original classifier C 0 to measure the similarity between their outputs.  e.g. the F-score for a classifier C i ( S-s i, S )

16 Definition actual class (observation) predicted class (expectation) tp Correct result fp Unexpected result fn Missing result tn Correct absence of result Precision = Recall =

17 Failure detection  Each of the |S| classifiers has an F-score associated with it, forming F-score vector:  Characterize the behavior of the system when there are no failures  If a failure occurs, based on the severity of the failure, it might affect some or even all of the values in the F-score vector.

18 Failure detection  Since C A, C B, and C 0 have all been trained with node C, their relative ratios will remain similar  They are affected in a similar way.  C C was trained by holding node C out and will therefore change in behavior relative to classifier C 0.  SMART can thus infer a failure has occurred and identify the cause of that failure.

19 Failure Recognition  Step 1: Is there a failure or not?  FDC (failure detection classifier)  Trained to distinguish between non-failure F-score vectors and failure F-score vectors.  Method: use historical data to generate both failure and non-failure F- score and train the FDC.  Run-time:  The system calculates the relative F-scores for the |S| classifiers and builds a F-score vector.  FDC determine if the F-score vector represents a system that has a failure or not. => determine the failure detection accuracy Artificially introducing failures in the historical data through modifying the readings of the “failed” nodes.

20 Failure Recognition  Step 2: Which node has failed?  FIC (failure identification classifier)  Determines which of the nodes has failed.  Is trained to distinguish between different node failures.  By modifying the readings of the “failed” nodes in the historical data.  Evaluate the failure identification accuracy.

21 Reaction to small fluctuations  Small fluctuations between F-score vectors might occur even without failures.  If residents of a home alter their normal activity patterns  Non-sever failures might cause very small changes to the behavior of the classifiers.  FDC -> false positive and false negative  To improve the accuracy of the failure detection:  Increase the size of the historical data used for training  Increase the failure detection latency

22 Node failure severity analysis  node failures are going to impact the application differently based on the available level of redundancy in the system.  e.g. cooking activity  2 sensors in the kitchen  the star one close to the sink  Different degree of impaction  More sensors?

23 Node failure severity analysis  Severity measurement when failure detected  Assume sensor s fails  To determine the effect of the failure on the appliction  If it decreases the detection accuracy below the severity threshold TH S, and maintenance is dispatched every time a severe failure is detected.  TH S, can be specified by per-activity severity

24 Maintaining detection accuracy under failure  SMART uses a variety of classifier types  Naïve Bayesian  Hidden Markov Model  Switch to the classifier instance among all types and training sets that performed best on the training data

25 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

26 Experimental Setup  Evaluate SMART on 3 houses in 2 publicly available activity recognition datasets.  1) CASAS datasets:  40 days of labeled data from over 60 sensors, 2-resident home  2) two single-resident homes: House A and House B  House A: 25 days from 14 sensors  House B: 13 days from 27 sensors

27 Experimental Setup  Complex activities detection (more than 1 sensor)  Severity threshold TH S = 0.3  The datasets do not contain failure information  All node failures in the experiments were simulated by modifying the values reported by the “failed” node.  “stuck at” failure: set the value of the failed node to 1  “misplacement” failure: replaced the data of the failed sensor s with data from a sensor located at s’ new position

28 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

29 Results  Detecting sensor node failures  Kitchen -> living room

30 Results  Node failure severity assessment  WSU: 1/8, HA: 3/8, HB: 3/7 significant sensors

31 Results  Evaluate SMART’s impact on the MTTF of the application  MTTF: number of time units after which the detection accuracy falls below TH S  Unlike the baseline, our approach determines that the application has filed not when the first node fails, but when the first severe node failure occurs.

32 Results  Detailed view of the MTTF for prepare breakfast from the WSU house.

33 Results  Maintaining high activity recognition accuracy under failures  Compare to a classifier trained on all nodes in the system, our approach achieves higher activity recognition accuracy in the presence of node failures

34 Results

35 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

36 Discussion  Effect of node use on the failure detection accuracy  The failure detection accuracy of a node v.s.  How frequently this node is used for a particular activity  node usage ratio per activity  The percentage of instances of that activity where node n is was used A positive correlation between the importance of that node for the activity and how accurately we can detect the node’s failure

37 Discussion  Effect of node use on the failure detection accuracy  The accuracy of detecting a “stuck at” failure for the important nodes for the kitchen activities in House A.  When only the important nodes are considered, the failure detection accuracy increases dramatically

38 Discussion  Limitations and future work  SMART cannot accurately detect the failures of sensors that are not frequently used in any of the activities.  SMART can be combined with state of the art health- monitoring systems, which can accurately detect fail-stop failures experienced by the rarely used nodes.  Assumption made: single-node failures not multiple failures  Plan to analyze how SMART’s failure detection accuracy is affected by multiple-node failures.

39 Outline  Introduction  Proposed Solution  SMART Approach Detail  Experimental Setup  Results  Discussions  Conclusions

40 Conclusions  SMART: a general failure detection, assessment, and adaptation approach for smart home applications  Decreases the number of maintenance dispatches by 55%  Triples the MTTF of the application on average  Maintains sufficient activity recognition accuracy in the presence of failures by dynamically updating the classifiers at runtime with over 85% accuracy  Improves the activity recognition accuracy under node failures by more than 15% on average.


Download ppt "Being SMART About Failures: Assessing Repairs in Smart Homes Krasimira Kapitanova, Enamul Hoque, John A. Stankovic, Kamin Whitehouse, Sang H. Son University."

Similar presentations


Ads by Google