Presentation is loading. Please wait.

Presentation is loading. Please wait.

Analysis of Classification-based Error Functions Mike Rimer Dr. Tony Martinez BYU Computer Science Dept. 18 March 2006.

Similar presentations


Presentation on theme: "Analysis of Classification-based Error Functions Mike Rimer Dr. Tony Martinez BYU Computer Science Dept. 18 March 2006."— Presentation transcript:

1 Analysis of Classification-based Error Functions Mike Rimer Dr. Tony Martinez BYU Computer Science Dept. 18 March 2006

2 Overview Machine learning Teaching artificial neural networks with an error function Problems with conventional error functions CB algorithms Experimental results Conclusion and future work

3 Machine Learning Goal: Automating learning of problem domains Given a training sample from a problem domain, induce a correct solution-hypothesis over the entire problem population The learning model is often used as a black box inputoutput f (x)

4 Teaching ANNs with an Error Function Used to train a multi-layer perceptron (MLP) to guide the gradient descent learning procedure to an optimal state Conventional error metrics are sum-squared error (SSE) and cross entropy (CE) SSE suited to function approximation CE aimed at classification problems CB error functions [Rimer & Martinez 06] work better for classification

5 SSE, CE Attempts to approximate 0-1 targets in order to represent making a decision 0 1 O2O2 O1O1 ERROR 2ERROR 1 Pattern labeled as class 2

6 Issues with approximating hard targets Requires weights to be large to achieve optimality Leads to premature weight saturation Weight decay, etc., can improve the situation Learns areas of the problem space unevenly and at different times during training Makes global learning problematic

7 Classification-based Error Functions Designed to more closely match the goal of learning a classification task (i.e. correct classifications, not low error on 0-1 targets), avoiding premature weight saturation and discouraging overfit CB1 [Rimer & Martinez 02, 06] CB2 [Rimer & Martinez 04] CB3 (submitted to ICML ‘06)

8 CB1 Only backpropagates error on misclassified training patterns 0 1 Correct T~T 0 1 Misclassified T~T ERROR

9 CB2 Adds a confidence margin, μ, that is increased globally as training progresses 0 1 Misclassified T~T ERROR μ 0 1 ~TT ERROR μ Correct, but doesn’t satisfy margin 0 1 Correct, and satisfies margin T~T μ

10 CB3 Learns a confidence C i for each training pattern i as training progresses Patterns often misclassified have low confidence Patterns consistently classified correctly gain confidence 0 1 Misclassified T~T ERROR 0 1 ~TT ERROR CiCi Correct with learned low confidence 0 1 ~TT ERROR CiCi Correct with learned high confidence

11 Neural Network Training Influenced by: Initial parameter (weight) settings Pattern order presentation (stochastic training) Learning rate # of hidden nodes Goal of training: High generalization Low bias and variance

12 Experiments Empirical comparison of six error functions SSE, CE, CE w/ WD, CB1-3 Used eleven benchmark problems from the UC Irvine Machine Learning Repository ann, balance, bcw, derm, ecoli, iono, iris, musk2, pima, sonar, wine Testing performed using stratified 10-fold cross- validation Model selection by hold-out set Results were averaged over ten tests LR = 0.1, M = 0.7

13 Classifier output difference (COD) Evaluation of behavioral difference of two hypotheses (e.g. classifiers) T is the test set I is the identity or characteristic function

14 Robustness to initial network weights Averaged 30 random runs over all datasets algorithm % Test accSt DevEpoch CB393.4684.7792200.67 CB292.8394.0800366.69 CB192.8285.3290514.14 CE92.7895.3937319.57 CE w/ WD92.2515.4735197.24 SSE91.9515.6131774.70

15 Robustness to initial network weights Averaged over all tests AlgorithmTest errorCOD CB30.06530.0221 CB20.07160.0274 CB10.07170.0244 CE0.07210.0248 CE w/ WD0.07740.0255 SSE0.08040.0368

16 Robustness to pattern presentation order Averaged 30 random runs over all datasets algorithm % Test accSt DevEpoch CB393.4465.0409200.46 CB292.6415.4197402.52 CB192.5425.473560.09 CE92.2905.6020329.65 CE w/ WD91.8185.6278221.21 SSE91.8175.6653593.30

17 Robustness to pattern presentation order Averaged over all tests AlgorithmTest errorCOD CB30.06550.0259 CB20.07360.0302 CB10.07460.0282 CE0.07710.0329 CE w/ WD0.08180.0338 SSE0.08180.0344

18 Robustness to learning rate Average of varying the learning rate from 0.01 – 0.3 AlgorithmTest accSt DevEpoch CB393.1753.514334.8 CB292.2853.437617.8 SSE92.2113.449525.7 CB191.9083.880505.4 CE91.6293.813466.2 CE w/ WD91.3303.845234.6

19 Robustness to learning rate

20 Robustness to number of hidden nodes Average of varying the number of nodes in the hidden layer from 1 - 30 AlgorithmTest accSt devEpoch CB393.0263.397303.9 CB192.2913.610381.0 CB292.1363.410609.4 SSE92.0663.402623.1 CE91.9563.563397.0 CE w/ WD91.743.493190.6

21 Robustness to number of hidden nodes

22 Conclusion CB1-3 are generally more robust than SSE, CE, and CE w/ WD, with respect to: Initial weight settings Pattern presentation order Pattern variance Learning rate # hidden nodes CB3 is superior, most robust, with most consistent results

23 Questions?

24

25

26

27

28

29


Download ppt "Analysis of Classification-based Error Functions Mike Rimer Dr. Tony Martinez BYU Computer Science Dept. 18 March 2006."

Similar presentations


Ads by Google