Download presentation
Presentation is loading. Please wait.
1
Review P(h i | d) – probability that the hypothesis is true, given the data (effect cause) Used by MAP: select the hypothesis that is most likely given the observed data P(d | h i ) – probability that the data is true, given the hypothesis (cause effect) Used by ML: select the hypothesis most likely to predict the observed data P(h i | d) = P(d | h i ) P(h i ) / P(d)
3
ML Learning θ = c/N Prediction is that actual proportion of candies in the bag is the same as the proportion in the observed data Problem: if some event has no yet been observed, it is assumed to have zero probability
4
MAP 50/50 bag
5
ML 50/50 bag
6
MAP 25/75 bag
7
ML 25/75 bag
8
Multiple Parameters New manufacturer wraps candies in different colors, probabilistically based in flavor
9
Observed data: c = cherry flavor, l = lime flavor, r c = cherry with red, g c = cherry with green r l = lime with red, g l = lime with green
10
With complete (observable) data, the maximum- likelihood parameter problem decomposes into separate problems, one for each variable.
11
Naïve Bayes Models We can expand to additional attributes, assuming attributes are conditionally independent: θ = P(C = true) θ i1 = P(X i = true | C = true) θ i2 = P(Xi = true | C = false) There 2n+1 parameters and this can be solved deterministically (no search)
13
Learning Bayes Nets If we don’t know the structure of the Bayes net, we can resort to search: Start with a model with no links Add parents, fitting parameters as above Measure accuracy (goal is best accuracy) Alternate: Start with initial guess of structure Make modifications by hill-climbing, simulated annealing: reverse, add or delete arcs
14
Simple reflex agents
15
Model-based reflex agents
16
Goal-based agents
17
Utility-based agents
18
Learning agents
Similar presentations
© 2024 SlidePlayer.com. Inc.
All rights reserved.