Download presentation
Presentation is loading. Please wait.
Published byΓάδ Βονόρτας Modified over 6 years ago
1
CHAPTER 7 BAYESIAN NETWORK INDEPENDENCE BAYESIAN NETWORK INFERENCE MACHINE LEARNING ISSUES
2
Review: Alarm Network
3
Causality? When Bayesian Networks reflect the true causal patterns:
Often simpler (nodes have fewer parents) Often easier to think about Often easier to elicit from experts BNs need not actually be causal Sometimes no causal net exists over the domain E.g. consider the variables Traffic and RoofDrips End up with arrows that reflect correlation, not causation What do the arrows really mean? Topology may happen to encode causal structure Topology really encodes conditional independencies
4
Creating Bayes’ Nets Last time: we talked about how any fixed Bayesian Network encodes a joint distribution Today: how to represent a fixed distribution as a Bayesian Network Key ingredient: conditional independence The exercise we did in “causal” assembly of BNs was a kind of intuitive use of conditional independence Now we have to formalize the process After that: how to answer queries (inference)
5
Conditional Independence
6
Conditional Independence
7
Independence in a BN
8
Causal Chains
9
Common Cause
10
Common Effect
11
The General Case
12
Reachability
13
Reachability (the Bayes Ball)
14
Example
15
Inference
16
Reminder: Alarm Network
17
Atomic Inference
18
Inference by Enumeration
19
Evaluation Tree
20
Variable Elimination Still lots of redundant work in the computation tree! We can save time if we cache all partial results This is the basic idea behind the variable elimination algorithm Compute and store factors over variables which represent results of intermediate computations All CPDs are factors, but not all factors are CPDs Thus not always “human interpretable” Just improves efficiency, doesn’t improve worst case time complexity Still exponential in the number of variables That’s all we’ll expect you to know!
21
Classification
22
Tuning on Held-Out Data
23
Confidences from a Classifier
24
Precision vs. Recall
25
Precision vs. Recall
26
Errors, and What to Do
27
What to Do About Errors? Need more features: words aren’t enough!
Have you ed the sender before? Have 1K other people just gotten the same ? Is the sending information consistent? Is the in ALL CAPS? Do inline URLs point where they say they point? Does the address you by (your) name? Naïve Bayes models can incorporate a variety of features, but tend to do best in homogeneous cases (e.g. all features are word occurrences)
28
Features A feature is a function which signals a property of the input
Examples: ALL_CAPS: value is 1 iff in all caps HAS_URL: value is 1 iff has a URL NUM_URLS: number of URLs in VERY_LONG: 1 iff is longer than 1K SUSPICIOUS_SENDER: 1 iff reply-to domain doesn’t match originating server Features are anything you can think of code to evaluate on an input Some cheap, some very very expensive to calculate Can even be the output of another classifier Domain knowledge goes here! In Naïve Bayes, how did we encode features?
29
Feature Extractors
30
Generative vs. Discriminative
Generative classifiers: E.g. Naïve Bayes We build a causal model of the variables We then query that model for causes, given evidence Discriminative classifiers: E.g. Perceptron (next) No causal model, no Bayes rule, often no probabilities Try to predict output directly Loosely: mistake driven rather than model driven
31
Some (Vague) Biology
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.