Download presentation
Presentation is loading. Please wait.
Published byGillian Grace Leonard Modified over 8 years ago
1
Text Classification and Naïve Bayes Formalizing the Naïve Bayes Classifier
2
Dan Jurafsky Bayes’ Rule Applied to Documents and Classes For a document d and a class c
3
Dan Jurafsky Naïve Bayes Classifier (I) MAP is “maximum a posteriori” = most likely class Bayes Rule Dropping the denominator
4
Dan Jurafsky Naïve Bayes Classifier (II) Document d represented as features x1..xn
5
Dan Jurafsky Naïve Bayes Classifier (IV) How often does this class occur? O(|X| n |C|) parameters We can just count the relative frequencies in a corpus Could only be estimated if a very, very large number of training examples was available.
6
Dan Jurafsky Multinomial Naïve Bayes Independence Assumptions Bag of Words assumption: Assume position doesn’t matter Conditional Independence: Assume the feature probabilities P(x i |c j ) are independent given the class c.
7
Dan Jurafsky Multinomial Naïve Bayes Classifier
8
Dan Jurafsky Applying Multinomial Naive Bayes Classifiers to Text Classification positions all word positions in test document
9
Text Classification and Naïve Bayes Formalizing the Naïve Bayes Classifier
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.