Presentation is loading. Please wait.

Presentation is loading. Please wait.

CRF Recitation Kevin Tang. Conditional Random Field Definition.

Similar presentations


Presentation on theme: "CRF Recitation Kevin Tang. Conditional Random Field Definition."— Presentation transcript:

1 CRF Recitation Kevin Tang

2 Conditional Random Field Definition

3 Meaning of Graphical Model

4 Discriminative v.s. Generative Y=0Y=1 X=1 1/20 X=2 1/4 Y=0Y=1 X=1 10 X=2 1/2 Stolen from: http://stackoverflow.com/questions/879432/what-is-the-difference-between-a-generative-and-discriminative-algorithmhttp://stackoverflow.com/questions/879432/what-is-the-difference-between-a-generative-and-discriminative-algorithm Also, see http://papers.nips.cc/paper/2020-on-discriminative-vs-generative-classifiers-a-comparison-of-logistic-regression-and-naive- bayes.pdfhttp://papers.nips.cc/paper/2020-on-discriminative-vs-generative-classifiers-a-comparison-of-logistic-regression-and-naive- bayes.pdf

5 Comparison To HMMs  Audience thoughts?

6 Comparison To HMMs  Similarities:  Both probabilistic models  Both use the Markov Property as an assumption  Differences  CRFs are discriminative while HMM’s are generative  CRFs may have more accuracy with sequence tagging as it directly models p(y|x)  HMMs use Bayes Rule to model tagging  HMMs can generate samples from the distribution p(x, y) and are often more robust (missing labels, unsupervised, or semisupervised)  Hmms can handle missing labels

7 Let’s summarize terminology and symbols

8 Other Formulae/Symbols we may see

9 Objective of Gradient Descent

10

11

12

13

14 Nesterov’s accelerated gradient descent

15 Summary of Gradient Descent  Pregenerate phis  Calculate dF  Calculate dlogZ  Generate Gs, generate alphas, betas  Run forward backwards algorithm with normalization  Calculate dw = dF – dlogZ  Update w = w + dw or use Nesterov  End after number of iterations, or when change hits a minimum, or percent change hits a minimum.

16 Some numbers for sanity purposes  Stuff that I got  ~250 iterations with Nesterov acceleration (will vary depending on your growth factor)  ~5 minutes computational time in Matlab Much faster when outside of a Matlab Class…(more like 1 minute)  ~30 minutes on a very unoptimized solution (but hey, it worked)  Could get faster with more vectorization, but I’m lazy.  You probably will have better luck in Python (grumble grumble)  ~50% hamming loss


Download ppt "CRF Recitation Kevin Tang. Conditional Random Field Definition."

Similar presentations


Ads by Google