Download presentation
Presentation is loading. Please wait.
1
Vapnik-Chervonenkis Dimension
Part I: Definition and Lower bound
2
PAC Learning model There exists a distribution D over domain X
Examples: <x, c(x)> use c for target function (rather than ct) Goal: With high probability (1-d) find h in H such that error(h,c ) < e e arbitrarily small.
3
VC: Motivation Handle infinite classes.
VC-dim “replaces” finite class size. Previous lecture (on PAC): specific examples rectangle. interval. Goal: develop a general methodology.
4
Definitions: Projection
Given a concept c over X associate it with a set (all positive examples) Projection (sets) For a concept class C and subset S PC(S) = { c S | c C} Projection (vectors) For a concept class C and S = {x1, … , xm} PC(S) = {<c(x1), … , cxm)> | c C}
5
Definition: VC-dim Clearly |PC(S) | 2m C shatters S if |PC(S) | =2m
VC dimension of a class C: The size d of the largest set S that shatters C. Can be infinite. For a finite class C VC-dim(C) log |C|
6
Example 1: Interval 1 C1={cz | z [0,1] } cz(x) = 1 x z
7
Example 2: line C2={cw | w=(a,b,c) } cw(x,y) = 1 ax+by c
8
Example 3: Parallel Rectangle
9
Example 4: Finite union of intervals
10
Example 5 : Parity n Boolean input variables T {1, …, n}
fT(x) = iT xi Lower bound: n unit vectors Upper bound Number of concepts Linear dependency
11
Example 6: OR n Boolean input variables P and N subsets {1, …, n}
fP,N(x) = ( iP xi) ( iN xi) Lower bound: n unit vectors Upper bound Trivial 2n Use ELIM (get n+1) Show second vector removes 2 (get n)
12
Example 7: Convex polygons
13
Example 7: Convex polygons
14
Example 8: Hyper-plane VC-dim(C8) = d+1 Lower bound Upper bound!
C8={cw,c | wd} cw,c(x) = 1 <w,x> c VC-dim(C8) = d+1 Lower bound unit vectors and zero vector Upper bound!
15
Radon Theorem Definitions: Theorem: Proof! Convex set.
Convex hull: conv(S) Theorem: Let T be a set of d+2 points in Rd There exists a subset S of T such that conv(S) conv(T \ S) Proof!
16
Hyper-plane: Finishing the proof
Assume d+2 points T can be shattered. Use Radon Theorem to find S such that conv(S) conv(T \ S) Assign point in S label 1 points not in S label 0 There is a separating hyper-plane How will it label conv(S) conv(T \ S)
17
Lower bounds: Setting Static learning algorithm:
asks for a sample S of size m(e,d) Based on S selects a hypothesis
18
Lower bounds: Setting Theorem: Proof: Expected error ¼. Finish proof!
if VC-dim(C) = then C is not learnable. Proof: Let m = m(0.1,0.1) Find 2m points which are shattered (set T) Let D be the uniform distribution on T Set ct(xi)=1 with probability ½. Expected error ¼. Finish proof!
19
Lower Bound: Feasible Theorem Proof: VC-dim(C)=d+1, then m(e,d)=W(d/e)
Let T be a set of d+1 points which is shattered. D samples: z0 with prob. 1-8e zi with prob. 8e/d
20
Continue Expected error 2e Bound confidence
Set ct(z0)=1 and ct(zi)=1 with probability ½ Expected error 2e Bound confidence for accuracy e
21
Lower Bound: Non-Feasible
Theorem For two hypoth. m(e,d)=W((log 1/d)/e2 ) Proof: Let H={h0, h1}, where hb(x)=b Two distributions: D0: Prob. <x,1> is ½ - g and <y,0> is ½ + g D1: Prob. <x,1> is ½ + g and <y,0> is ½ - g
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.