DOC PREVIEW
CMU CS 10701 - Support Vector Machines

This preview shows page 1-2-15-16-17-32-33 out of 33 pages.

Save
View full document
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
View full document
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience
Premium Document
Do you want full access? Go Premium and unlock all 33 pages.
Access to all documents
Download any document
Ad free experience

Unformatted text preview:

1Nov 23rd, 2001Copyright © 2001, 2003, Andrew W. MooreSupport Vector Machines Andrew W. MooreProfessorSchool of Computer ScienceCarnegie Mellon Universitywww.cs.cmu.edu/[email protected] to other teachers and users of these slides. Andrew would be delighted if you found this source material useful in giving your own lectures. Feel free to use these slides verbatim, or to modify them to fit your own needs. PowerPoint originals are available. If you make use of a significant portion of these slides in your own lecture, please include this message, or the following link to the source repository of Andrew’s tutorials: http://www.cs.cmu.edu/~awm/tutorials. Comments and corrections gratefully received. Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 2Linear Classifiersf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)How would you classify this data?2Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 3Linear Classifiersf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)How would you classify this data?Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 4Linear Classifiersf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)How would you classify this data?3Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 5Linear Classifiersf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)How would you classify this data?Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 6Linear Classifiersf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)Any of these would be fine....but which is best?4Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 7Classifier Marginf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)Define the marginof a linear classifier as the width that the boundary could be increased by before hitting a datapoint.Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 8Maximum Marginf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)The maximum margin linear classifier is the linear classifier with the, um, maximum margin.This is the simplest kind of SVM (Called an LSVM)Linear SVM5Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 9Maximum Marginf xαyestdenotes +1denotes -1f(x,w,b) = sign(w. x - b)The maximum margin linear classifier is the linear classifier with the, um, maximum margin.This is the simplest kind of SVM (Called an LSVM)Support Vectors are those datapoints that the margin pushes up againstLinear SVMCopyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 10Why Maximum Margin?denotes +1denotes -1f(x,w,b) = sign(w. x - b)The maximum margin linear classifier is the linear classifier with the, um, maximum margin.This is the simplest kind of SVM (Called an LSVM)Support Vectors are those datapoints that the margin pushes up against1. Intuitively this feels safest. 2. If we’ve made a small error in the location of the boundary (it’s been jolted in its perpendicular direction) this gives us least chance of causing a misclassification.3. LOOCV is easy since the model is immune to removal of any non-support-vector datapoints.4. There’s some theory (using VC dimension) that is related to (but not the same as) the proposition that this is a good thing.5. Empirically it works very very well.6Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 11Specifying a line and margin• How do we represent this mathematically?•…in minput dimensions?Plus-PlaneMinus-PlaneClassifier Boundary“Predict Class = +1”zone“Predict Class = -1”zoneCopyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 12Plus-PlaneMinus-PlaneClassifier Boundary“Predict Class = +1”zone“Predict Class = -1”zone7Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 13Specifying a line and margin• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }Plus-PlaneMinus-PlaneClassifier Boundary“Predict Class = +1”zone“Predict Class = -1”zone-1 < w . x + b < 1ifUniverse explodesw . x + b <= -1if-1w . x + b >= 1if+1Classify as..wx+b=1wx+b=0wx+b=-1Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 14Computing the margin width• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }Claim: The vector w is perpendicular to the Plus Plane. Why?“Predict Class = +1”zone“Predict Class = -1”zonewx+b=1wx+b=0wx+b=-1M =Margin WidthHow do we compute Min terms of wand b?8Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 15Computing the margin width• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }Claim: The vector w is perpendicular to the Plus Plane. Why?“Predict Class = +1”zone“Predict Class = -1”zonewx+b=1wx+b=0wx+b=-1M =Margin WidthHow do we compute Min terms of wand b?Let u and v be two vectors on the Plus Plane. What is w . ( u – v ) ?And so of course the vector w is also perpendicular to the Minus PlaneCopyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 16Computing the margin width• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }• The vector w is perpendicular to the Plus Plane•Let x-be any point on the minus plane•Let x+be the closest plus-plane-point to x-.“Predict Class = +1”zone“Predict Class = -1”zonewx+b=1wx+b=0wx+b=-1M =Margin WidthHow do we compute Min terms of wand b?x-x+Any location in ¨m: not necessarily a datapointAny location in Rm: not necessarily a datapoint9Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 17Computing the margin width• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }• The vector w is perpendicular to the Plus Plane•Let x-be any point on the minus plane•Let x+be the closest plus-plane-point to x-.• Claim: x+= x-+ λ wfor some value of λ. Why?“Predict Class = +1”zone“Predict Class = -1”zonewx+b=1wx+b=0wx+b=-1M =Margin WidthHow do we compute Min terms of wand b?x-x+Copyright © 2001, 2003, Andrew W. Moore Support Vector Machines: Slide 18Computing the margin width• Plus-plane = { x : w . x + b = +1 }• Minus-plane = { x : w . x + b = -1 }• The vector w is perpendicular to the Plus Plane•Let x-be any point on the minus plane•Let x+be the closest plus-plane-point to x-.• Claim: x+= x-+ λ wfor some


View Full Document

CMU CS 10701 - Support Vector Machines

Documents in this Course
lecture

lecture

12 pages

lecture

lecture

17 pages

HMMs

HMMs

40 pages

lecture

lecture

15 pages

lecture

lecture

20 pages

Notes

Notes

10 pages

Notes

Notes

15 pages

Lecture

Lecture

22 pages

Lecture

Lecture

13 pages

Lecture

Lecture

24 pages

Lecture9

Lecture9

38 pages

lecture

lecture

26 pages

lecture

lecture

13 pages

Lecture

Lecture

5 pages

lecture

lecture

18 pages

lecture

lecture

22 pages

Boosting

Boosting

11 pages

lecture

lecture

16 pages

lecture

lecture

20 pages

Lecture

Lecture

20 pages

Lecture

Lecture

39 pages

Lecture

Lecture

14 pages

Lecture

Lecture

18 pages

Lecture

Lecture

13 pages

Exam

Exam

10 pages

Lecture

Lecture

27 pages

Lecture

Lecture

15 pages

Lecture

Lecture

24 pages

Lecture

Lecture

16 pages

Lecture

Lecture

23 pages

Lecture6

Lecture6

28 pages

Notes

Notes

34 pages

lecture

lecture

15 pages

Midterm

Midterm

11 pages

lecture

lecture

11 pages

lecture

lecture

23 pages

Boosting

Boosting

35 pages

Lecture

Lecture

49 pages

Lecture

Lecture

22 pages

Lecture

Lecture

16 pages

Lecture

Lecture

18 pages

Lecture

Lecture

35 pages

lecture

lecture

22 pages

lecture

lecture

24 pages

Midterm

Midterm

17 pages

exam

exam

15 pages

Lecture12

Lecture12

32 pages

lecture

lecture

19 pages

Lecture

Lecture

32 pages

boosting

boosting

11 pages

pca-mdps

pca-mdps

56 pages

bns

bns

45 pages

mdps

mdps

42 pages

svms

svms

10 pages

Notes

Notes

12 pages

lecture

lecture

42 pages

lecture

lecture

29 pages

lecture

lecture

15 pages

Lecture

Lecture

12 pages

Lecture

Lecture

24 pages

Lecture

Lecture

22 pages

Midterm

Midterm

5 pages

mdps-rl

mdps-rl

26 pages

Load more
Download Support Vector Machines
Our administrator received your request to download this document. We will send you the file to your email shortly.
Loading Unlocking...
Login

Join to view Support Vector Machines and access 3M+ class-specific study document.

or
We will never post anything without your permission.
Don't have an account?
Sign Up

Join to view Support Vector Machines 2 2 and access 3M+ class-specific study document.

or

By creating an account you agree to our Privacy Policy and Terms Of Use

Already a member?