214 Support Vector Machines
15.8 Exercises
- Show that the hard-SVM rule, namely,
argmax
(w,b):‖w‖=1
min
i∈[m]|〈w,xi〉+b| s.t. ∀i, yi(〈w,xi〉+b)> 0 ,is equivalent to the following formulation:
argmax
(w,b):‖w‖=1min
i∈[m]yi(〈w,xi〉+b). (15.13)Hint:DefineG={(w,b) :∀i,yi(〈w,xi〉+b)> 0 }.- Show that
argmax
(w,b):‖w‖=1
min
i∈[m]yi(〈w,xi〉+b)∈G- Show that∀(w,b)∈G,
min
i∈[m]
yi(〈w,xi〉+b) = min
i∈[m]|〈w,xi〉+b|2.Margin and the PerceptronConsider a training set that is linearly sep-
arable with a marginγand such that all the instances are within a ball of
radiusρ. Prove that the maximal number of updates the Batch Perceptron
algorithm given in Section 9.1.2 will make when running on this training set
is (ρ/γ)^2.
3.Hard versus soft SVM:Prove or refute the following claim:
There existsλ > 0 such that for every sampleSofm > 1 examples, which
is separable by the class of homogenous halfspaces, the hard-SVM and the
soft-SVM (with parameterλ) learning rules return exactly the same weight
vector.
4.Weak duality:Prove that for any functionf of two vector variablesx∈
X,y∈Y, it holds that
minx∈Xmaxy∈Yf(x,y)≥maxy∈Yminx∈Xf(x,y).