This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision Next revision Both sides next revision | ||
assignments:assignment4 [2016/10/03 10:00] asa [Part 2: leave-one-out error for linearly separable data] |
assignments:assignment4 [2016/10/05 11:41] asa [Part 3: Soft-margin SVM for separable data] |
||
---|---|---|---|
Line 27: | Line 27: | ||
Suppose you are given a linearly separable dataset, and you are training the soft-margin SVM, which uses slack variables with the soft-margin constant $C$ set | Suppose you are given a linearly separable dataset, and you are training the soft-margin SVM, which uses slack variables with the soft-margin constant $C$ set | ||
- | with the soft margin constant $C$ set | ||
to some positive value. | to some positive value. | ||
Consider the following statement: | Consider the following statement: | ||
Since increasing the $\xi_i$ can only increase the objective of the primal problem (which | Since increasing the $\xi_i$ can only increase the objective of the primal problem (which | ||
- | we are trying to minimize), at the optimal solution to the primal problem, all the | + | we are trying to minimize), at the solution to the primal problem, all the |
training examples will have $\xi_i$ equal | training examples will have $\xi_i$ equal | ||
to zero. | to zero. |