This shows you the differences between two versions of the page.
Next revision | Previous revision Next revision Both sides next revision | ||
papers:bakir-bottou-weston-2005 [2006/04/12 19:14] 127.0.0.1 (old revision restored) |
papers:bakir-bottou-weston-2005 [2006/08/09 12:49] leonb |
||
---|---|---|---|
Line 1: | Line 1: | ||
===== Breaking SVM Complexity with Cross-Training ===== | ===== Breaking SVM Complexity with Cross-Training ===== | ||
+ | // | ||
+ | We propose to selectively remove examples from the training set using | ||
+ | probabilistic estimates related to editing algorithms | ||
+ | (Devijver and Kittler, 1982). | ||
+ | separable distribution of training examples with minimal impact on the | ||
+ | position of the decision boundary. | ||
+ | the number of SVs and the number of training examples, and sharply reduces the | ||
+ | complexity of SVMs during both the training and prediction stages. | ||
<box 99% orange> | <box 99% orange> | ||
- | Gökhan Bakir, Léon Bottou and Jason Weston: Breaking SVM Complexity with Cross-Training, | + | Gökhan Bakir, Léon Bottou and Jason Weston: |
[[http:// | [[http:// | ||
Line 23: | Line 30: | ||
} | } | ||
+ | ==== Notes ==== | ||
+ | |||
+ | **Much better** solutions for this problem are discussed in | ||
+ | [[bordes-ertekin-weston-bottou-2005|(Bordes et al., 2006)]] | ||
+ | and [[collobert-weston-bottou-2006|(Collobert et al., 2006)]]. |