User Tools

Site Tools


Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
papers:bordes-bottou-gallinari-2009 [2009/07/28 12:34]
leonb
papers:bordes-bottou-gallinari-2009 [2017/11/29 10:27] (current)
leonb [Errata]
Line 9: Line 9:
 descent but requires less iterations  descent but requires less iterations 
 to achieve the same accuracy.  to achieve the same accuracy. 
-This algorithm won the ``Wild Track'' of the first +This algorithm won the "Wild Trackof the first 
 PASCAL Large Scale Learning Challenge. PASCAL Large Scale Learning Challenge.
  
 +<html><font color=blue></html>
 +//Errata//: 
 +Please see section [[#Errata]] below.
 +<html></font></html>
  
 <box 99% orange> <box 99% orange>
-Antoine Bordes, Léon Bottou and Patrick Gallinari:  **SGD-QN: Careful Quasi-Newton Stochastic Gradient Descent**,  //Journal of Machine Learning Research//, 10:1737--1754, 2009. (to appear)+Antoine Bordes, Léon Bottou and Patrick Gallinari:  **SGD-QN: Careful Quasi-Newton Stochastic Gradient Descent**,  //Journal of Machine Learning Research//, 10:1737--1754, July 2009.
  
-<html><!-- [[http://leon.bottou.org/publications/djvu/jmlr-2009.djvu|jmlr-2009.djvu]]+[[http://jmlr.csail.mit.edu/papers/v10/bordes09a.html|JMLR Link]] 
 +[[http://jmlr.csail.mit.edu/papers/v11/bordes10a.html|JMLR Erratum]] 
 +<html>&nbsp;&nbsp;</html> 
 +[[http://leon.bottou.org/publications/djvu/jmlr-2009.djvu|jmlr-2009.djvu]]
 [[http://leon.bottou.org/publications/pdf/jmlr-2009.pdf|jmlr-2009.pdf]] [[http://leon.bottou.org/publications/pdf/jmlr-2009.pdf|jmlr-2009.pdf]]
-[[http://leon.bottou.org/publications/psgz/jmlr-2009.ps.gz|jmlr-2009.ps.gz]] --> </html>+[[http://leon.bottou.org/publications/psgz/jmlr-2009.ps.gz|jmlr-2009.ps.gz]]
 </box> </box>
  
Line 28: Line 35:
     volume = {10},     volume = {10},
     pages = {1737--1754},     pages = {1737--1754},
 +    month = {July},
     url = {http://leon.bottou.org/papers/bordes-bottou-gallinari-2009},     url = {http://leon.bottou.org/papers/bordes-bottou-gallinari-2009},
   }   }
 +
 +==== Implementation ====
 +
 +The complete source code of 
 +[[http://webia.lip6.fr/~bordes/mywiki/doku.php?id=sgdqn|LibSGDQN]]
 +is available on 
 +[[http://webia.lip6.fr/~bordes/mywiki/doku.php|Antoine's]] web site.
 +This source code comes with a script that replicates the
 +experiments discussed in this paper.
 +
 +
 +==== Appendix ====
 +
 +The appendix contains a derivation of upper and lower bounds 
 +on the asymptotic convergence speed of stochastic gradient algorithm.
 +The constants are exact in the case of second order stochastic gradient.
 +
 +
 +==== Errata ====
 +
 +The SGDQN algorithm as described in this paper contains a subtle flaw
 +described in a subsequent [[:papers:bordes-2010|erratum]].
 +
 +There is a missing 1/2 factor in the bounds of theorem 1.
 +
 +\[
 + \def\w{\mathbf{w}}
 + {\frac{1}{2}} \frac{{\mathrm tr}(\mathbf{HBGB})}{2\lambda_{\max}-1}\,t^{-1} + {\mathrm o}(t^{-1})
 +  ~\leq~ \mathbb{E}_{\sigma}\big[\:{\cal P}_n(\w_t)-{\cal P}_n(\w^*_n)\:\big] ~\leq~ 
 +  {\frac{1}{2}} \frac{{\mathrm tr}(\mathbf{HBGB})}{2\lambda_{\min}-1}\,t^{-1} + {\mathrm o}(t^{-1}) 
 +\]
 +
 +The version of the paper found on this site contains the correct theorem and proof.
 +
 +
 +
 +
 +
  
papers/bordes-bottou-gallinari-2009.1248798864.txt.gz · Last modified: 2009/07/28 12:34 by leonb

Page Tools