Are You Still Tuning Hyperparameters? Parameter-free Model Selection and Learning

Publication
Dec 13, 2014
Abstract

Stochastic gradient descent algorithms for training linear and kernel predictors are gaining more and more importance, thanks to their scalability. While various methods have been proposed to speed up their convergence, the model selection phase is often ignored. In this paper, we propose a new kernel-based stochastic gradient descent algorithm that performs model selection while training, with no parameters to tune, nor any form of cross-validation. The algorithm estimates over time the right regularization in a data-dependent way. Optimal rates of convergence are proved under standard smoothness assumptions on the target function.

  • Workshop "Modern Nonparametrics 3: Automating the Learning Pipeline" at NIPS 2014
  • Conference/Workshop Paper

BibTeX