Adaptive sampling for incremental optimization using stochastic gradient descent

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

A wide collection of popular statistical learning methods, ranging from K-means to Support Vector Machines through Neural Networks, can be formulated as a stochastic gradient descent (SGD) algorithm in a specific setup. In practice, the main limitation of this incremental optimization technique is due to the stochastic noise induced by the choice at random of the data involved in the gradient estimate computation at each iteration. It is the purpose of this paper to introduce a novel implementation of the SGD algorithm, where the data subset used at a given step is not picked uniformly at random among all possible subsets but drawn from a specific adaptive sampling scheme, depending on the past iterations in a Markovian manner, in order to refine the current statistical estimation of the gradient. Beyond an algorithmic description of the approach we propose, rate bounds are established and illustrative numerical results are displayed in order to provide theoretical and empirical evidence of its statistical performance, compared to more “naive” SGD implementations. Computational issues are also discussed at length, revealing the practical advantages of the method promoted.

Original languageEnglish
Title of host publicationAlgorithmic Learning Theory - 26th International Conference, ALT 2015
EditorsClaudio Gentile, Sandra Zilles, Kamalika Chaudhuri
PublisherSpringer Verlag
Pages317-331
Number of pages15
ISBN (Print)9783319244853
DOIs
Publication statusPublished - 1 Jan 2015
Externally publishedYes
Event26th International Conference on Algorithmic Learning Theory, ALT 2015 - Banff, Canada
Duration: 4 Oct 20156 Oct 2015

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume9355
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference26th International Conference on Algorithmic Learning Theory, ALT 2015
Country/TerritoryCanada
CityBanff
Period4/10/156/10/15

Fingerprint

Dive into the research topics of 'Adaptive sampling for incremental optimization using stochastic gradient descent'. Together they form a unique fingerprint.

Cite this