New optimization algorithms for neural network training using operator splitting techniques

Cristian Daniel Alecsa, Titus Pinţa, Imre Boros

Research output: Contribution to journalArticlepeer-review

Abstract

In the following paper we present a new type of optimization algorithms adapted for neural network training. These algorithms are based upon sequential operator splitting technique for some associated dynamical systems. Furthermore, we investigate through numerical simulations the empirical rate of convergence of these iterative schemes toward a local minimum of the loss function, with some suitable choices of the underlying hyper-parameters. We validate the convergence of these optimizers using the results of the accuracy and of the loss function on the MNIST, MNIST-Fashion and CIFAR 10 classification datasets.

Original languageEnglish
Pages (from-to)178-190
Number of pages13
JournalNeural Networks
Volume126
DOIs
Publication statusPublished - 1 Jun 2020
Externally publishedYes

Keywords

  • CIFAR10
  • Dynamical system
  • MNIST
  • Nesterov
  • Neural network
  • Splitting

Fingerprint

Dive into the research topics of 'New optimization algorithms for neural network training using operator splitting techniques'. Together they form a unique fingerprint.

Cite this