This paper deals with the steplength selection in stochastic gradient methods for large scale optimization problems arising in machine learning. We introduce an adaptive steplength selection derived by tailoring a limited memory steplength rule, recently developed in the deterministic context, to the stochastic gradient approach. The proposed steplength rule provides values within an interval, whose bounds need to be prefixed by the user. A suitable choice of the interval bounds allows to perform similarly to the standard stochastic gradient method equipped with the best-tuned steplength. Since the setting of the bounds slightly affects the performance, the new rule makes the tuning of the parameters less expensive with respect to the choice of the optimal prefixed steplength in the standard stochastic gradient method. We evaluate the behaviour of the proposed steplength selection in training binary classifiers on well known data sets and by using different loss functions.

On the Steplength Selection in Stochastic Gradient Methods / Franchini, G.; Ruggiero, V.; Zanni, L.. - 11973:(2020), pp. 186-197. (Intervento presentato al convegno 3rd Triennial International Conference and Summer School on Numerical Computations: Theory and Algorithms, NUMTA 2019 tenutosi a ita nel 2019) [10.1007/978-3-030-39081-5_17].

On the Steplength Selection in Stochastic Gradient Methods

Franchini G.;Zanni L.
2020

Abstract

This paper deals with the steplength selection in stochastic gradient methods for large scale optimization problems arising in machine learning. We introduce an adaptive steplength selection derived by tailoring a limited memory steplength rule, recently developed in the deterministic context, to the stochastic gradient approach. The proposed steplength rule provides values within an interval, whose bounds need to be prefixed by the user. A suitable choice of the interval bounds allows to perform similarly to the standard stochastic gradient method equipped with the best-tuned steplength. Since the setting of the bounds slightly affects the performance, the new rule makes the tuning of the parameters less expensive with respect to the choice of the optimal prefixed steplength in the standard stochastic gradient method. We evaluate the behaviour of the proposed steplength selection in training binary classifiers on well known data sets and by using different loss functions.
2020
3rd Triennial International Conference and Summer School on Numerical Computations: Theory and Algorithms, NUMTA 2019
ita
2019
11973
186
197
Franchini, G.; Ruggiero, V.; Zanni, L.
On the Steplength Selection in Stochastic Gradient Methods / Franchini, G.; Ruggiero, V.; Zanni, L.. - 11973:(2020), pp. 186-197. (Intervento presentato al convegno 3rd Triennial International Conference and Summer School on Numerical Computations: Theory and Algorithms, NUMTA 2019 tenutosi a ita nel 2019) [10.1007/978-3-030-39081-5_17].
File in questo prodotto:
Non ci sono file associati a questo prodotto.
Pubblicazioni consigliate

Licenza Creative Commons
I metadati presenti in IRIS UNIMORE sono rilasciati con licenza Creative Commons CC0 1.0 Universal, mentre i file delle pubblicazioni sono rilasciati con licenza Attribuzione 4.0 Internazionale (CC BY 4.0), salvo diversa indicazione.
In caso di violazione di copyright, contattare Supporto Iris

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11380/1199144
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 5
  • ???jsp.display-item.citation.isi??? 1
social impact