EURO 2024 Copenhagen
Abstract Submission

EURO-Online login

2211. A Novel Criterion for Batch Size Adaptation in Stochastic Gradient Methods

Invited abstract in session TD-3: Data science meets strongly NP-Hard CO , stream Data Science Meets Optimization.

Tuesday, 14:30-16:00
Room: 1005 (building: 202)

Authors (first author is the speaker)

1. Marco Boresta
Istituto di analisi dei sistemi ed informatica "Antonio Ruberti" (IASI), Consiglio Nazionale delle Ricerche (CNR)
2. Alberto De Santis
Sapienza University of Rome
3. Stefano Lucidi
Department of Computer, Control, and Management Science, University of Rome "La Sapienza"

Abstract

In this work, we introduce a novel interpretation of the variance associated with batch gradient estimates in machine learning model training. By reevaluating this variance, we introduce a novel strategy for the dynamic adjustment of batch sizes throughout the training process. This approach diverges from traditional approaches by prioritizing a more detailed representation of variance behavior, which can lead to more efficient and effective training regimes. We test our methodology against a series of problems, benchmarking it against current state-of-the-art methods to demonstrate its advantages and applicability.

Keywords

Status: accepted


Back to the list of papers