JOURNAL ARTICLE

Stochastic Gradient Descent with Pre-Conditioned Polyak Step-Size

Farshed AbdukhakimovCaiping XiangDmitry KamzolovMartin Takáč

Year: 2024 Journal:   Журнал вычислительной математики и математической физики Vol: 64 (4)Pages: 575-586   Publisher: Russian Academy of Sciences

Abstract

Stochastic Gradient Descent (SGD) is one of the many iterative optimization methods that are widely used in solving machine learning problems. These methods display valuable properties and attract researchers and industrial machine learning engineers with their simplicity. However, one of the weaknesses of this type of methods is the necessity to tune learning rate (step-size) for every loss function and dataset combination to solve an optimization problem and get an efficient performance in a given time budget. Stochastic Gradient Descent with Polyak Step-size (SPS) is a method that offers an update rule that alleviates the need of fine-tuning the learning rate of an optimizer. In this paper, we propose an extension of SPS that employs preconditioning techniques, such as Hutchinson’s method, Adam, and AdaGrad, to improve its performance on badly scaled and/or ill-conditioned datasets.

Keywords:
Stochastic gradient descent Descent (aeronautics) Mathematics Computer science Artificial intelligence Physics Artificial neural network

Metrics

1
Cited By
0.64
FWCI (Field Weighted Citation Impact)
10
Refs
0.71
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Stochastic Gradient Optimization Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Complexity and Algorithms in Graphs
Physical Sciences →  Computer Science →  Computational Theory and Mathematics
Sparse and Compressive Sensing Techniques
Physical Sciences →  Engineering →  Computational Mechanics

Related Documents

JOURNAL ARTICLE

Stochastic Gradient Descent with Preconditioned Polyak Step-Size

Farshed AbdukhakimovCheng XiangDmitry KamzolovMartin Takáč

Journal:   Computational Mathematics and Mathematical Physics Year: 2024 Vol: 64 (4)Pages: 621-634
JOURNAL ARTICLE

Adaptive Polyak Step-Size for Momentum Accelerated Stochastic Gradient Descent With General Convergence Guarantee

Jiawei ZhangCheng JinYuantao Gu

Journal:   IEEE Transactions on Signal Processing Year: 2025 Vol: 73 Pages: 462-476
JOURNAL ARTICLE

Accelerated stochastic gradient descent with step size selection rules

Zhuang YangCheng WangZhemin ZhangJonathan Li

Journal:   Signal Processing Year: 2019 Vol: 159 Pages: 171-186
JOURNAL ARTICLE

New logarithmic step size for stochastic gradient descent

Mahsa Soheil ShamaeeSajad Fathi HafshejaniZeinab Saeidian

Journal:   Frontiers of Computer Science Year: 2024 Vol: 19 (1)
© 2026 ScienceGate Book Chapters — All rights reserved.