Bluesky Facebook Reddit Email

New logarithmic step size for stochastic gradient descent

04.22.24 | Higher Education Press

Apple iPhone 17 Pro

Apple iPhone 17 Pro delivers top performance and advanced cameras for field documentation, data collection, and secure research communications.


The step size, often referred to as the learning rate, plays a pivotal role in optimizing the efficiency of the stochastic gradient descent (SGD) algorithm. In recent times, multiple step size strategies have emerged for enhancing SGD performance. However, a significant challenge associated with these step sizes is related to their probability distribution, denoted as ηt/Σ T t=1 ηt . This distribution has been observed to avoid assigning exceedingly small values to the final iterations. For instance, the widely used cosine step size, while effective in practice, encounters this issue by assigning very low probability distribution values to the last iterations.

To address this challenge, a research team led by M. Soheil Shamaee published their new research on 14 Mar 2024 in Frontiers of Computer Science co-published by Higher Education Press and Springer Nature.

The team introduces a new logarithmic step size for the SGD approach. This new step size has proven to be particularly effective during the final iterations, where it enjoys a significantly higher probability of selection compared to the conventional cosine step size. As a result, the new step size method surpasses the performance of the cosine step size method in these critical concluding iterations, benefiting from their increased likelihood of being chosen as the selected solution. The obtained numerical results serve as a testament to the efficiency of the newly proposed step size, particularly on the FashionMinst, CIFAR10, and CIFAR100 datasets. Additionally, the new logarithmic step size has shown remarkable improvements in test accuracy, achieving a 0.9% increase for the CIFAR100 dataset when utilized with a convolutional neural network (CNN) model.

DOI: 10.1007/s11704-023-3245-z

Frontiers of Computer Science

10.1007/s11704-023-3245-z

Experimental study

Not applicable

New logarithmic step size for stochastic gradient descent

14-Mar-2024

Keywords

Article Information

Contact Information

Rong Xie
Higher Education Press
xierong@hep.com.cn

Source

How to Cite This Article

APA:
Higher Education Press. (2024, April 22). New logarithmic step size for stochastic gradient descent. Brightsurf News. https://www.brightsurf.com/news/LN2Z9O41/new-logarithmic-step-size-for-stochastic-gradient-descent.html
MLA:
"New logarithmic step size for stochastic gradient descent." Brightsurf News, Apr. 22 2024, https://www.brightsurf.com/news/LN2Z9O41/new-logarithmic-step-size-for-stochastic-gradient-descent.html.