| Authors | مهسا سهیل شمائی,سجاد فتحی هفشجانی |
| Journal | Mathematics |
| Page number | 829 |
| Volume number | 13 |
| IF | 2.3 |
| Paper Type | Full Paper |
| Published At | 2025-03-01 |
| Journal Grade | Scientific - research |
| Journal Type | Electronic |
| Journal Country | Iran, Islamic Republic Of |
| Journal Index | JCR ,SCOPUS |
Abstract
This paper presents a second examination of trigonometric step sizes and their impact on Warm Restart Stochastic Gradient Descent (SGD), an essential optimization technique in deep learning. Building on prior work with cosine-based step sizes, this study introduces three novel trigonometric step sizes aimed at enhancing warm restart methods. These step sizes are formulated to address the challenges posed by non-smooth and non-convex objective functions, ensuring that the algorithm can converge effectively toward the global minimum. Through rigorous theoretical analysis, we demonstrate that the proposed approach achieves an ????(1????−−√)
convergence rate for smooth non-convex functions and extend the analysis to non-smooth and non-convex scenarios. Experimental evaluations on FashionMNIST, CIFAR10, and CIFAR100 datasets reveal significant improvements in test accuracy, including a notable 2.14%
increase on CIFAR100 compared to existing warm restart strategies. These results underscore the effectiveness of trigonometric step sizes in enhancing optimization performance for deep learning models.