ExplorerMathematicsMathematics
Research PaperResearchia:202604.16031

Stochastic Trust-Region Methods for Over-parameterized Models

Aike Yang

Abstract

Under interpolation-type assumptions such as the strong growth condition, stochastic optimization methods can attain convergence rates comparable to full-batch methods, but their performance, particularly for SGD, remains highly sensitive to step-size selection. To address this issue, we propose a unified stochastic trust-region framework that eliminates manual step-size tuning and extends naturally to equality-constrained problems. For unconstrained optimization, we develop a first-order stocha...

Submitted: April 16, 2026Subjects: Mathematics; Mathematics

Description / Details

Under interpolation-type assumptions such as the strong growth condition, stochastic optimization methods can attain convergence rates comparable to full-batch methods, but their performance, particularly for SGD, remains highly sensitive to step-size selection. To address this issue, we propose a unified stochastic trust-region framework that eliminates manual step-size tuning and extends naturally to equality-constrained problems. For unconstrained optimization, we develop a first-order stochastic trust-region algorithm and show that, under the strong growth condition, it achieves an iteration and stochastic first-order oracle complexity of O(ε2log(1/ε))O(\varepsilon^{-2} \log(1/\varepsilon)) for finding an ε\varepsilon-stationary point. For equality-constrained problems, we introduce a quadratic-penalty-based stochastic trust-region method with penalty parameter μμ, and establish an iteration and oracle complexity of O(ε4log(1/ε))O(\varepsilon^{-4} \log(1/\varepsilon)) to reach an ε\varepsilon-stationary point of the penalized problem, corresponding to an O(ε)O(\varepsilon)-approximate KKT point of the original constrained problem. Numerical experiments on deep neural network training and orthogonally constrained subspace fitting demonstrate that the proposed methods achieve performance comparable to well-tuned stochastic baselines, while exhibiting stable optimization behavior and effectively handling hard constraints without manual learning-rate scheduling.


Source: arXiv:2604.14017v1 - http://arxiv.org/abs/2604.14017v1 PDF: https://arxiv.org/pdf/2604.14017v1 Original Link: http://arxiv.org/abs/2604.14017v1

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

Access Paper
View Source PDF
Submission Info
Date:
Apr 16, 2026
Topic:
Mathematics
Area:
Mathematics
Comments:
0
Bookmark
Stochastic Trust-Region Methods for Over-parameterized Models | Researchia