Frank-Wolfe Beyond 1/t Convergence
Abstract
We consider smooth convex minimization over compact convex sets, i.e., $\min_{x \in C} f(x)$ with the (vanilla) Frank-Wolfe algorithm. Well-known lower bounds establish a worst-case $Ω(1/t)$ primal-gap barrier in the general smooth convex case, and faster convergence usually requires favorable function properties such as Hölder error bounds or strong convexity. We present a new Local Dual Sharpness (LDS) condition, essentially a property of the feasible region and its LMO, under which the Frank-...
Description / Details
We consider smooth convex minimization over compact convex sets, i.e., with the (vanilla) Frank-Wolfe algorithm. Well-known lower bounds establish a worst-case primal-gap barrier in the general smooth convex case, and faster convergence usually requires favorable function properties such as Hölder error bounds or strong convexity. We present a new Local Dual Sharpness (LDS) condition, essentially a property of the feasible region and its LMO, under which the Frank-Wolfe algorithm converges in for any smooth convex function, ruling out an lower bound under LDS. The condition is a generalization (and localization) of uniform convexity of sets and it is satisfied by any uniformly convex set. To our knowledge, this is the first unconditional convergence result for uniformly convex sets. Combining LDS with stronger function properties, e.g., a local variant of Hölder error bounds, allows us to quantify the actual rates.
Source: arXiv:2604.28006v1 - http://arxiv.org/abs/2604.28006v1 PDF: https://arxiv.org/pdf/2604.28006v1 Original Link: http://arxiv.org/abs/2604.28006v1
Please sign in to join the discussion.
No comments yet. Be the first to share your thoughts!
May 1, 2026
Mathematics
Mathematics
0