AI & ML Paradigm Challenge

New 'Broximal Alignment' math allows us to find the absolute best solution in complex landscapes without needing 'smooth' data.

April 16, 2026

Original Paper

Broximal Alignment for Global Non-Convex Optimization

Kaja Gruntkowska, Hanmin Li, Xun Qian, Peter Richtárik

arXiv · 2604.13483

The Takeaway

Most AI optimization is 'local'—it finds a nearby valley and stops. This paper introduces a method that can provably find the 'global' minimum (the deepest valley) even in non-convex, non-smooth landscapes. It bypasses the standard requirements for convexity and Lipschitz continuity that have limited optimization theory for decades. This is 'the hammer' for complex machine learning problems where the loss landscape is a jagged mess. It provides a rigorous mathematical path to better model training and system tuning where standard gradient descent fails. It fundamentally upgrades the toolkit for global non-convex optimization.

From the abstract

Most non-convex optimization theory is built around gradient dynamics, leaving global convergence largely unexplored. The dominant paradigm focuses on stationarity, certifying only that the gradient norm vanishes, which is often a weak proxy for actual optimization success. In practice, gradient norms can stagnate or even increase during training, and stationary points may be far from global solutions. In this work, we propose a new framework for global non-convex optimization that avoids gradie