The Structure-Adaptive Acceleration of Stochastic Proximal Gradient Algorithms

With Junqi Tang, University of Edinburgh

The Structure-Adaptive Acceleration of Stochastic Proximal Gradient Algorithms

Stochastic gradient methods have become the de-facto techniques in data science, signal processing and machine learning, due to their computational efficiency in large-scale optimization problems. Throughout the past few years, accelerated stochastic gradient algorithms are extensively studied and developed, which are not only excellent numerically, but also worse-case optimal theoretically for convex and smooth objective functions. In many real-world applications, we often consider composite optimization tasks where non-smooth regularizers are used for better estimation or generalization. Such regularizers usually enforce the solutions to have low-dimensional structure, such as sparsity, group-sparsity, low-rank and piece-wise smoothness. In this talk, we present structure-adaptive variants of randomized optimization algorithms, including accelerated variance-reduced SGD , and accelerated proximal coordinate descent, for more efficiently solving large-scale composite optimization problems. These algorithms are tailored to exploit the low-dimensional structure of the solution, by judiciously designed restart schemes according to restricted strong-convexity property of the objective function due to non-smooth regularization. The convergence analysis demonstrates that our approach leads to provably improved iteration complexity, while we also validate the efficiency of our algorithms numerically on large-scale sparse regression tasks.

Add to your calendar or Include in your list

How can mathematics help us to understand the behaviour of ants? Read more about the fanscinating work being carri… https://t.co/iCODvvxqE6 View on Twitter