Proximal regularization for the saddle point gradient dynamics

Abstract
This paper concerns the solution of a convex optimization problem through the saddle point gradient dynamics. Instead of using the standard Lagrangian as is classical in this method, we consider a regularized Lagrangian obtained through a proximal minimization step.We show that, without assumptions of smoothness or strict convexity in the original problem, the regularized Lagrangian is smooth and leads to globally convergent saddle point dynamics. The method is demonstrated through an application to resource allocation in cloud computing.
En
Thesis note
Thesis degree name
8 p.
Notes
Incluye bibliografía.
Subject
CONVEX OPTIMIZATION, PROXIMAL METHOD, SADDLE POINT DYNAMICS
Type
Preprint
Access the full text
Citation
View in library
Rights license