Team, Visitors, External Collaborators
Overall Objectives
Research Program
Application Domains
Highlights of the Year
New Software and Platforms
New Results
Bilateral Contracts and Grants with Industry
Partnerships and Cooperations
XML PDF e-pub
PDF e-Pub

Section: New Results

A probabilistic incremental proximal gradient method

Participant: Emilie Chouzenoux (Collaboration: Omer Deniz Akyildiz, Alan Turing Institute, London, UK, Victor Elvira, University of Edinburgh, Joaquin Miguez, Universidad Carlos III de Madrid, Spain)

In the paper [3], we proposed a probabilistic optimization method, named probabilistic incremental proximal gradient (PIPG) method, by developing a probabilistic interpretation of the incremental proximal gradient algorithm. We explicitly model the update rules of the incremental proximal gradient method and develop a systematic approach to propagate the uncertainty of the solution estimate over iterations. The PIPG algorithm takes the form of Bayesian filtering updates for a state-space model constructed by using the cost function. Our framework makes it possible to utilize well-known exact or approximate Bayesian filters, such as Kalman or extended Kalman filters, to solve large scale regularized optimization problems.