-
Notifications
You must be signed in to change notification settings - Fork 7
Description
We consider TV-like problem of the form
where
We let
Depending on
Data fidelity loss and metric on D x
Huber:
- Smooth
- prox-easy but
$f \circ A$ is not prox-easy except if$A$ is diagonalizable in Fourier
L1:
- Nonsmooth
- prox-easy but
$f \circ A$ is not prox-easy
L1-L2:
- Only make sense for 2D
- Nonsmooth
- prox-easy but
$f \circ A$ is not prox-easy
MSE:
- Smooth
- prox-easy but
$f \circ A$ is not prox-easy except if$A$ is diagonalizable in Fourier
Note: Huber and L1 can be used either as a data fidelity metric, or as a way to measure the gradient (
Solvers
Direct methods
-
$A = \operatorname{Id}$ or orthogonal design,$D$ is whatever finite diff. -
$f$ is the MSE -
$g$ is the$\ell^{1}$ -norm -
Graph-cut based (Boykov, Veksler, and Zabih 2001) or (Kolmogorov and Zabin 2004) (note: this is the fastest way)
-
Alternative discretization (maybe not relevant)
Forward-Backward on the dual
-
$A = \operatorname{Id}$ or orthogonal design,$D$ is whatever finite diff. -
$f$ is such that$f^{\star}$ is$L$ -smooth with closed-form gradient. -
$g$ is prox-easy.
The starting point is to consider the dual problem of the primal problem reads
Variants:
- Vanilla
- Momentum (Heavy-ball (Polyak 1964), ADAM …)
- Accelerated (Nesterov , FISTA (Beck and Teboulle 2009), …)
- Extrapolation (Celer (Massias, Gramfort, and Salmon 2018))
- Mirror descent (Bregman-prox) see (Ben-Tal and Nemirovski 1987) for a review (I am not a specialist)
Proximal-dual hybrid gradient
The starting point is two possible saddle point problems equivalent \eqref{eq:tv-gen}
if
if
- Alternating Direction Method of Multipliers (ADMM) (Eckstein 1989) or (Gabay and Mercier 1976) a.k.a Douglas–Rachford (Lions and Mercier 1979) (there is a paper of Douglas–Rachford but harder to read in the modern language) on the dual.
- Arrow, Hurwicz (not convergent in theory in general but efficient) (book of 1958)
- PDHG (Esser 2009) or (Pock et al. 2010) (
$1/k$ ) - PDHG with over-relaxation a.k.a Chambolle-Pock (Chambolle and Pock 2011) (
$1/k^{2}$ ) - Condat-Vu (Vũ 2013), (Condat 2013) (useful for the Huber case) #14
Variants:
- Acceleration of PDHG: FISTA-like for strongly convex case (Chambolle and Pock 2011)
- Preconditionning of PDHG: (Chambolle and Pock 2011) for a diagonal preconditionning easy to implement
- Accelerated ADMM