Proximal gradient method
Introduction
Convex Optimization is a sub field of Optimization which can produce reliable solutions and can be solved exactly. Many signal processing problems can be formulated as convex optimization problems of form
where are convex functions defined from where some of the functions are non-differentiable, this which rules out our conventional smooth optimization techniques like Steepest decent method, conjugate gradient method etc. The class of algorithms which can solve above optimization problem. These methods proceed by splitting, in that the functions are used individually so as to yield an easily implementable algorithm. They are called proximal because each non smooth function among is involved via its proximity operator. Iterative thresholding, projected Landweber, projected gradient, alternating projections, alternating-direction method of multipliers, alternating split Bregman are special instances of proximal algorithms. Details of proximal methods are discussed in [1]
Notations and Terminology
Let is the -dimensional euclidean space and domain of function . Suppose be the non-empty convex subset of . The indicator function of is defined as
-norm is defined as ( )
The distance from to is defined as
If is closed and convex, the projection of onto is the unique point such that .
Sub differential of is given by
Proximal Gradient methods
One of the widely used convex optimization algorithm is POCS (Projection Onto Convex Sets). This algorithm is employed to recover/synthesize a signal satisfying simultaneously several convex constraints. Let be the indicator function of non-empty closed convex set modeling a constraint. This reduces to convex feasibility problem, which require us to find a solution such that it lies in the intersection of all convex sets . In POCS method each set is incorporated by its projection operator . So in each iteration is updated as
However beyond such problems Projection operators are not appropriate and more general operators are required to tackle them. Among the various generalizations of the notion of a convex projection operator that exist, proximity operators are best suited for other purposes.
Definition
Proximity operators of function at is defined as
For every , the minimization problem,
admits a unique solution which is denoted by .
The proximity operator of is characterized by inclusion
If is differentiable then above equation reduces to
References
- Rockafellar, R. T. (1970). Convex analysis. Princeton: Princeton University Press.
- Combettes, Patrick L.; Pesquet, Jean-Christophe (2011). Springer's Fixed-Point Algorithms for Inverse Problems in Science and Engineering. Vol. 49. pp. 185–212.
Notes
External links
- Stephen Boyd and Lieven Vandenberghe Book , Convex optimization
- EE364a: Convex Optimization I and EE364b: Convex Optimization II, Stanford course homepages
- EE227A: Lieven Vandenberghe Notes Lecture 18