Mirror descent

In mathematics, mirror descent is an iterative optimization algorithm for finding a local minimum of a differentiable function.

It generalizes algorithms such as gradient descent and multiplicative weights.

Mirror descent was originally proposed by Nemirovski and Yudin in 1983.

[1] In gradient descent with the sequence of learning rates

applied to a differentiable function

and considers the sequence

such that This can be reformulated by noting that In other words,

minimizes the first-order approximation to

with added proximity term

This squared Euclidean distance term is a particular example of a Bregman distance.

Using other Bregman distances will yield other algorithms such as Hedge which may be more suited to optimization over particular geometries.

to optimize over a convex set

We are also given differentiable convex function

-strongly convex with respect to the given norm.

This is called the distance-generating function, and its gradient

is known as the mirror map.

, in each iteration of Mirror Descent: Mirror descent in the online optimization setting is known as Online Mirror Descent (OMD).

Graph of a strictly concave quadratic function with unique maximum.
Optimization computes maxima and minima.