Skip to article frontmatterSkip to article content

1.5Standard optimization problems

Several optimization problems are so frequently encountered that they are given specific names. Recall the most general form of optimization problems encountered in this course:

minimizef0(x)subject tofi(x)0i=1,,mhi(x)=0i=1,,p\begin{array}{lll} \text{minimize} & f_0(x) & \\ \text{subject to} & f_i(x) \leq 0 & i = 1, \ldots, m\\ & h_i(x) = 0 & i = 1, \ldots, p \end{array}

For specific choices of objective and constraint functions, one obtains different classes of optimization problems.

Some important examples are:

For each of these cases, we will review the standard form and give some examples.

Linear programming

When both the objective function and constraints are affine, one obtains a Linear Program (LP).

LP general form

minimizecx+dsubject toAx=b,Gxh\begin{split} \minimize\quad & c^\top x + d \\ \text{subject to}\quad& Ax= b,\\ &Gx\leq h \end{split}

Comments

Example: diet problem

This example is reproduced from Boyd & Vandenberghe (2004, Section 4.3.1.).

A healthy diet contains mm different nutrients in quantities at least equal to b1,,bmb_1, \ldots, b_m. We can compose such a diet by choosing quantities of nn different foods. One unit of food jj contains aija_{ij} of nutrient ii and has a cost of cjc_j.

Question: Write the optimization problem corresponding to choosing the cheapest healthy diet.

Quadratic programming

When the objective function is quadratic and the constraints are affine, we obtain a Quadratic Program (QP).

QP general form

minimize12xPx+qx+rsubject toAx=b,Gxh\begin{split} \minimize\quad & \frac{1}{2}x^\top P x + q^\top x + r \\ \text{subject to}\quad& Ax= b,\\ &Gx\leq h \end{split}

where PP is a symmetric matrix (but not necessarily positive semidefinite).

Comments

Example: Gaussian maximum likelihood estimation

We want to estimate the mean θRn{\theta} \in \mathbb{R}^n of a Gaussian random variable N(θ,Σ)\mathcal{N}({\theta}, {\Sigma}), with known covariance matrix Σ{\Sigma}, from mm i.i.d. measurements y1,,ymN(θ,Σ)y_1, \ldots, y_m \sim \mathcal{N}({\theta}, {\Sigma}).

Recall that

p(yi;θ)=1(2π)n/2Σ1/2exp(12(yiθ)Σ1(yiθ))p(y_i; {\theta}) = \frac{1}{(2\pi)^{n/2}\vert {\Sigma}\vert^{1/2}}\exp\left(-\frac{1}{2}(y_i - {\theta})^\top {\Sigma}^{-1}(y_i - {\theta})\right)

Maximum likelihood (ML) principle

find θarg maxθRnL(θ;y1,,ym):=p(y1,,ym;θ)\text{find } {\theta}^\star \in \argmax_{{\theta}\in \mathbb{R}^n} \mathcal{L}({{\theta}}; y_1, \ldots, y_m):= p(y_1, \ldots, y_m ; {\theta})

The function L(θ;y1,,ym)\mathcal{L}({\theta}; y_1, \ldots, y_m) is called the likelihood function. The goal of ML estimation is to find the parameter θ{\theta} for which the data y1,,ymy_1, \ldots, y_m has the highest joint probability.

Question: what is the corresponding optimization problem?

Convex programming

LPs and QPs (under some conditions) are instances of convex optimization problems.

In the next chapter, we will precisely define this notion and see why it plays such an important role in optimization problems.

General form of a convex problem

minimizef0(x)subject tofi(x)0,i=1,,maix=bi,i=1,,p\begin{split} \minimize\quad & f_0(x) \\ \text{subject to}\quad& f_i(x) \leq 0, \quad i=1, \ldots, m\\ &a_i^\top x= b_i, \quad i=1, \ldots, p \end{split}

where f0f_0 and f1,,fmf_1, \ldots, f_m are convex functions.

Comments

In this course, we will mostly focus on convex optimization problems.

References
  1. Boyd, S. P., & Vandenberghe, L. (2004). Convex optimization. Cambridge University Press.