Skip to article frontmatterSkip to article content
Site not loading correctly?

This may be due to an incorrect BASE_URL configuration. See the MyST Documentation for reference.

Introduction

In this chapter, we will focus on conjugate gradient methods, a popular class of iterative algorithms widely used for solving large-scale optimization and linear algebra problems.

The original motivation comes from the study of (square) positive definite linear systems of nn equations,

Qx=pQ x = p

where xRnx \in \mathbb{R}^n, QRn×nQ \in \mathbb{R}^{n\times n} and pRnp\in \mathbb{R}^n. We assume that Q0Q \succ 0.

The solution to (1) is of course x=Q1px^\star = Q^{-1}p. So, why does it matter to study this problem? Because, for large nn, direct computation or even, memory storage become too costly to be used in practice.

Instead, we seek iterative methods to solve (1): this is known as the linear conjugate gradient method, which exploits the specific structure of the problem. (also closely related to least squares problems, as we will recall next)

References
  1. Nocedal, J., & Wright, S. J. (2006). Numerical optimization (Second Edition). Springer.