Introduction¶
In this chapter, we will focus on conjugate gradient methods, a popular class of iterative algorithms widely used for solving large-scale optimization and linear algebra problems.
The original motivation comes from the study of (square) positive definite linear systems of equations,
where , and . We assume that .
The solution to (1) is of course . So, why does it matter to study this problem? Because, for large , direct computation or even, memory storage become too costly to be used in practice.
Instead, we seek iterative methods to solve (1): this is known as the linear conjugate gradient method, which exploits the specific structure of the problem. (also closely related to least squares problems, as we will recall next)
- Nocedal, J., & Wright, S. J. (2006). Numerical optimization (Second Edition). Springer.