Schur complement: Difference between revisions
imported>Amirhusenjihed |
|||
| (2 intermediate revisions by 2 users not shown) | |||
| Line 1: | Line 1: | ||
{{Short description|Tool in linear algebra and matrix analysis}} | {{Short description|Tool in linear algebra and matrix analysis}} | ||
The '''Schur complement''' is a key tool in the fields of [[linear algebra]], the theory of [[matrix (mathematics)|matrices]], numerical analysis, and statistics. | The '''Schur complement''' is a key tool in the fields of [[linear algebra]], the theory of [[matrix (mathematics)|matrices]], [[numerical analysis]], and [[statistics]]. | ||
It is defined for a [[block matrix]]. Suppose ''p'', ''q'' are [[nonnegative integers]] such that ''p + q > 0'', and suppose ''A'', ''B'', ''C'', ''D'' are respectively ''p'' × ''p'', ''p'' × ''q'', ''q'' × ''p'', and ''q'' × ''q'' matrices of complex numbers. Let | It is defined for a [[block matrix]]. Suppose ''p'', ''q'' are [[nonnegative integers]] such that ''p + q > 0'', and suppose ''A'', ''B'', ''C'', ''D'' are respectively ''p'' × ''p'', ''p'' × ''q'', ''q'' × ''p'', and ''q'' × ''q'' matrices of complex numbers. Let | ||
| Line 15: | Line 15: | ||
|author=Schur, J. | |author=Schur, J. | ||
|title=Über Potenzreihen die im Inneren des Einheitskreises beschränkt sind | |title=Über Potenzreihen die im Inneren des Einheitskreises beschränkt sind | ||
|journal=J. | |journal=J. Reine U. Angewandte Mathematik | ||
|volume=147 | |volume=147 | ||
|year=1917 | |year=1917 | ||
| Line 91: | Line 91: | ||
: <math> \operatorname{rank}(M) = \operatorname{rank}(D) + \operatorname{rank}\left(A - BD^{-1} C\right)</math> | : <math> \operatorname{rank}(M) = \operatorname{rank}(D) + \operatorname{rank}\left(A - BD^{-1} C\right)</math> | ||
* ([[Haynsworth inertia additivity formula]]) If ''A'' is invertible, then the ''inertia'' of the block matrix ''M'' is equal to the inertia of ''A'' plus the inertia of ''M''/''A''. | * ([[Haynsworth inertia additivity formula]]) If ''A'' is invertible, then the ''inertia'' of the block matrix ''M'' is equal to the inertia of ''A'' plus the inertia of ''M''/''A''. | ||
* (Quotient identity) <math>A/B = ((A/C)/(B/C))</math>.<ref>{{Cite journal |last1=Crabtree |first1=Douglas E. |last2=Haynsworth |first2=Emilie V. |date=1969 |title=An identity for the Schur complement of a matrix |url=https://www.ams.org/proc/1969-022-02/S0002-9939-1969-0255573-1/ |journal=Proceedings of the American Mathematical Society |language=en |volume=22 |issue=2 |pages=364–366 |doi=10.1090/S0002-9939-1969-0255573-1 |s2cid=122868483 |issn=0002-9939|doi-access=free }}</ref> | * (Quotient identity) <math>A/B = ((A/C)/(B/C))</math>.<ref>{{Cite journal |last1=Crabtree |first1=Douglas E. |last2=Haynsworth |first2=Emilie V. |date=1969 |title=An identity for the Schur complement of a matrix |url=https://www.ams.org/proc/1969-022-02/S0002-9939-1969-0255573-1/ |journal=Proceedings of the American Mathematical Society |language=en |volume=22 |issue=2 |pages=364–366 |doi=10.1090/S0002-9939-1969-0255573-1 |s2cid=122868483 |issn=0002-9939|doi-access=free |url-access=subscription }}</ref> | ||
* The Schur complement of a [[Laplacian matrix]] is also a Laplacian matrix.<ref>{{Cite journal |last=Devriendt |first=Karel |date=2022 |title=Effective resistance is more than distance: Laplacians, Simplices and the Schur complement |url=https://linkinghub.elsevier.com/retrieve/pii/S0024379522000039 |journal=Linear Algebra and Its Applications |language=en |volume=639 |pages=24–49 |doi=10.1016/j.laa.2022.01.002|arxiv=2010.04521 |s2cid=222272289 }}</ref> | * The Schur complement of a [[Laplacian matrix]] is also a Laplacian matrix.<ref>{{Cite journal |last=Devriendt |first=Karel |date=2022 |title=Effective resistance is more than distance: Laplacians, Simplices and the Schur complement |url=https://linkinghub.elsevier.com/retrieve/pii/S0024379522000039 |journal=Linear Algebra and Its Applications |language=en |volume=639 |pages=24–49 |doi=10.1016/j.laa.2022.01.002|arxiv=2010.04521 |s2cid=222272289 }}</ref> | ||
== Application to solving linear equations == | == Application to solving linear equations == | ||
The Schur complement arises naturally in solving a system of linear equations such as<ref name="Boyd 2004" /> | The Schur complement arises naturally in solving a [[system of linear equations]] such as<ref name="Boyd 2004" /> | ||
<math> | <math> | ||
| Line 149: | Line 149: | ||
In practice, one needs <math>A</math> to be [[condition number|well-conditioned]] in order for this algorithm to be numerically accurate. | In practice, one needs <math>A</math> to be [[condition number|well-conditioned]] in order for this algorithm to be numerically accurate. | ||
This method is useful in electrical engineering to reduce the dimension of a network's equations. It is especially useful when element(s) of the output vector are zero. For example, when <math>u</math> or <math>v</math> is zero, we can eliminate the associated rows of the coefficient matrix without any changes to the rest of the output vector. If <math>v</math> is null then the above equation for <math>x</math> reduces to <math>x = \left(A^{-1} + A^{-1} B S^{-1} C A^{-1}\right) u</math>, thus reducing the dimension of the coefficient matrix while leaving <math>u</math> unmodified. This is used to advantage in electrical engineering where it is referred to as node elimination or [[Kron reduction]]. | This method is useful in electrical engineering to reduce the dimension of a network's equations. It is especially useful when element(s) of the output vector are zero. For example, when <math>u</math> or <math>v</math> is zero, we can eliminate the associated rows of the [[coefficient matrix]] without any changes to the rest of the output vector. If <math>v</math> is null then the above equation for <math>x</math> reduces to <math>x = \left(A^{-1} + A^{-1} B S^{-1} C A^{-1}\right) u</math>, thus reducing the dimension of the coefficient matrix while leaving <math>u</math> unmodified. This is used to advantage in electrical engineering where it is referred to as node elimination or [[Kron reduction]]. | ||
==Applications to probability theory and statistics== | ==Applications to probability theory and statistics== | ||
| Line 157: | Line 157: | ||
:<math>\Sigma = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right],</math> | :<math>\Sigma = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right],</math> | ||
where <math display="inline">A \in \mathbb{R}^{n \times n}</math> is the covariance matrix of ''X'', <math display="inline">C \in \mathbb{R}^{m \times m}</math> is the covariance matrix of ''Y'' and <math display="inline">B \in \mathbb{R}^{n \times m}</math> is the covariance matrix between ''X'' and ''Y''. | where <math display="inline">A \in \mathbb{R}^{n \times n}</math> is the [[covariance matrix]] of ''X'', <math display="inline">C \in \mathbb{R}^{m \times m}</math> is the covariance matrix of ''Y'' and <math display="inline">B \in \mathbb{R}^{n \times m}</math> is the covariance matrix between ''X'' and ''Y''. | ||
Then the [[Conditional variance|conditional covariance]] of ''X'' given ''Y'' is the Schur complement of ''C'' in <math display="inline">\Sigma</math>:<ref name="von Mises 1964">{{cite book |title=Mathematical theory of probability and statistics |url=https://archive.org/details/mathematicaltheo0057vonm |url-access=registration |first=Richard |last=von Mises |year=1964|publisher=Academic Press| chapter=Chapter VIII.9.3|isbn=978-1483255385}}</ref> | Then the [[Conditional variance|conditional covariance]] of ''X'' given ''Y'' is the Schur complement of ''C'' in <math display="inline">\Sigma</math>:<ref name="von Mises 1964">{{cite book |title=Mathematical theory of probability and statistics |url=https://archive.org/details/mathematicaltheo0057vonm |url-access=registration |first=Richard |last=von Mises |year=1964|publisher=Academic Press| chapter=Chapter VIII.9.3|isbn=978-1483255385}}</ref> | ||
| Line 166: | Line 166: | ||
\end{align}</math> | \end{align}</math> | ||
If we take the matrix <math>\Sigma</math> above to be, not a covariance of a random vector, but a ''sample'' covariance, then it may have a [[Wishart distribution]]. In that case, the Schur complement of ''C'' in <math>\Sigma</math> also has a Wishart distribution.{{Citation needed|date=January 2014}} | If we take the matrix <math>\Sigma</math> above to be, not a covariance of a [[Multivariate random variable|random vector]], but a ''sample'' covariance, then it may have a [[Wishart distribution]]. In that case, the Schur complement of ''C'' in <math>\Sigma</math> also has a Wishart distribution.{{Citation needed|date=January 2014}} | ||
== Conditions for positive definiteness and semi-definiteness == | == Conditions for positive definiteness and semi-definiteness == | ||
Let ''X'' be a symmetric matrix of real numbers given by | Let ''X'' be a [[symmetric matrix]] of real numbers given by | ||
<math display="block">X = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right].</math> | <math display="block">X = \left[\begin{matrix} A & B \\ B^\mathrm{T} & C\end{matrix}\right].</math> | ||
Then | Then by the [[Haynsworth inertia additivity formula]], we find | ||
* If ''A'' is invertible, then ''X'' is positive definite if and only if ''A'' and its complement ''X/A'' are both positive definite:<ref name="Zhang 2005"></ref>{{rp|34}} | * If ''A'' is invertible, then ''X'' is positive definite if and only if ''A'' and its complement ''X/A'' are both positive definite:<ref name="Zhang 2005"></ref>{{rp|34}} | ||
:<math display="block">X \succ 0 \Leftrightarrow A \succ 0, X/A = C - B^\mathrm{T} A^{-1} B \succ 0.</math> | :<math display="block">X \succ 0 \Leftrightarrow A \succ 0, X/A = C - B^\mathrm{T} A^{-1} B \succ 0.</math> | ||
| Line 182: | Line 182: | ||
:<math display="block">\text{If } C \succ 0,\text{ then } X \succeq 0 \Leftrightarrow X/C = A - B C^{-1} B^\mathrm{T} \succeq 0.</math> | :<math display="block">\text{If } C \succ 0,\text{ then } X \succeq 0 \Leftrightarrow X/C = A - B C^{-1} B^\mathrm{T} \succeq 0.</math> | ||
The first and third statements can be derived<ref name="Boyd 2004">Boyd, S. and Vandenberghe, L. (2004), "Convex Optimization", Cambridge University Press (Appendix A.5.5)</ref> by considering the minimizer of the quantity | The first and third statements can also be derived<ref name="Boyd 2004">Boyd, S. and Vandenberghe, L. (2004), "Convex Optimization", Cambridge University Press (Appendix A.5.5)</ref> by considering the minimizer of the quantity | ||
<math display="block">u^\mathrm{T} A u + 2 v^\mathrm{T} B^\mathrm{T} u + v^\mathrm{T} C v, \,</math> | <math display="block">u^\mathrm{T} A u + 2 v^\mathrm{T} B^\mathrm{T} u + v^\mathrm{T} C v, \,</math> | ||
as a function of ''v'' (for fixed ''u''). | as a function of ''v'' (for fixed ''u''). | ||
| Line 193: | Line 193: | ||
and similarly for positive semi-definite matrices, the second (respectively fourth) statement is immediate from the first (resp. third) statement. | and similarly for positive semi-definite matrices, the second (respectively fourth) statement is immediate from the first (resp. third) statement. | ||
There is also a sufficient and necessary condition for the positive semi-definiteness of ''X'' in terms of a generalized Schur complement.<ref name="Zhang 2005" /> Precisely, | There is also a sufficient and [[Necessity and sufficiency|necessary condition]] for the positive semi-definiteness of ''X'' in terms of a generalized Schur complement.<ref name="Zhang 2005" /> Precisely, | ||
* <math>X \succeq 0 \Leftrightarrow A \succeq 0, C - B^\mathrm{T} A^g B \succeq 0, \left(I - A A^{g}\right)B = 0 \, </math> and | * <math>X \succeq 0 \Leftrightarrow A \succeq 0, C - B^\mathrm{T} A^g B \succeq 0, \left(I - A A^{g}\right)B = 0 \, </math> and | ||
* <math>X \succeq 0 \Leftrightarrow C \succeq 0, A - B C^g B^\mathrm{T} \succeq 0, \left(I - C C^g\right)B^\mathrm{T} = 0, </math> | * <math>X \succeq 0 \Leftrightarrow C \succeq 0, A - B C^g B^\mathrm{T} \succeq 0, \left(I - C C^g\right)B^\mathrm{T} = 0, </math> | ||
Latest revision as of 06:10, 29 October 2025
Template:Short description The Schur complement is a key tool in the fields of linear algebra, the theory of matrices, numerical analysis, and statistics.
It is defined for a block matrix. Suppose p, q are nonnegative integers such that p + q > 0, and suppose A, B, C, D are respectively p × p, p × q, q × p, and q × q matrices of complex numbers. Let so that M is a (p + q) × (p + q) matrix.
If D is invertible, then the Schur complement of the block D of the matrix M is the p × p matrix defined by If A is invertible, the Schur complement of the block A of the matrix M is the q × q matrix defined by In the case that A or D is singular, substituting a generalized inverse for the inverses on M/A and M/D yields the generalized Schur complement.
The Schur complement is named after Issai Schur[1] who used it to prove Schur's lemma, although it had been used previously.[2] Emilie Virginia Haynsworth was the first to call it the Schur complement.[3] The Schur complement is sometimes referred to as the Feshbach map after a physicist Herman Feshbach.[4]
Background
The Schur complement arises when performing a block Gaussian elimination on the matrix M. In order to eliminate the elements below the block diagonal, one multiplies the matrix M by a block lower triangular matrix on the right as follows: where Ip denotes a p×p identity matrix. As a result, the Schur complement appears in the upper-left p×p block.
Continuing the elimination process beyond this point (i.e., performing a block Gauss–Jordan elimination), leads to an LDU decomposition of M, which reads Thus, the inverse of M may be expressed involving D−1 and the inverse of Schur's complement, assuming it exists, as The above relationship comes from the elimination operations that involve D−1 and M/D. An equivalent derivation can be done with the roles of A and D interchanged. By equating the expressions for M−1 obtained in these two different ways, one can establish the matrix inversion lemma, which relates the two Schur complements of M: M/D and M/A (see "Derivation from LDU decomposition" in Template:Slink).
Properties
- If p and q are both 1 (i.e., A, B, C and D are all scalars), we get the familiar formula for the inverse of a 2-by-2 matrix:
- provided that AD − BC is non-zero.
- In general, if A is invertible, then
- whenever this inverse exists.
- (Schur's formula) When A, respectively D, is invertible, the determinant of M is also clearly seen to be given by
- , respectively
- ,
- which generalizes the determinant formula for 2 × 2 matrices.
- (Guttman rank additivity formula) If D is invertible, then the rank of M is given by
- (Haynsworth inertia additivity formula) If A is invertible, then the inertia of the block matrix M is equal to the inertia of A plus the inertia of M/A.
- (Quotient identity) .[5]
- The Schur complement of a Laplacian matrix is also a Laplacian matrix.[6]
Application to solving linear equations
The Schur complement arises naturally in solving a system of linear equations such as[7]
.
Assuming that the submatrix is invertible, we can eliminate from the equations, as follows.
Substituting this expression into the second equation yields
We refer to this as the reduced equation obtained by eliminating from the original equation. The matrix appearing in the reduced equation is called the Schur complement of the first block in :
- .
Solving the reduced equation, we obtain
Substituting this into the first equation yields
We can express the above two equation as:
Therefore, a formulation for the inverse of a block matrix is:
In particular, we see that the Schur complement is the inverse of the block entry of the inverse of .
In practice, one needs to be well-conditioned in order for this algorithm to be numerically accurate.
This method is useful in electrical engineering to reduce the dimension of a network's equations. It is especially useful when element(s) of the output vector are zero. For example, when or is zero, we can eliminate the associated rows of the coefficient matrix without any changes to the rest of the output vector. If is null then the above equation for reduces to , thus reducing the dimension of the coefficient matrix while leaving unmodified. This is used to advantage in electrical engineering where it is referred to as node elimination or Kron reduction.
Applications to probability theory and statistics
Suppose the random column vectors X, Y live in Rn and Rm respectively, and the vector (X, Y) in Rn + m has a multivariate normal distribution whose covariance is the symmetric positive-definite matrix
where is the covariance matrix of X, is the covariance matrix of Y and is the covariance matrix between X and Y.
Then the conditional covariance of X given Y is the Schur complement of C in :[8]
If we take the matrix above to be, not a covariance of a random vector, but a sample covariance, then it may have a Wishart distribution. In that case, the Schur complement of C in also has a Wishart distribution.Script error: No such module "Unsubst".
Conditions for positive definiteness and semi-definiteness
Let X be a symmetric matrix of real numbers given by Then by the Haynsworth inertia additivity formula, we find
- If A is invertible, then X is positive definite if and only if A and its complement X/A are both positive definite:[2]Template:Rp
- If C is invertible, then X is positive definite if and only if C and its complement X/C are both positive definite:
- If A is positive definite, then X is positive semi-definite if and only if the complement X/A is positive semi-definite:[2]Template:Rp
- If C is positive definite, then X is positive semi-definite if and only if the complement X/C is positive semi-definite:
The first and third statements can also be derived[7] by considering the minimizer of the quantity as a function of v (for fixed u).
Furthermore, since and similarly for positive semi-definite matrices, the second (respectively fourth) statement is immediate from the first (resp. third) statement.
There is also a sufficient and necessary condition for the positive semi-definiteness of X in terms of a generalized Schur complement.[2] Precisely,
- and
where denotes a generalized inverse of .
See also
- Woodbury matrix identity
- Quasi-Newton method
- Haynsworth inertia additivity formula
- Gaussian process
- Total least squares
- Guyan reduction in computational mechanics
References
<templatestyles src="Reflist/styles.css" />
- ↑ Script error: No such module "Citation/CS1".
- ↑ a b c d Script error: No such module "citation/CS1".
- ↑ Haynsworth, E. V., "On the Schur Complement", Basel Mathematical Notes, #BNB 20, 17 pages, June 1968.
- ↑ Script error: No such module "Citation/CS1".
- ↑ Script error: No such module "Citation/CS1".
- ↑ Script error: No such module "Citation/CS1".
- ↑ a b Boyd, S. and Vandenberghe, L. (2004), "Convex Optimization", Cambridge University Press (Appendix A.5.5)
- ↑ Script error: No such module "citation/CS1".
Script error: No such module "Check for unknown parameters".