In linear algebra, LU decomposition (where 'LU' stands for 'Lower Upper', and also called LU factorization) factors a matrix as the product of a lower triangular matrix and an upper triangular matrix. The product sometimes includes a permutation matrix as well. The LU decomposition can be viewed as the matrix form of Gaussian elimination. Computers usually solve square systems of linear equations using the LU decomposition, and it is also a key step when inverting a matrix, or computing the determinant of a matrix. The LU decomposition was introduced by mathematician Alan Turing.^{[1]}
Definitions
Let A be a square matrix. An LU factorization refers to the factorization of A, with proper row and/or column orderings or permutations,
into two factors, a lower triangular matrix L and an upper triangular matrix U,
 $A\; =\; LU,\; \backslash ,$
In the lower triangular matrix all elements above the diagonal are zero, in the upper triangular matrix, all the elements below
the diagonal are zero. For example, for a 3by3 matrix A, its LU decomposition looks like this:
 $$
\begin{bmatrix}
a_{11} & a_{12} & a_{13} \\
a_{21} & a_{22} & a_{23} \\
a_{31} & a_{32} & a_{33} \\
\end{bmatrix} =
\begin{bmatrix}
l_{11} & 0 & 0 \\
l_{21} & l_{22} & 0 \\
l_{31} & l_{32} & l_{33} \\
\end{bmatrix}
\begin{bmatrix}
u_{11} & u_{12} & u_{13} \\
0 & u_{22} & u_{23} \\
0 & 0 & u_{33} \\
\end{bmatrix}.
Without a proper ordering or permutations in the matrix, the factorization may fail to materialize. For example, it is easy to verify (by expanding the matrix multiplication) that $a\_\{11\}\; =\; l\_\{11\}\; u\_\{11\}$. If $a\_\{11\}\; =\; 0$, then at least one of $l\_\{11\}$ and $u\_\{11\}$ has to be zero, which
implies either L or U is singular. This is impossible if A is nonsingular. This is a procedural problem. It can be removed by simply reordering the rows of A
so that the first element of the permuted matrix is nonzero. The same problem in subsequent factorization steps can be removed the same way, see the basic procedure below.
It turns out that a proper permutation in rows (or columns) is sufficient for the LU factorization. The LU factorization with Partial Pivoting refers often to the LU factorization with row permutations only,
 $PA\; =\; LU,\; \backslash ,$
where L and U are again lower and upper triangular matrices, and P is a permutation matrix which, when leftmultiplied to A, reorders the rows of A. It turns out that all square matrices can be factorized in this form,^{[2]} and the factorization is numerically stable in practice.^{[3]} This makes LUP decomposition a useful technique in practice.
An LU factorization with full pivoting involves both row and column permutations,
 $PAQ\; =\; LU,\; \backslash ,$
where L, U and P are defined as before, and Q is a permutation matrix that reorders the columns of A.^{[4]}
An LDU decomposition is a decomposition of the form
 $A\; =\; LDU,\; \backslash ,$
where D is a diagonal matrix and L and U are unit triangular matrices, meaning that all the entries on the diagonals of L and U are one.
Above we required that A be a square matrix, but these decompositions can all be generalized to rectangular matrices as well. In that case, L and P are square matrices which each have the same number of rows as A, while U is exactly the same shape as A. Upper triangular should be interpreted as having only zero entries below the main diagonal, which starts at the upper left corner.
Example
We factorize the following 2by2 matrix:
 $$
\begin{bmatrix}
4 & 3 \\
6 & 3 \\
\end{bmatrix} =
\begin{bmatrix}
l_{11} & 0 \\
l_{21} & l_{22} \\
\end{bmatrix}
\begin{bmatrix}
u_{11} & u_{12} \\
0 & u_{22} \\
\end{bmatrix}.
One way to find the LU decomposition of this simple matrix would be to simply solve the linear equations by inspection. Expanding the matrix multiplication gives
 $l\_\{11\}\; \backslash cdot\; u\_\{11\}\; +\; 0\; \backslash cdot\; 0\; =\; 4$
 $l\_\{11\}\; \backslash cdot\; u\_\{12\}\; +\; 0\; \backslash cdot\; u\_\{22\}\; =\; 3$
 $l\_\{21\}\backslash cdot\; u\_\{11\}\; +\; l\_\{22\}\; \backslash cdot\; 0\; =\; 6$
 $l\_\{21\}\backslash cdot\; u\_\{12\}\; +\; l\_\{22\}\; \backslash cdot\; u\_\{22\}\; =\; 3.$
This system of equations is underdetermined. In this case any two nonzero elements of L and U matrices are parameters of the solution and can be set arbitrarily to any nonzero value. Therefore to find the unique LU decomposition, it is necessary to put some restriction on L and U matrices. For example, we can conveniently require the lower triangular matrix L to be a unit triangular matrix (i.e. set all the entries of its main diagonal to ones). Then the system of equations has the following solution:
 $l\_\{21\}\; =\; 1.5$
 $u\_\{11\}\; =\; 4$
 $u\_\{12\}\; =\; 3$
 $u\_\{22\}\; =\; 1.5.$
Substituting these values into the LU decomposition above yields
 $$
\begin{bmatrix}
4 & 3 \\
6 & 3 \\
\end{bmatrix} =
\begin{bmatrix}
1 & 0 \\
1.5 & 1 \\
\end{bmatrix}
\begin{bmatrix}
4 & 3 \\
0 & 1.5 \\
\end{bmatrix}.
Existence and uniqueness
Square matrices
Any square matrix $A$ admits an LUP factorization.^{[2]} If $A$ is invertible, then it admits an LU (or LDU) factorization if and only if all its leading principal minors are nonzero.^{[5]} If $A$ is a singular matrix of rank $k$, then it admits an LU factorization if the first $k$ leading principal minors are nonzero, although the converse is not true.^{[6]}
If a square, invertible matrix has an LDU factorization, then it is unique.^{[5]} In that case, the LU factorization is also unique if we require that the diagonal of $L$ (or $U$) consists of ones.
Symmetric positive definite matrices
If A is a symmetric (or Hermitian, if A is complex) positive definite matrix, we can arrange matters so that U is the conjugate transpose of L. That is, we can write A as
 $A\; =\; L\; L^\{*\}.\; \backslash ,$
This decomposition is called the Cholesky decomposition. The Cholesky decomposition always exists and is unique. Furthermore, computing the Cholesky decomposition is more efficient and numerically more stable than computing some other LU decompositions.
General matrices
For a (not necessarily invertible) matrix over any field, the exact necessary and sufficient conditions under which it has an LU factorization are known. The conditions are expressed in terms of the ranks of certain submatrices. The Gaussian elimination algorithm for obtaining LU decomposition has also been extended to this most general case.^{[7]}
Algorithms
The LU decomposition is basically a modified form of Gaussian elimination. We transform the matrix A into an upper triangular matrix U by eliminating the entries below the main diagonal. The Doolittle algorithm does the elimination column by column starting from the left, by multiplying A to the left with atomic lower triangular matrices. It results in a unit lower triangular matrix and an upper triangular matrix. The Crout algorithm is slightly different and constructs a lower triangular matrix and a unit upper triangular matrix.
Computing the LU decomposition using either of these algorithms requires 2n^{3} / 3 floating point operations, ignoring lower order terms. Partial pivoting adds only a quadratic term; this is not the case for full pivoting.^{[8]}
Closed formula
When an LDU factorization exists and is unique there is a closed (explicit) formula for the elements of L, D, and U in terms of ratios of determinants of certain submatrices of the original matrix A.^{[9]} In particular, $D\_1\; =\; A\_\{1,1\}$ and for $i\; =\; 2,\; \backslash ldots,\; n$, $D\_i$ is the ratio of the $i^\{th\}$ principal submatrix to the $(i1)^\{th\}$ principal submatrix.
Doolittle algorithm
Given an N × N matrix
 $$
A= (a_{n,n})
we define
 $A^\{(0)\}:=\; A$
and then we iterate n = 1,...,N1 as follows.
We eliminate the matrix elements below the main diagonal in the nth column
of A^{(n1)}
by adding to the ith row of this matrix the nth row multiplied
by
 $l\_\{i,n\}:=\; \backslash frac\{a\_\{i,n\}^\{(n1)\}\}\{a\_\{n,n\}^\{(n1)\}\}$
for $i\; =\; n+1,\backslash ldots,N$. This can be done by
multiplying A^{(n1)} to the left with the
lower triangular matrix
 $$
L_n =
\begin{pmatrix}
1 & & & & & 0 \\
& \ddots & & & & \\
& & 1 & & & \\
& & l_{n+1,n} & \ddots & & \\
& & \vdots & & \ddots & \\
0 & & l_{N,n} & & & 1 \\
\end{pmatrix}.
We set
 $A^\{(n)\}:=\; L\_n\; A^\{(n1)\}.$
After N1 steps, we eliminated all the matrix elements below the main diagonal, so we obtain an upper triangular matrix
A^{(N1)}. We find the decomposition
 $$
A = L_{1}^{1} L_{1} A^{(0)}
L_{1}^{1} A^{(1)} = L_{1}^{1} L_{2}^{1} L_{2} A^{(1)}
L_{1}^{1}L_{2}^{1} A^{(2)} =\ldots = L_{1}^{1} \ldots L_{N1}^{1} A^{(N1)}.
Denote the upper triangular matrix
A^{(N1)} by U, and $L=L\_\{1\}^\{1\}\; \backslash ldots\; L\_\{N1\}^\{1\}$. Because the inverse of a lower triangular matrix L_{n} is again a lower triangular matrix, and the multiplication of two lower triangular matrices is again a lower triangular matrix, it follows that L is a lower triangular matrix.
Moreover, it can be seen that
 $$
L =
\begin{pmatrix}
1 & & & & & 0 \\
l_{2,1} & \ddots & & & & \\
& & 1 & & & \\
\vdots & & l_{n+1,n} & \ddots & & \\
& & \vdots & & 1 & \\
l_{N,1} & & l_{N,n} & & l_{N,N1} & 1 \\
\end{pmatrix}.
We obtain $A=LU$.
It is clear that in order for this algorithm to work, one needs to have $a\_\{n,n\}^\{(n1)\}\backslash not=0$
at each step (see the definition of $l\_\{i,n\}$). If this assumption fails at some point, one needs to interchange nth row with another row below it before continuing. This is why the LU decomposition in general looks like $P^\{1\}A\; =\; L\; U$.
Crout and LUP algorithms
The LUP decomposition algorithm by Cormen et al. generalizes Crout matrix decomposition. It can be described as follows.
 If $A$ has a nonzero entry in its first row, then take a permutation matrix $P\_1$ such that $A\; P\_1$ has a nonzero entry in its upper left corner. Otherwise, take for $P\_1$ the identity matrix. Let $A\_1\; =\; A\; P\_1$.
 Let $A\_2$ be the matrix that one gets from $A\_1$ by deleting both the first row and the first column. Decompose $A\_2\; =\; L\_2\; U\_2\; P\_2$ recursively. Make $L$ from $L\_2$ by first adding a zero row above and then adding the first column of $A\_1$ at the left.
 Make $U\_3$ from $U\_2$ by first adding a zero row above and a zero column at the left and then replacing the upper left entry (which is 0 at this point) by 1. Make $P\_3$ from $P\_2$ in a similar manner and define $A\_3\; =\; A\_1\; /\; P\_3\; =\; A\; P\_1\; /\; P\_3$. Let $P$ be the inverse of $P\_1\; /\; P\_3$.
 At this point, $A\_3$ is the same as $L\; U\_3$, except (possibly) at the first row. If the first row of $A$ is zero, then $A\_3\; =\; L\; U\_3$, since both have first row zero, and $A\; =\; L\; U\_3\; P$ follows, as desired. Otherwise, $A\_3$ and $L\; U\_3$ have the same nonzero entry in the upper left corner, and $A\_3\; =\; L\; U\_3\; U\_1$ for some upper triangular square matrix $U\_1$ with ones on the diagonal ($U\_1$ clears entries of $L\; U\_3$ and adds entries of $A\_3$ by way of the upper left corner). Now $A\; =\; L\; U\_3\; U\_1\; P$ is a decomposition of the desired form.
Theoretical complexity
If two matrices of order n can be multiplied in time M(n), where M(n)≥n^{a} for some a>2, then the LU decomposition can be computed in time O(M(n)).^{[10]} This means, for example, that an O(n^{2.376}) algorithm exists based on the Coppersmith–Winograd algorithm.
Sparse matrix decomposition
Special algorithms have been developed for factorizing large sparse matrices. These algorithms attempt to find sparse factors L and U. Ideally, the cost of computation is determined by the number of nonzero entries, rather than by the size of the matrix.
These algorithms use the freedom to exchange rows and columns to minimize fillin (entries which change from an initial zero to a nonzero value during the execution of an algorithm).
General treatment of orderings that minimize fillin can be addressed using graph theory.
Applications
Solving linear equations
Given a system of linear equations in matrix form
 $A\; x\; =\; b\; ,\; \backslash ,$
we want to solve the equation for x given A and b. Suppose we have already obtained the LUP decomposition of A such that $PA\; =\; LU$, (or the LU composition if one exists, in which case $P\; =\; I$) we can rewrite the equation equivalently as
 $L\; U\; x\; =\; P\; b\; .\; \backslash ,$
In this case the solution is done in two logical steps:
 First, we solve the equation $L\; y\; =\; P\; b$ for y;
 Second, we solve the equation $U\; x\; =\; y$ for x.
Note that in both cases we are dealing with triangular matrices (L and U) which can be solved directly by forward and backward substitution without using the Gaussian elimination process (however we do need this process or equivalent to compute the LU decomposition itself).
The above procedure can be repeatedly applied to solve the equation multiple times for different b. In this case it is faster (and more convenient) to do an LU decomposition of the matrix A once and then solve the triangular matrices for the different b, rather than using Gaussian elimination each time. The matrices L and U could be thought to have "encoded" the Gaussian elimination process.
The cost of solving a system of linear equations is approximately $\backslash frac23\; n^3$ floating point operations if the matrix $A$ has size $n$. This makes it twice as fast as algorithms based on the QR decomposition, which costs about $\backslash frac43\; n^3$ floating point operations when Householder reflections is used. For this reason, the LU decomposition is usually preferred.^{[11]}
Inverting a matrix
When solving systems of equations, b is usually treated as a vector with a length equal to the height of matrix A. Instead of vector b, we have matrix B, where B is an nbyp matrix, so that we are trying to find a matrix X (also a nbyp matrix):
 $$
A X = L U X = B.
We can use the same algorithm presented earlier to solve for each column of matrix X. Now suppose that B is the identity matrix of size n. It would follow that the result X must be the inverse of A.^{[12]}
Computing the determinant
Given the LUP decomposition $A\; =\; P^\{1\}\; L\; U$ of a square matrix A, the determinant of A can be computed straightforwardly as
 $\backslash det(A)\; =\; \backslash det(P^\{1\})\; \backslash det(L)\; \backslash det(U)\; =\; (1)^S\; \backslash left(\; \backslash prod\_\{i=1\}^n\; l\_\{ii\}\; \backslash right)\; \backslash left(\; \backslash prod\_\{i=1\}^n\; u\_\{ii\}\; \backslash right)\; .$
The second equation follows from the fact that the determinant of a triangular matrix is simply the product of its diagonal entries, and that the determinant of a permutation matrix is equal to (−1)^{S} where S is the number of row exchanges in the decomposition.
The same method readily applies to LU decomposition by setting P to the identity matrix.
See also
Notes
References
 .
 .
 .
 . See Section 3.5.
 .
 .
 .
 .
 .
External links
References
 LU decomposition on MathWorld.
 LU decomposition on MathLinux.
 California State University, Fullerton
 LU decomposition at Holistic Numerical Methods Institute
 LU matrix factorization. MATLAB reference.
Computer code
 LAPACK is a collection of FORTRAN subroutines for solving dense linear algebra problems
 ALGLIB includes a partial port of the LAPACK to C++, C#, Delphi, etc.
 University of Dayton
 C code, Mathematics Source Library
 LU in X10
Online resources
 WebApp descriptively solving systems of linear equations with LU Decomposition
 Matrix Calculator, bluebit.gr
 LU Decomposition Tool, unibonn.de
 The Wolfram Demonstrations Project, 2007.


 Key concepts  

 Problems  

 Hardware  

 Software  


de:Gaußsches Eliminationsverfahren#LRZerlegung
This article was sourced from Creative Commons AttributionShareAlike License; additional terms may apply. World Heritage Encyclopedia content is assembled from numerous content providers, Open Access Publishing, and in compliance with The Fair Access to Science and Technology Research Act (FASTR), Wikimedia Foundation, Inc., Public Library of Science, The Encyclopedia of Life, Open Book Publishers (OBP), PubMed, U.S. National Library of Medicine, National Center for Biotechnology Information, U.S. National Library of Medicine, National Institutes of Health (NIH), U.S. Department of Health & Human Services, and USA.gov, which sources content from all federal, state, local, tribal, and territorial government publication portals (.gov, .mil, .edu). Funding for USA.gov and content contributors is made possible from the U.S. Congress, EGovernment Act of 2002.
Crowd sourced content that is contributed to World Heritage Encyclopedia is peer reviewed and edited by our editorial staff to ensure quality scholarly research articles.
By using this site, you agree to the Terms of Use and Privacy Policy. World Heritage Encyclopedia™ is a registered trademark of the World Public Library Association, a nonprofit organization.