KNOWPIA
WELCOME TO KNOWPIA

In mathematics, the **polar decomposition** of a square real or complex matrix is a factorization of the form , where is a unitary matrix and is a positive semi-definite Hermitian matrix ( is an orthogonal matrix and is a positive semi-definite symmetric matrix in the real case), both square and of the same size.^{[1]}

If a real matrix is interpreted as a linear transformation of -dimensional space , the polar decomposition separates it into a rotation or reflection of , and a scaling of the space along a set of orthogonal axes.

The polar decomposition of a square matrix always exists. If is invertible, the decomposition is unique, and the factor will be positive-definite. In that case, can be written uniquely in the form , where is unitary and is the unique self-adjoint logarithm of the matrix .^{[2]} This decomposition is useful in computing the fundamental group of (matrix) Lie groups.^{[3]}

The polar decomposition can also be defined as where is a symmetric positive-definite matrix with the same eigenvalues as but different eigenvectors.

The polar decomposition of a matrix can be seen as the matrix analog of the polar form of a complex number as , where is its absolute value (a non-negative real number), and is a complex number with unit norm (an element of the circle group).

The definition may be extended to rectangular matrices by requiring to be a semi-unitary matrix and to be a positive-semidefinite Hermitian matrix. The decomposition always exists and is always unique. The matrix is unique if and only if has full rank. ^{[4]}

A real square matrix can be interpreted as the linear transformation of that takes a column vector to . Then, in the polar decomposition , the factor is an real orthonormal matrix. The polar decomposition then can be seen as expressing the linear transformation defined by into a scaling of the space along each eigenvector of by a scale factor (the action of ), followed by a rotation of (the action of ).

Alternatively, the decomposition expresses the transformation defined by as a rotation ( ) followed by a scaling ( ) along certain orthogonal directions. The scale factors are the same, but the directions are different.

The polar decomposition of the complex conjugate of is given by Note that

The positive-semidefinite matrix *P* is always unique, even if *A* is singular, and is denoted as

In terms of the singular value decomposition (SVD) of , , one has

One can also decompose in the form

The **polar decomposition** of a square invertible real matrix is of the form

The matrix with polar decomposition is normal if and only if and commute: , or equivalently, they are simultaneously diagonalizable.

The core idea behind the construction of the polar decomposition is similar to that used to compute the singular-value decomposition.

If is normal, then it is unitarily equivalent to a diagonal matrix: for some unitary matrix and some diagonal matrix . This makes the derivation of its polar decomposition particularly straightforward, as we can then write

The polar decomposition is thus , with and diagonal in the eigenbasis of and having eigenvalues equal to the phases and absolute values of those of , respectively.

From the singular-value decomposition, it can be shown that a matrix is invertible if and only if (equivalently, ) is. Moreover, this is true if and only if the eigenvalues of are all not zero.^{[6]}

In this case, the polar decomposition is directly obtained by writing

In this expression, is unitary because is. To show that also is unitary, we can use the SVD to write , so that

Yet another way to directly show the unitarity of is to note that, writing the SVD of in terms of rank-1 matrices as , where are the singular values of , we have

Note how, from the above construction, it follows that *the unitary matrix in the polar decomposition of an invertible matrix is uniquely defined*.

The SVD of a square matrix reads , with unitary matrices, and a diagonal, positive semi-definite matrix. By simply inserting an additional pair of s or s, we obtain the two forms of the polar decomposition of :

As an explicit example of this more general case, consider the SVD of the following matrix:

The **polar decomposition** of any bounded linear operator *A* between complex Hilbert spaces is a canonical factorization as the product of a partial isometry and a non-negative operator.

The polar decomposition for matrices generalizes as follows: if *A* is a bounded linear operator then there is a unique factorization of *A* as a product *A* = *UP* where *U* is a partial isometry, *P* is a non-negative self-adjoint operator and the initial space of *U* is the closure of the range of *P*.

The operator *U* must be weakened to a partial isometry, rather than unitary, because of the following issues. If *A* is the one-sided shift on *l*^{2}(**N**), then |*A*| = {*A ^{*}A*}

The existence of a polar decomposition is a consequence of Douglas' lemma:

**Lemma** — If *A*, *B* are bounded operators on a Hilbert space *H*, and *A ^{*}A* ≤

The operator *C* can be defined by *C*(*Bh*) := *Ah* for all *h* in *H*, extended by continuity to the closure of *Ran*(*B*), and by zero on the orthogonal complement to all of *H*. The lemma then follows since *A ^{*}A* ≤

In particular. If *A ^{*}A* =

When *H* is finite-dimensional, *U* can be extended to a unitary operator; this is not true in general (see example above). Alternatively, the polar decomposition can be shown using the operator version of singular value decomposition.

By property of the continuous functional calculus, |*A*| is in the C*-algebra generated by *A*. A similar but weaker statement holds for the partial isometry: *U* is in the von Neumann algebra generated by *A*. If *A* is invertible, the polar part *U* will be in the C*-algebra as well.

If *A* is a closed, densely defined unbounded operator between complex Hilbert spaces then it still has a (unique) **polar decomposition**

The proof uses the same lemma as above, which goes through for unbounded operators in general. If dom(*A ^{*}A*) = dom(

If an unbounded operator *A* is affiliated to a von Neumann algebra **M**, and *A* = *UP* is its polar decomposition, then *U* is in **M** and so is the spectral projection of *P*, 1_{B}(*P*), for any Borel set *B* in [0, ∞).

The polar decomposition of quaternions **H** depends on the unit 2-dimensional sphere of square roots of minus one, known as right versors. Given any *r* on this sphere, and an angle −π < *a* ≤ π, the versor is on the unit 3-sphere of **H**. For *a* = 0 and *a* = π, the versor is 1 or −1 regardless of which *r* is selected. The norm *t* of a quaternion *q* is the Euclidean distance from the origin to *q*. When a quaternion is not just a real number, then there is a *unique* polar decomposition:

- . Here
*r*,*a*,*t*are all uniquely determined such that r is a right versor (*r*^{2}= –1),*a*satisfies 0 <*a*< π, and*t*> 0 .

In the Cartesian plane, alternative planar ring decompositions arise as follows:

- If
*x*≠ 0,*z*=*x*(1 + ε(*y*/*x*)) is a polar decomposition of a dual number*z*=*x*+*yε*, where*ε*^{2}= 0; i.e.,*ε*is nilpotent. In this polar decomposition, the unit circle has been replaced by the line*x*= 1, the polar angle by the slope*y*/*x*, and the radius*x*is negative in the left half-plane. - If
*x*^{2}≠*y*^{2}, then the unit hyperbola*x*^{2}−*y*^{2}= 1 and its conjugate*x*^{2}−*y*^{2}= −1 can be used to form a polar decomposition based on the branch of the unit hyperbola through (1, 0). This branch is parametrized by the hyperbolic angle*a*and is written*j*^{2}= +1 and the arithmetic^{[7]}of split-complex numbers is used. The branch through (−1, 0) is traced by −*e*^{aj}. Since the operation of multiplying by j reflects a point across the line*y*=*x*, the second hyperbola has branches traced by*je*^{aj}or −*je*^{aj}. Therefore a point in one of the quadrants has a polar decomposition in one of the forms:*j*, −*j*} has products that make it isomorphic to the Klein four-group. Evidently polar decomposition in this case involves an element from that group.

To compute an approximation of the polar decomposition *A* = *UP*, usually the unitary factor *U* is approximated.^{[8]}^{[9]} The iteration is based on Heron's method for the square root of *1* and computes, starting from , the sequence

The combination of inversion and Hermite conjugation is chosen so that in the singular value decomposition, the unitary factors remain the same and the iteration reduces to Heron's method on the singular values.

This basic iteration may be refined to speed up the process:

- Every step or in regular intervals, the range of the singular values of is estimated and then the matrix is rescaled to to center the singular values around
*1*. The scaling factor is computed using matrix norms of the matrix and its inverse. Examples of such scale estimates are: - The QR decomposition can be used in a preparation step to reduce a singular matrix
*A*to a smaller regular matrix, and inside every step to speed up the computation of the inverse. - Heron's method for computing roots of can be replaced by higher order methods, for instance based on Halley's method of third order, resulting in
*A*.

**^**Hall 2015 Section 2.5**^**Hall 2015 Theorem 2.17**^**Hall 2015 Section 13.3**^**Higham, Nicholas J.; Schreiber, Robert S. (1990). "Fast polar decomposition of an arbitrary matrix".*SIAM J. Sci. Stat. Comput*.**11**(4). Philadelphia, PA, USA: Society for Industrial and Applied Mathematics: 648–655. CiteSeerX 10.1.1.111.9239. doi:10.1137/0911038. ISSN 0196-5204. S2CID 14268409.**^**Hall 2015 Lemma 2.18**^**Note how this implies, by the positivity of , that the eigenvalues are all real and strictly positive.**^**Sobczyk, G.(1995) "Hyperbolic Number Plane",*College Mathematics Journal 26:268–80***^**Higham, Nicholas J. (1986). "Computing the polar decomposition with applications".*SIAM J. Sci. Stat. Comput*.**7**(4). Philadelphia, PA, USA: Society for Industrial and Applied Mathematics: 1160–1174. CiteSeerX 10.1.1.137.7354. doi:10.1137/0907079. ISSN 0196-5204.**^**Byers, Ralph; Hongguo Xu (2008). "A New Scaling for Newton's Iteration for the Polar Decomposition and its Backward Stability".*SIAM J. Matrix Anal. Appl*.**30**(2). Philadelphia, PA, USA: Society for Industrial and Applied Mathematics: 822–843. CiteSeerX 10.1.1.378.6737. doi:10.1137/070699895. ISSN 0895-4798.

- Conway, J.B.: A Course in Functional Analysis. Graduate Texts in Mathematics. New York: Springer 1990
- Douglas, R.G.: On Majorization, Factorization, and Range Inclusion of Operators on Hilbert Space. Proc. Amer. Math. Soc.
**17**, 413–415 (1966) - Hall, Brian C. (2015),
*Lie Groups, Lie Algebras, and Representations: An Elementary Introduction*, Graduate Texts in Mathematics, vol. 222 (2nd ed.), Springer, ISBN 978-3319134666. - Helgason, Sigurdur (1978),
*Differential geometry, Lie groups, and symmetric spaces*, Academic Press, ISBN 0-8218-2848-7