Jump to content

Invertible matrix

From Wikipedia, the free encyclopedia
(Redirected fromInverse matrix)

Inlinear algebra,ann-by-nsquare matrixAis calledinvertible(alsononsingular,nondegenerateor rarelyregular) if there exists ann-by-nsquare matrixBsuch thatwhereIndenotes then-by-nidentity matrixand the multiplication used is ordinarymatrix multiplication.[1]If this is the case, then the matrixBis uniquely determined byA,and is called the(multiplicative)inverseofA,denoted byA−1.Matrix inversionis the process of finding the matrix which when multiplied by the original matrix gives the identity matrix.[2]

Over afield,a square matrix that isnotinvertible is calledsingularordegenerate.A square matrix with entries in a field is singularif and only ifitsdeterminantis zero. Singular matrices are rare in the sense that if a square matrix's entries are randomly selected from any bounded region on thenumber lineorcomplex plane,theprobabilitythat the matrix is singular is 0, that is, it will"almost never"be singular. Non-square matrices, i.e.m-by-nmatrices for whichmn,do not have an inverse. However, in some cases such a matrix may have aleft inverseorright inverse.IfAism-by-nand therankofAis equal ton,(nm), thenAhas a left inverse, ann-by-mmatrixBsuch thatBA=In.IfAhas rankm(mn), then it has a right inverse, ann-by-mmatrixBsuch thatAB=Im.

While the most common case is that of matrices over therealorcomplexnumbers, all these definitions can be given for matrices over anyalgebraic structureequipped withadditionandmultiplication(i.e.rings). However, in the case of a ring beingcommutative,the condition for a square matrix to be invertible is that its determinant is invertible in the ring, which in general is a stricter requirement than it being nonzero. For anoncommutative ring,the usual determinant is not defined. The conditions for existence of left-inverse or right-inverse are more complicated, since a notion of rank does not exist over rings.

The set ofn×ninvertible matrices together with the operation ofmatrix multiplicationand entries from ringRform agroup,thegeneral linear groupof degreen,denotedGLn(R).

Properties[edit]

The invertible matrix theorem[edit]

LetAbe a squaren-by-nmatrix over afieldK(e.g., the fieldof real numbers). The following statements are equivalent, i.e., they are either all true or all false for any given matrix:[3]

  • Ais invertible, i.e. it has an inverse under matrix multiplication, i.e., there exists aBsuch thatAB=In=BA.(In this statement, "invertible" can equivalently be replaced with "left-invertible" or "right-invertible", in which one-sided inverses are considered.)
  • The linear transformation mappingxtoAxis invertible, i.e., has an inverse under function composition. (Here, again, "invertible" can equivalently be replaced with either "left-invertible" or "right-invertible" )
  • ThetransposeATis an invertible matrix.
  • Aisrow-equivalentto then-by-nidentity matrixIn.
  • Aiscolumn-equivalentto then-by-nidentity matrixIn.
  • Ahasnpivot positions.
  • Ahas fullrank:rankA=n.
  • Ahas a trivialkernel:ker(A) = {0}.
  • The linear transformation mappingxtoAxis bijective; that is, the equationAx=bhas exactly one solution for eachbinKn.(Here, "bijective" can equivalently be replaced with "injective"or"surjective")
  • The columns ofAform abasisofKn.(In this statement, "basis" can equivalently be replaced with either "linearly independent set" or "spanning set" )
  • The rows ofAform a basis ofKn.(Similarly, here, "basis" can equivalently be replaced with either "linearly independent set" or "spanning set" )
  • ThedeterminantofAis nonzero:detA≠ 0.(In general, a square matrix over acommutative ringis invertible if and only if its determinant is aunit(i.e. multiplicatively invertible element) of that ring.
  • The number 0 is not aneigenvalueofA.(More generally, a numberis an eigenvalue ofAif the matrixis singular, whereIis the identity matrix.)
  • The matrixAcan be expressed as a finite product ofelementary matrices.

Other properties[edit]

Furthermore, the following properties hold for an invertible matrixA:

  • for nonzero scalark
  • ifAhas orthonormal columns, where+denotes theMoore–Penrose inverseandxis a vector
  • For any invertiblen-by-nmatricesAandB,More generally, ifare invertiblen-by-nmatrices, then

The rows of the inverse matrixVof a matrixUareorthonormalto the columns ofU(and vice versa interchanging rows for columns). To see this, suppose thatUV=VU=Iwhere the rows ofVare denoted asand the columns ofUasforThen clearly, theEuclidean inner productof any twoThis property can also be useful in constructing the inverse of a square matrix in some instances, where a set oforthogonalvectors (but not necessarily orthonormal vectors) to the columns ofUare known. In which case, one can apply the iterativeGram–Schmidt processto this initial set to determine the rows of the inverseV.

A matrix that is its own inverse (i.e., a matrixAsuch thatA=A−1,and consequentlyA2=I), is called aninvolutory matrix.

In relation to its adjugate[edit]

Theadjugateof a matrixAcan be used to find the inverse ofAas follows:

IfAis an invertible matrix, then

In relation to the identity matrix[edit]

It follows from theassociativityof matrix multiplication that if

forfinite squarematricesAandB,then also

[4]

Density[edit]

Over the field of real numbers, the set of singularn-by-nmatrices, considered as asubsetofis anull set,that is, hasLebesguemeasure zero.This is true because singular matrices are the roots of thedeterminantfunction. This is acontinuous functionbecause it is apolynomialin the entries of the matrix. Thus in the language ofmeasure theory,almost alln-by-nmatrices are invertible.

Furthermore, then-by-ninvertible matrices are adenseopen setin thetopological spaceof alln-by-nmatrices. Equivalently, the set of singular matrices isclosedandnowhere densein the space ofn-by-nmatrices.

In practice however, one may encounter non-invertible matrices. And innumerical calculations,matrices which are invertible, but close to a non-invertible matrix, can still be problematic; such matrices are said to beill-conditioned.

Examples[edit]

An example with rank ofn− 1is a non-invertible matrix

We can see the rank of this 2-by-2 matrix is 1, which isn− 1 ≠n,so it is non-invertible.

Consider the following 2-by-2 matrix:

The matrixis invertible. To check this, one can compute that,which is non-zero.

As an example of a non-invertible, or singular, matrix, consider the matrix

The determinant ofis 0, which is anecessary and sufficient conditionfor a matrix to be non-invertible.

Methods of matrix inversion[edit]

Gaussian elimination[edit]

Gaussian eliminationis a useful and easy way to compute the inverse of a matrix. To compute a matrix inverse using this method, anaugmented matrixis first created with the left side being the matrix to invert and the right side being theidentity matrix.Then, Gaussian elimination is used to convert the left side into the identity matrix, which causes the right side to become the inverse of the input matrix.

For example, take the following matrix:

The first step to compute its inverse is to create the augmented matrix

Call the first row of this matrixand the second row.Then, add row 1 to row 2This yields

Next, subtract row 2, multiplied by 3, from row 1which yields

Finally, multiply row 1 by −1and row 2 by 2This yields the identity matrix on the left side and the inverse matrix on the right:

Thus,

The reason it works is that the process of Gaussian elimination can be viewed as a sequence of applying left matrix multiplication using elementary row operations usingelementary matrices(), such as

Applying right-multiplication usingwe getAnd the right sidewhich is the inverse we want.

To obtainwe create the augumented matrix by combiningAwithIand applyingGaussian elimination.The two portions will be transformed using the same sequence of elementary row operations. When the left portion becomesI,the right portion applied the same elementary row operation sequence will becomeA−1.

Newton's method[edit]

A generalization ofNewton's methodas used for amultiplicative inverse algorithmmay be convenient, if it is convenient to find a suitable starting seed:

Victor PanandJohn Reifhave done work that includes ways of generating a starting seed.[5][6]

Newton's method is particularly useful when dealing withfamiliesof related matrices that behave enough like the sequence manufactured for thehomotopyabove: sometimes a good starting point for refining an approximation for the new inverse can be the already obtained inverse of a previous matrix that nearly matches the current matrix, for example, the pair of sequences of inverse matrices used in obtainingmatrix square roots by Denman–Beavers iteration;this may need more than one pass of the iteration at each new matrix, if they are not close enough together for just one to be enough. Newton's method is also useful for "touch up" corrections to the Gauss–Jordan algorithm which has been contaminated by small errors due toimperfect computer arithmetic.

Cayley–Hamilton method[edit]

TheCayley–Hamilton theoremallows the inverse ofAto be expressed in terms ofdet(A),traces and powers ofA:[7]

wherenis size ofA,andtr(A)is thetraceof matrixAgiven by the sum of themain diagonal.The sum is taken oversand the sets of allsatisfying the linearDiophantine equation

The formula can be rewritten in terms of completeBell polynomialsof argumentsas

This is described in more detail underCayley–Hamilton method.

Eigendecomposition[edit]

If matrixAcan be eigendecomposed, and if none of its eigenvalues are zero, thenAis invertible and its inverse is given by

whereQis the square(N×N)matrix whoseith column is theeigenvectorofA,andΛis thediagonal matrixwhose diagonal entries are the corresponding eigenvalues, that is,If Ais symmetric,Qis guaranteed to be anorthogonal matrix,thereforeFurthermore, becauseΛis a diagonal matrix, its inverse is easy to calculate:

Cholesky decomposition[edit]

If matrixAispositive definite,then its inverse can be obtained as

whereLis thelower triangularCholesky decompositionofA,andL*denotes theconjugate transposeofL.

Analytic solution[edit]

Writing the transpose of thematrix of cofactors,known as anadjugate matrix,can also be an efficient way to calculate the inverse ofsmallmatrices, but thisrecursivemethod is inefficient for large matrices. To determine the inverse, we calculate a matrix of cofactors:

so that

where|A|is thedeterminantofA,Cis thematrix of cofactors,andCTrepresents the matrixtranspose.

Inversion of 2 × 2 matrices[edit]

Thecofactor equationlisted above yields the following result for2 × 2matrices. Inversion of these matrices can be done as follows:[8]

This is possible because1/(adbc)is thereciprocalof the determinant of the matrix in question, and the same strategy could be used for other matrix sizes.

The Cayley–Hamilton method gives

Inversion of 3 × 3 matrices[edit]

Acomputationally efficient3 × 3matrix inversion is given by

(where thescalarAis not to be confused with the matrixA).

If the determinant is non-zero, the matrix is invertible, with the entries of the intermediary matrix on the right side above given by

The determinant ofAcan be computed by applying therule of Sarrusas follows:

The Cayley–Hamilton decomposition gives

The general3 × 3inverse can be expressed concisely in terms of thecross productandtriple product.If a matrix(consisting of three column vectors,,,and) is invertible, its inverse is given by

The determinant ofA,det(A),is equal to the triple product ofx0,x1,andx2—the volume of theparallelepipedformed by the rows or columns:

The correctness of the formula can be checked by using cross- and triple-product properties and by noting that for groups, left and right inverses always coincide. Intuitively, because of the cross products, each row ofA–1is orthogonal to the non-corresponding two columns ofA(causing the off-diagonal terms ofbe zero). Dividing by

causes the diagonal entries ofI=A−1Ato be unity. For example, the first diagonal is:

Inversion of 4 × 4 matrices[edit]

With increasing dimension, expressions for the inverse ofAget complicated. Forn= 4,the Cayley–Hamilton method leads to an expression that is still tractable:

Blockwise inversion[edit]

Matrices can also beinverted blockwiseby using the following analytic inversion formula:[9]

(1)

whereA,B,CandDarematrix sub-blocksof arbitrary size. (Amust be square, so that it can be inverted. Furthermore,AandDCA−1Bmust be nonsingular.[10]) This strategy is particularly advantageous ifAis diagonal andDCA−1B(theSchur complementofA) is a small matrix, since they are the only matrices requiring inversion.

This technique was reinvented several times and is due toHans Boltz(1923),[citation needed]who used it for the inversion ofgeodeticmatrices, andTadeusz Banachiewicz(1937), who generalized it and proved its correctness.

Thenullity theoremsays that the nullity ofAequals the nullity of the sub-block in the lower right of the inverse matrix, and that the nullity ofBequals the nullity of the sub-block in the upper right of the inverse matrix.

The inversion procedure that led to Equation (1) performed matrix block operations that operated onCandDfirst. Instead, ifAandBare operated on first, and providedDandABD−1Care nonsingular,[11]the result is

(2)

Equating Equations (1) and (2) leads to

(3)

where Equation (3) is theWoodbury matrix identity,which is equivalent to thebinomial inverse theorem.

IfAandDare both invertible, then the above two block matrix inverses can be combined to provide the simple factorization

(2)

By theWeinstein–Aronszajn identity,one of the two matrices in the block-diagonal matrix is invertible exactly when the other is.

This formula simplifies significantly when the upper right block matrixBis thezero matrix.This formulation is useful when the matricesAandDhave relatively simple inverse formulas (orpseudo inversesin the case where the blocks are not all square. In this special case, the block matrix inversion formula stated in full generality above becomes

If the given invertible matrix is a symmetric matrix with invertible blockAthe following block inverse formula holds[12]

(4)

where.This requires 2 inversions of the half-sized matricesAandSand only 4 multiplications of half-sized matrices, if organized properlytogether with some additions, subtractions, negations and transpositions of negligible complexity. Any matrixhas an associated positive semidefinite, symmetric matrix,which is exactly invertible (and positive definite), if and only ifis invertible. By writingmatrix inversion can be reduced to inverting symmetric matrices and 2 additional matrix multiplications, because thepositive definite matrixsatisfies the invertibility condition for its left upper blockA.

These formulas together allow to construct adivide and conquer algorithmthat uses blockwise inversion of associated symmetric matrices to invert a matrix with the same time complexity as thematrix multiplication algorithmthat is used internally.[12]Research into matrix multiplication complexityshows that there exist matrix multiplication algorithms with a complexity ofO(n2.371552)operations, while the best proven lower bound isΩ(n2logn).[13]

By Neumann series[edit]

If a matrixAhas the property that

thenAis nonsingular and its inverse may be expressed by aNeumann series:[14]

Truncating the sum results in an "approximate" inverse which may be useful as apreconditioner.Note that a truncated series can be accelerated exponentially by noting that the Neumann series is ageometric sum.As such, it satisfies

.

Therefore, only2L− 2matrix multiplications are needed to compute2Lterms of the sum.

More generally, ifAis "near" the invertible matrixXin the sense that

thenAis nonsingular and its inverse is

If it is also the case thatAXhasrank1 then this simplifies to

p-adic approximation[edit]

IfAis a matrix withintegerorrationalentries and we seek a solution inarbitrary-precisionrationals, then ap-adicapproximation method converges to an exact solution inO(n4log2n),assuming standardO(n3)matrix multiplication is used.[15]The method relies on solvingnlinear systems via Dixon's method ofp-adic approximation (each inO(n3log2n)) and is available as such in software specialized in arbitrary-precision matrix operations, for example, in IML.[16]

Reciprocal basis vectors method[edit]

Given ann×nsquare matrix,,withnrows interpreted asnvectors(Einstein summationassumed) where theare a standardorthonormal basisofEuclidean space(), then usingClifford algebra(orgeometric algebra) we compute the reciprocal (sometimes calleddual) column vectors:

as the columns of the inverse matrixNote that, the place ""indicates that""is removed from that place in the above expression for.We then have,whereis theKronecker delta.We also have,as required. If the vectorsare not linearly independent, thenand the matrixis not invertible (has no inverse).

Derivative of the matrix inverse[edit]

Suppose that the invertible matrixAdepends on a parametert.Then the derivative of the inverse ofAwith respect totis given by[17]

To derive the above expression for the derivative of the inverse ofA,one can differentiate the definition of the matrix inverseand then solve for the inverse ofA:

Subtractingfrom both sides of the above and multiplying on the right bygives the correct expression for the derivative of the inverse:

Similarly, ifis a small number then

More generally, if

then,

Given a positive integer,

Therefore,

Generalized inverse[edit]

Some of the properties of inverse matrices are shared bygeneralized inverses(for example, theMoore–Penrose inverse), which can be defined for anym-by-nmatrix.[18]

Applications[edit]

For most practical applications, it isnotnecessary to invert a matrix to solve asystem of linear equations;however, for a unique solution, itisnecessary that the matrix involved be invertible.

Decomposition techniques likeLU decompositionare much faster than inversion, and various fast algorithms for special classes of linear systems have also been developed.

Regression/least squares[edit]

Although an explicit inverse is not necessary to estimate the vector of unknowns, it is the easiest way to estimate their accuracy, found in the diagonal of a matrix inverse (the posterior covariance matrix of the vector of unknowns). However, faster algorithms to compute only the diagonal entries of a matrix inverse are known in many cases.[19]

Matrix inverses in real-time simulations[edit]

Matrix inversion plays a significant role incomputer graphics,particularly in3D graphicsrendering and 3D simulations. Examples include screen-to-worldray casting,world-to-subspace-to-world object transformations, and physical simulations.

Matrix inverses in MIMO wireless communication[edit]

Matrix inversion also plays a significant role in theMIMO(Multiple-Input, Multiple-Output) technology inwireless communications.The MIMO system consists ofNtransmit andMreceive antennas. Unique signals, occupying the samefrequency band,are sent viaNtransmit antennas and are received viaMreceive antennas. The signal arriving at each receive antenna will be alinear combinationof theNtransmitted signals forming anN×Mtransmission matrixH.It is crucial for the matrixHto be invertible for the receiver to be able to figure out the transmitted information.

See also[edit]

References[edit]

  1. ^Axler, Sheldon(18 December 2014).Linear Algebra Done Right.Undergraduate Texts in Mathematics(3rd ed.).Springer Publishing(published 2015). p. 296.ISBN978-3-319-11079-0.
  2. ^J.-S. Roger Jang (March 2001)."Matrix Inverse in Block Form".
  3. ^Weisstein, Eric W."Invertible Matrix Theorem".mathworld.wolfram.com.Retrieved2020-09-08.
  4. ^Horn, Roger A.; Johnson, Charles R. (1985).Matrix Analysis.Cambridge University Press.p. 14.ISBN978-0-521-38632-6..
  5. ^Pan, Victor; Reif, John (1985),Efficient Parallel Solution of Linear Systems,Proceedings of the 17th Annual ACM Symposium on Theory of Computing, Providence:ACM
  6. ^ Pan, Victor; Reif, John (1985),Harvard University Center for Research in Computing Technology Report TR-02-85,Cambridge, MA:Aiken Computation Laboratory
  7. ^A proof can be found in the Appendix B ofKondratyuk, L. A.; Krivoruchenko, M. I. (1992)."Superconducting quark matter in SU(2) color group".Zeitschrift für Physik A.344(1): 99–115.Bibcode:1992ZPhyA.344...99K.doi:10.1007/BF01291027.S2CID120467300.
  8. ^Strang, Gilbert (2003).Introduction to linear algebra(3rd ed.). SIAM. p. 71.ISBN978-0-9614088-9-3.,Chapter 2, page 71
  9. ^Tzon-Tzer, Lu; Sheng-Hua, Shiou (2002). "Inverses of 2 × 2 block matrices".Computers & Mathematics with Applications.43(1–2): 119–129.doi:10.1016/S0898-1221(01)00278-4.
  10. ^ Bernstein, Dennis (2005).Matrix Mathematics.Princeton University Press. p. 44.ISBN978-0-691-11802-4.
  11. ^ Bernstein, Dennis (2005).Matrix Mathematics.Princeton University Press. p. 45.ISBN978-0-691-11802-4.
  12. ^abT. H. Cormen, C. E. Leiserson, R. L. Rivest, C. Stein,Introduction to Algorithms,3rd ed., MIT Press, Cambridge, MA, 2009, §28.2.
  13. ^Ran Raz.On the complexity of matrix product. In Proceedings of the thirty-fourth annual ACM symposium on Theory of computing. ACM Press, 2002.doi:10.1145/509907.509932.
  14. ^ Stewart, Gilbert (1998).Matrix Algorithms: Basic decompositions.SIAM. p. 55.ISBN978-0-89871-414-2.
  15. ^Haramoto, H.; Matsumoto, M. (2009)."A p-adic algorithm for computing the inverse of integer matrices".Journal of Computational and Applied Mathematics.225(1): 320–322.Bibcode:2009JCoAM.225..320H.doi:10.1016/j.cam.2008.07.044.
  16. ^"IML - Integer Matrix Library".cs.uwaterloo.ca.Retrieved14 April2018.
  17. ^Magnus, Jan R.; Neudecker, Heinz (1999).Matrix Differential Calculus: with Applications in Statistics and Econometrics(Revised ed.). New York: John Wiley & Sons. pp. 151–152.ISBN0-471-98633-X.
  18. ^Roman, Stephen(2008),Advanced Linear Algebra,Graduate Texts in Mathematics(Third ed.), Springer, p. 446,ISBN978-0-387-72828-5.
  19. ^Lin, Lin; Lu, Jianfeng; Ying, Lexing; Car, Roberto; E, Weinan (2009)."Fast algorithm for extracting the diagonal of the inverse matrix with application to the electronic structure analysis of metallic systems".Communications in Mathematical Sciences.7(3): 755–777.doi:10.4310/CMS.2009.v7.n3.a12.

Further reading[edit]

External links[edit]