Processing math: 100%
Skip to main content
Library homepage
 

Text Color

Text Size

 

Margin Size

 

Font Type

Enable Dyslexic Font
Physics LibreTexts

19.2: Appendix - Matrix Algebra

( \newcommand{\kernel}{\mathrm{null}\,}\)

Matrices

Matrix algebra provides an elegant and powerful representation of multivariate operators, and coordinate transformations that feature prominently in classical mechanics. For example they play a pivotal role in finding the eigenvalues and eigenfunctions for coupled equations that occur in rigid-body rotation, and coupled oscillator systems. An understanding of the role of matrix mechanics in classical mechanics facilitates understanding of the equally important role played by matrix mechanics in quantal physics.

It is interesting that although determinants were used by physicists in the late 19th century, and the concept of matrix algebra was developed by Arthur Cayley in England in 1855, many of these ideas were the work of Hamilton, and the discussion of matrix algebra was buried in a more general discussion of determinants. Matrix algebra was an esoteric branch of mathematics, little known by the physics community, until 1925 when Heisenberg proposed his innovative new quantum theory. The striking feature of this new theory was its representation of physical quantities by sets of time-dependent complex numbers and a peculiar multiplication rule. Max Born recognized that Heisenberg’s multiplication rule is just the standard “row times column” multiplication rule of matrix algebra; a topic that he had encountered as a young student in a mathematics course. In 1924 Richard Courant had just completed the first volume of the new text Methods of Mathematical Physics during which Pascual Jordan had served as his young assistant working on matrix manipulation. Fortuitously, Jordan and Born happened to share a carriage on a train to Hanover during which Jordan overheard Born talk about his problems trying to work with matrices. Jordan introduced himself to Born and offered to help. This led to publication, in September 1925, of the famous Born-Jordan paper[Bor25a] that gave the first rigorous formulation of matrix mechanics in physics. This was followed in November by the Born-Heisenberg-Jordan sequel[Bor25b] that established a logical consistent general method for solving matrix mechanics problems plus a connection between the mathematics of matrix mechanics and linear algebra. Matrix algebra developed into an important tool in mathematics and physics during World War 2 and now it is an integral part of undergraduate linear algebra courses.

Most applications of matrix algebra in this book are restricted to real, symmetric, square matrices. The size of a matrix is defined by the rank, which equals the row rank and column rank, i.e. the number of independent row vectors or column vectors in the square matrix. It is presumed that you have studied matrices in a linear algebra course. Thus the goal of this review is to list simple manipulation of symmetric matrices and matrix diagonalization that will be used in this course. You are referred to a linear algebra textbook if you need further details.

Matrix definition

A matrix is a rectangular array of numbers with M rows and N columns. The notation used for an element of a matrix is Aij where i designates the row and j designates the column of this matrix element in the matrix A. Convention denotes a matrix A as

A(A11A12A1(N1)A1NA21A22..A2(N1)A2N::Aij::A(M1)1A(M1)2..A(M1)(N1)A(M1)NAM1AM2AM(N1)AMN)

Matrices can be square, M=N, or rectangular MN. Matrices having only one row or column are called row or column vectors respectively, and need only a single subscript label. For example,

A=(A1A2:AM1AM)

Matrix manipulation

Matrices are defined to obey certain rules for matrix manipulation as given below.

1) Multiplication of a matrix by a scalar λ simply multiplies each matrix element by λ.

Cij=λAij

2) Addition of two matrices A and B having the same rank, i.e. the number of columns, is given by

Cij=Aij+Bij

3) Multiplication of a matrix A by a matrix B is defined only if the number of columns in A equals the number of rows in B. The product matrix C is given by the matrix product

C=AB

Cij=[AB]ij=kAikBkj

For example, if both A and B are rank three symmetric matrices then

C=AB=(A11A12A13A21A22A23A31A32A33)(B11B12B13B21B22B23B31B32B33)=(A11B11+A12B21+A13B31A11B12+A12B22+A13B32A11B13+A12B23+A13B33A21B11+A22B21+A23B31A21B12+A22B22+A23B32A21B13+A22B23+A23B33A31B11+A32B21+A33B31A31B12+A32B22+A33B32A31B13+A32B23+A33B33)

In general, multiplication of matrices A and B is noncommutative, i.e.

ABBA

In the special case when AB=BA then the matrices are said to commute.

Transposed matrix AT

The transpose of a matrix A will be denoted by AT and is given by interchanging rows and columns, that is

(AT)ij=Aji

The transpose of a column vector is a row vector. Note that older texts use the symbol ˜A for the transpose.

Identity (unity) matrix I

The identity (unity) matrix I is diagonal with diagonal elements equal to 1, that is

Iij=δij

where the Kronecker delta symbol is defined by

δik=0 if ik=1 if i=k

Inverse matrix A1

If a matrix is non-singular, that is, its determinant is non-zero, then it is possible to define an inverse matrix A1. A square matrix has an inverse matrix for which the product

AA1=I

Orthogonal matrix

A matrix with real elements is orthogonal if

AT=A1

That is

k(AT)ikAkj=kAkiAkj=δij

Adjoint matrix A

For a matrix with complex elements, the adjoint matrix, denoted by A is defined as the transpose of the complex conjugate

(A)ij=Aji

Hermitian matrix

The Hermitian conjugate of a complex matrix H is denoted as H and is defined as

H=(HT)=(H)T

Therefore

Hij=Hji

A matrix is Hermitian if it is equal to its adjoint

H=H

that is

Hij=Hji=Hij

A matrix that is both Hermitian and has real elements is a symmetric matrix since complex conjugation has no effect.

Unitary matrix

A matrix with complex elements is unitary if its inverse is equal to the adjoint matrix

U=U1

which is equivalent to

UU=I

A unitary matrix with real elements is an orthogonal matrix as given in Equation ???.

Trace of a square matrix TrA

The trace of a square matrix, denoted by TrA, is defined as the sum of the diagonal matrix elements.

TrA=Ni=1Aii

Inner product of column vectors

Real vectors

The generalization of the scalar (dot) product in Euclidean space is called the inner product. Exploiting the rules of matrix multiplication requires taking the transpose of the first column vector to form a row vector which then is multiplied by the second column vector using the conventional rules for matrix multiplication. That is, for rank N vectors

[X][Y]=(X1X2:XN)(Y1Y2:YN)=[X]T[Y]=(X1X2..XN)(Y1Y2:YN)=Ni=1XiYi

For rank N=3 this inner product agrees with the conventional definition of the scalar product and gives a result that is a scalar. For the special case when [A][B]=0 then the two matrices are called orthogonal. The magnitude squared of a column vector is given by the inner product

[X][X]=Ni=1(Xi)20

Note that this is only positive.

Complex vectors

For vectors having complex matrix elements the inner product is generalized to a form that is consistent with Equation ??? when the column vector matrix elements are real.

[X][Y]=[X][Y]=(X1X2..XN1XN)(Y1Y2:YN1YN)=Ni=1XiYi

For the special case

[X][X]=[X][X]=Ni=1XiXi0

Determinants

Definition

The determinant of a square matrix with N rows equals a single number derived using the matrix elements of the matrix. The determinant is denoted as detA or |A| where

|A|=Nj=1ε(j1,j2,.jN)A1j1A2j2ANjN

where ε(j1,j2,.jN) is the permutation index which is either even or odd depending on the number of permutations required to go from the normal order (1,2,3,N) to the sequence (j1j2j3jN).

For example for N=3 the determinant is

|A|=A11A22A33+A12A23A31+A13A21A32A13A22A31A11A23A32A12A21A33

Properties

  1. The value of a determinant |A|=0, if
    1. all elements of a row (column) are zero.
    2. all elements of a row (column) are identical with, or multiples of, the corresponding elements of another row (column).
  2. The value of a determinant is unchanged if
    1. rows and columns are interchanged.
    2. a linear combination of any number of rows is added to any one row.
  3. The value of a determinant changes sign if two rows, or any two columns, are interchanged.
  4. Transposing a square matrix does not change its determinant. |AT|=|A|
  5. If any row (column) is multiplied by a constant factor then the value of the determinant is multiplied by the same factor.
  6. The determinant of a diagonal matrix equals the product of the diagonal matrix elements. That is, when Aij=λiδij then |A|=λ1λ2λ3λN
  7. The determinant of the identity (unity) matrix |I|=1.
  8. The determinant of the null matrix, for which all matrix elements are zero, |0|=0
  9. A singular matrix has a determinant equal to zero.
  10. If each element of any row (column) appears as the sum (difference) of two or more quantities, then the determinant can be written as a sum (difference) of two or more determinants of the same order. For example for order N=2, |A11±B11A12±B12A21A22|=|A11A12A21A22|±|B11B12A21A22|
  11. A determinant of a matrix product equals the product of the determinants. That is, if C=AB then |C|=|A||B|

Cofactor of a square matrix

For a square matrix having N rows the cofactor is obtained by removing the ith row and the jth column and then collapsing the remaining matrix elements into a square matrix with N1 rows while preserving the order of the matrix elements. This is called the complementary minor which is denoted as A(ij). The matrix elements of the cofactor square matrix a are obtained by multiplying the determinant of the (ij) complementary minor by the phase factor (1)i+j. That is

aij=(1)i+j|A(ij)|

The cofactor matrix has the property that

Nk=1Aikajk=δij|A|=Nk=1Akiakj

Cofactors are used to expand the determinant of a square matrix in order to evaluate the determinant.

Inverse of a non-singular matrix

The (i,j) matrix elements of the inverse matrix A1 of a non-singular matrix A are given by the ratio of the cofactor aji and the determinant |A|, that is

A1ij=1|A|aji

Equations ??? and ??? can be used to evaluate the i,j element of the matrix product (A1A)

(A1A)ij=Nk=1A1ikAkj=1|A|Nk=1ajiAkj=1|A|δji|A|=δij=Iij

This agrees with Equation ??? that AA1=I.

The inverse of rank 2 or 3 matrices is required frequently when determining the eigen-solutions for rigidbody rotation, or coupled oscillator, problems in classical mechanics as described in chapters 11 and 12. Therefore it is convenient to list explicitly the inverse matrices for both rank 2 and rank 3 matrices.

Inverse for rank 2 matrices:

A1=[abcd]1=1|A|[dbca]=1(adbc)[dbca]

where the determinant of A is written explicitly in Equation ???.

Inverse for rank 3 matrices:

A1=[abcdefghi]1=1|A|[ABCDEFGHI]T=1|A|[ADGBEHCFI]=1aA+bB+cC[A=(eifh)D=(bich)G=(bfce)B=(difg)E=(aicg)H=(afcd)C=(dheg)F=(ahbg)I=(aebd)]

where the functions A,B,C,D,E,F,G,H,I, are equal to rank 2 determinants listed in Equation ???.

Reduction of a matrix to diagonal form

Solving coupled linear equations can be reduced to diagonalization of a matrix. Consider the matrix A operating on the vector X to produce a vector Y, that are expressed as components with respect to the unprimed coordinate frame, i.e.

AX=Y

Consider that the unitary real matrix R with rank n, rotates the n-dimensional un-primed coordinate frame into the primed coordinate frame such that A, X and Y are transformed to A, X and Y in the rotated primed coordinate frame. Then

X=RXY=RY

With respect to the primed coordinate frame Equation ??? becomes

R(AX)=RY

RAR1RX=RY

RAR1X=AX=Y

using the fact that the identity matrix I=RR1=RRT since the rotation matrix in n dimensions is orthogonal.

Thus we have that the rotated matrix

A=RART

Let us assume that this transformed matrix is diagonal, then it can be written as the product of the unit matrix I and a vector of scalar numbers called the characteristic roots λ as

A=RART=λI

using the fact that RT=R1 then gives

RT(λI)=ART

Let both sides of Equation ??? act on X which gives

λIX=AX

or

[λIA]X=0

This represents a set of n homogeneous linear algebraic equations in n unknowns X where λ is a set of characteristic roots, (eigenvalues) with corresponding eigenfunctions X. Ignoring the trivial case of X being zero, then ??? requires that the secular determinant of the bracket be zero, that is

|λIA|=0

The determinant can be expanded and factored into the form

(λλ1)(λλ2)(λλ3).(λλn)=0

where the n eigenvalues are λ=λ1,λ2,λn of the matrix A.

The eigenvectors X corresponding to each eigenvalue are determined by substituting a given eigenvalue λi into the relation

XTAX=[λiδij]

If all the eigenvalues are distinct, i.e. different, then this set of n equations completely determines the ratio of the components of each eigenvector along the axes of the coordinate frame. However, when two or more eigenvalues are identical, then the reduction to a true diagonal form is not possible and one has the freedom to select an appropriate eigenvector that is orthogonal to the remaining axes.

In summary, the matrix can only be fully diagonalized if

(a) all the eigenvalues are distinct,

(b) the real matrix is symmetric,

(c) it is unitary.

A frequent application of matrices in classical mechanics is for solving a system of homogeneous linear equations of the form

A11x1+A12x2+A1nxn=0A11x1+A12x2+A1nxn=0......=.An1x1+An2x2..+Annxn=0

Making the following definitions

A=(A11A12A1nA21A22A2nAn1An2Ann)

X=(x1x2xn)

Then the set of linear equations can be written in a compact form using the matrices

AX=0

which can be solved using Equation ???. Ensure that you are able to diagonalize a matrices with rank 2 and 3. You can use Mathematica, Maple, MatLab, or other such mathematical computer programs to diagonalize larger matrices.

Example 19.2.1: Eigenvalues and eigenvectors of a real symmetric matrix

Consider the matrix

A=(010100000)

The secular determinant is given by ???

|λ101λ000λ|=0

This expands to

λ(λ+1)(λ1)=0

Thus the three eigen values are λ=1,0,1.

To find each eigenvectors we substitute the corresponding eigenvalue into Equation ???.

(λ101λ000λ)(xyz)=(000)

The eigenvalue λ=1 yields x+y=0 and z=0. Thus the eigen vector is r1=(12,12,0). The eigenvalue λ=0 yields x=0 and y=0. Thus the eigen vector is r2=(0,0,1). The eigenvalue λ=1 yields x+y=0 and z=0. Thus the eigen vector is r3=(12,12,0). The orthogonality of these three eigen vectors, which correspond to three distinct eigenvalues, can be verified.


This page titled 19.2: Appendix - Matrix Algebra is shared under a CC BY-NC-SA 4.0 license and was authored, remixed, and/or curated by Douglas Cline via source content that was edited to the style and standards of the LibreTexts platform.

Support Center

How can we help?