Matrix

The matrix of a linear map M(T) is defined with respect to a set of basis for V and W. The entries of M(T) are defined as Tvk=A1,kw1+A2,kw2++Am,kwmk=1,,n where v1,,vn and w1,,wm are the basis vectors of V and W respectively. M(T) is a m-by-n matrix and each element F. Unless stated otherwise, we will use the standard basis. It is important to note that the matrix of a linear map is always defined with respect to a set of basis vectors.

Addition and Multiplication

Addition of matrices of the same size is defined as matrix whose each element is the sum of the corresponding elements of the two matrices. (A+B)i,j=Ai,j+Bi,j

Thus, the matrix of sum of linear maps is the sum of matrices of those linear maps M(S+T)=M(S)+M(T)S,TL(V,W)

The scalar multiplication with λF is same as multiplying each element of the original matrix with λ (λA)i,j=λ(Ai,j)

Thus, the matrix of a scalar times a linear map is same as the scalar times the matrix of the linear map M(λT)=λM(T)TL(V,W)

We will denote the set of all mbyn matrices with elements in F by Fm,n. This set of all matrices is also a vector space on the addition and scalar multiplication rules for matrices defined above. The basis for such space is the collection of all matrices with all but one element zeros, and one element 1. There are mn such matrices meaning dim Fm,n=mn.

Matrix Multiplication

Matrix multiplication is important to define in order to work with product of linear maps. We wish to have the following hold M(ST)=M(S)M(T)TL(U,V),SL(V,W) For two matrices A and B of sizes m-by-n and n-by-p respectively, their matrix multiplication is defined as (AB)i,j=k=1pAi,kBk,j Which is the sum product of the ith row of A and jth column of B. Notice that the matrices have one dimension same. It is a necessary condition for the multiplication to be valid and also implies that the range of T is same as domain of S. Furthermore, AB is of size m-by-p.

Matrix multiplication is not commutative (since the multiplication may not be defined or the products may not be the same otherwise), but is distributive and associative.

Inverse Linear Map

A linear map TL(V,W) is invertible if there exists TL(W,V) such that ST is the identity map on V and TS is the identity map on W. S is said to be the inverse of T with ST=I and TS=I, and there exists a unique inverse for a linear map (can be proven by contradiction).

We denote the unique inverse of a linear map TL(V,W) with T1L(W,V) and they satisfy TT1=I and T1T=I.

A linear map is invertible if and only if it is both injective and surjective. To prove this, we first assume the linear map to be invertible and show that the latter part of statement is true and vice versa.

Two vector spaces are isomorphic if there is an isomorphism (an invertible linear map) from one vector space onto the other.

Operator

A linear map from a vector space onto itself is called an operator. L(V) denotes the set of all operators from V onto itself. It is the same notation as L(V)=L(V,V).

On a finite dimensional vector space V, any linear map TL(V) is \textbf{invertible}, \textbf{injective} and \textbf{surjective}.

Rank of a Matrix

Suppose AFm,n. Row and column ranks are two non negative integers defined as

  • Row Rank: Dimension of the span of the rows of A in F1,n
  • Column Rank: Dimension of the spaan of the columns of A in Fm,1

But for any AFm,n, row rank = column rank (can be proved using duality). Hence, we simply use the word rank to denote the column rank of a matrix.

Rank is defined for a matrix, whereas dimension is defined for a vector space. This is the important distinction between the two. The rank can be found by counting the number of non-zero rows in the row echelon form of a matrix (see here).

Rank-Nullity Theorem

We restate the theorem from fundamental theorem of linear maps using the concepts of dimension and ranks. For any vector spaces V and W, and linear map TL(V,W) rank(T)+Nullity(T)=dim(V) where dim(V) is the number of basis vectors of the vector space V. Usually, the dim(V) is the number of columns in the matrix T. For a matrix, the dimension of the null space will be determined by the number of basis vectors of the solution set to Ax=0.

Block Matrix

Sometimes, block matrices can help simplify cumbersome operations. For the purpose of demonstration, consider a matrix A of dimensions n×n, which we break into four blocks: B of size m×m, C of size m×nm, D of size nm×m, and E of size nm×nm. Then, A=[BCDE] Then, the calculation of A2 is A2=[BCDE][BCDE]=[B2+CDBC+CEDB+EDDC+E2]

We obtained B2 using 2×2 multiplication on the four elements of the block matrix. All the multiplications agree in terms of the dimensions. This technique is especially useful when the matrix can be broken in such a way that individual matrix are identity matrix, diagonal matrix etc.

Orthogonal Matrix

Orthogonal or \textbf{orthonormal} matrix is a matrix whose rows and columns are orthonormal vectors. An orthogonal matrix Q will satisfy QQT=QTQ=IQ1=QTdet(Q)=±1

The last one follows from the fact that 1=det(I)=det(QQT)=det(Q)2

Orthogonal matrix play an important role in QR decomposition and SVD.

Determinant

Determinant is defined for square matrices and represents a transformation from Rn×nR. Its defined as follows for 2×2 and 3×3 matrices A=[abcd]det(A)=acbdA=[abcdefghi]det(A)=a(eihf)b(digf)+c(dhge)

The term outside the bracket is the entry from the first row, and the entry inside the brackets is the determinant of the n1×n1 matrix formed by not considering the first row and the column corresponding to the entry outside the bracket. We put alternating + and signs before each term. This way the definition extends to any n×n matrix.

If the determinant of a matrix is non zero, then the matrix has full rank, or all its rows and columns are linearly independent, and the matrix is invertible.