The note here aims to provide shallow references to important aspects of linear maps that relate to the juicy part of linear algebra, matrix. Proofs are mostly omitted in the note as the author was in a hurry and lazy. Like the previous notes, we use V to denote vector space over filed F.1. Linear Map and Its Properties 1.1. Injectivity and surjectivity of linear maps 1.2. Invertibility and Isomorphism of linear maps 2. Matrix, A Representation of Linear Map 2.1. Rank of a matrix 2.2. Notations for matrix manipulations 3. Duality 3.1. Null space and range of the dual of a linear map 4. Things missed out in this note 1. Linear Map and Its PropertiesIn mathematics, mapping builds the correspondence from one set to another set. In linear algebra, the sets are vector spaces. A map is called linear if the following conditions are satisfied:
Definition1A map T from V to W is called linear ifT(v+u)=Tv+Tu for v,uV, andT(𝜆v)=𝜆Tv for 𝜆F and vV
For the simplicity, we use T:VW to denote a linear map from vector space V to W. T here stands for transformation as you will soon see why. Similar to the definition of vector space, the definition of linear map is accompanied by the definition of "zero" or "null":
Definition2The null space of linear map T:VW is defined asnullT={Tv=0:vV}
To give an example of null space, let V=R3, and for arbitrary (x1,x2,x3)V, we have T(x1,x2,x3)=(x1,x2), i.e. T:R3R2. In this case, nullT={(0,0,x3):x3R}. We also have a name for the set of vectors that lead to nonzero vectors upon application of T:
Definition3the range of a linear map T:VW, denoted as rangeT, is a subset of W containing vectors that satisfies the following conditionrangeT={wW:vV,Tv=w}
With these definitions introduced, we are ready to give the theorem that relates T, rangeT, and nullT. Because we use it so frequently in the proofs of linear algebra, we give it a dramatic name as
Theorem1Fundemental theorem of linear maps: Suppose V is finite-dimensional and T is a linear map from V to W, then we havedimV=dimnullT+dimrangeT.
Definition 1 indicates that linear maps are closed under addition and multiplication, so it should not be a surprise that all the linear maps T:VW form a linear space, denoted as L(V,W). Before we discuss special properties of linear maps, we need to clarify two important concepts related to linear maps: injectivity and surjectivity.1.1. Injectivity and surjectivity of linear mapsA linear map is called injective if either one of the following conditions holds:
Theorem2Injectivity of a linear map T:VW means:a. for v,uV, Tv=Tu iff v=u, andb. nullT={0} and dimnullT=0
With Theorem 2, it should not be hard to imagine that a map is not injective when cordiality of V is larger than W. For vector spaces, cordinal numbers are usually infinite. Thus, a formal way of saying this is that a linear map is not injective when dimV>dimW. To see this, we notice that rangeTW in Theorem 2. According to Theorem 1, we have dimnullT=dimV-dimrangeTdimeV-dimW>0, contradicting to Theorem 2b. Thus, there is at least one pair of v and u satisfying Tv=Tu with uv. As for the surjectivity, we have the following theorem holds
Theorem3Surjectivity of a linear map T:VW means rangeT=W.
Theorem 3 indicates that a linear map is not surjective if dimVdimW, as can be proved by using Theorem 1. In other words, there is at least one wW which no vV can map to when dimVdimW. So what is so special about injectivity and surjectivity? How can we solve problems by exploiting the two properties of a linear map? Arguably the easiest application is to check if (in)homogeneous system of linear equations have a solution or not. A homogeneous system of m linear equations can be written in the following form:a1,1x1+a1,2x2+a1,nxn=0a2,1x1+a2,2x2+a2,nxn=0am,1x1+am,2x2+am,nxn=0anchor-tagIf we let V=Fn, then we can define a linear map T:FnFm as the followingT(x1,x2,...,xn)=(j=1na1,jxj,j=1na2,jxj,...,j=1nam,jxj)anchor-tagSo having a solution to the system above is equavalent to having nonzero element in nullT, i.e., dimnullT>0. Acoording to Theorem 1, such condition can only be satisfied when dimFn>dimrangeT. Because dimrangeTdimFm, we must have solutions to the equation system when n>m. Thus, a homogeneous system of linear equations has solutions when its corresponding map T:FnFm is not injective.On the other hand, for inhomogeneous system of linear equations, having a solution means there is at least one vector v=(x1,x2,...,xn)Fn satisfying the condition of Tv=(c1,c2,...,cm), where some ci0. Given the linear map defined in Eqn.(2), we claim that solution for the inhomogeneous sytem does not exist for some choices of ci when m>n as the linear map T is not surjective.Other than this simple application, surjectivity and injectivity also dictate two properties of linear maps, invertibility and isomorphism, as we shall now see.1.2. Invertibility and Isomorphism of linear mapsWe first define the invertibility of a linear map as the following
Definition4A linear map TL(V,W) is invertible if there exists another linear map SL(W,V) that make S(Tv)=Iv=v hold for all vV, where I is identity map.
As a matter of fact, the inverted map S in Definition 4 is usually written as T-1 (i.e. inverse of T). If a linear map has an inverse, we have
Proposition1Inverse of a linear map is unique.
as can be shown by assuming two inverses, S1 and S2, of a linear map T:S1=S1I=S1TS2=IS2=S2According to Definition 4, STv=v for all vV, which indicates
Proposition2Invertibility of a linear map indicates injectivity and surjectivity.
Being invertible for a linear map is very important to quantum-mechanical calculations, as an invertible map, if it is unitary, is equivalent to a reversible operator whose physical realization is a logic gate on quantum circuit. Unlike quantum computer, classical logic gates, such as AND, OR, XOR are irreversible as numbers of their inputs are larger than numbers of their outputs. From the perspective of information theory, being reversible means that we can recover input information by simply knowing the outputs, which can be achieved by gates that follow invertible linear map from inputs to outputs. To be streamlined with possible later discussion of quantum operators(gates). Here we give the definition of operator in linear algebra:
Definition5An operator is a linear map from a vector space onto itself, i,e, T:VV, and the set of operators on vector space V forms a vector space as well, denoted as L(V).
The concept of isomorphism is a natural extension of inversibility of vector space, as we shall introduce now
Definition6An isomorphism is a invertible linear map, and thus, two vector spaces are called isomorphic if an invertible linear map exists from one to another.
Isomorphism gives the equivalence of two vector spaces. This is very powerful as we can now make a problem, being hard in one vector space, much easier by simply reformulating it onto another vector space, if we know the two vector spaces are isomorphic. In linear algebra, arguably the most important isomorphism is between the vector space of linear map L(V,W), and the vector space of matrices, Fm,n. We will come back to the proof of this proposition once we introduce the concept of matrix.2. Matrix, A Representation of Linear MapThe title of this section indicates on intention of explaining representation theory used in quantum theories, even though the author would much love to. Instead, we will introduce the form of matrix, and the interpretation of it. However, we will not focus on standard contents such as matrix addition, multiplication, etc., for which readers can refer to introductory textbooks or Wikipedia pages. The primary purpose of this section is to reveal how one can understand matrix in terms of linear map, and mechanistic calculation of matrices is never the fun part.To begin with, suppose we have a vector space V with its basis being v1,v2,...,vn, and another vector space W with its basis of w1,w2,...,wm. Therefore, dimV=n and dimW=m. Now we define a linear map T:VW asTvi=a1,iw1+a2,iw2++am,iwmfor i=1,2,...,nanchor-tagwhere ai,jF. Then, the linear map can be represented by a m-row, n-column matrix as shown belowv1vivnw1w2wm(a1,1a1,ia1,na2,1a2,ia2,nam,1am,iam,n)anchor-tagBy multiplying elements in each column with its associated wj outside of the matrix, we recover the pesentation of Tvi in W. To see how an arbitrary vector vV is mapped to W using the matrix in (4), let us first represent v as a conlumn vector, i.e.,M(v)=(c1c2cn)anchor-tagwhich indicates v=c1v1+c2v2++cnvn and ciF for i=1,2,...,n. M here maps vV to column vector, and it is also a linear map. We will discuss more about this later. Given Eqn.(3), we can calculate Tv asTv=c1Tv1+c2Tv2++cnTvn=c1j=1maj,1wj+c2j=1maj,2wj++cnj=1maj,nwj=(i=1ncia1,i)w1+(i=1ncia2,i)w2++(i=1nciam,i)wmanchor-tagIf we use the rule of matrix multiplication to multiply matrix in (4) with the vector in (5), we obtain a m-row column vector beingw=((i=1ncia1,i)(i=1ncia2,i)(i=1nciam,i))anchor-tagComparing (7) with (6), it should be obvious that the resulted vector w is indeed the image of v in W. In other words, v is mapped to w through the linear map T represented as the matrix in (4). To be formal, we can define the matrix of a linear map as the following
Definition7Suppose TL(V,W), and v1,v2,...,vn is a basis of V, and w1,...,wm is a basis of W. The matrix of T with respect to these bases is a m-by-n matrix M(T) whose entries ai,j are defined by Eqn.(3).
Definition 7 emphasizes the use of bases, implying that matrix is always related to the change of basis. If we treat M as a map that maps T to a matrix, then it's easy to show that M itself is a linear map as it is closed under addition and multiplication. Also, For the simplicity in later discussion, we use Fm,n to denote a m-by-n matrix.So far, we have taken for granted that linear map can always be represented by a matrix. But the validity of such a claim need to be examined by proving the following proposition:
Proposition3Suppose v1,v2,...,vn is a basis of V and w1,...,wm a basis of W, then M is an isomorphism between L(V,W) and Fm,n.
Proof:M is a linear map as for any T,SL(V,W), M(T+S)=M(T)+M(S), M(𝜆T)=𝜆M(T) for 𝜆F. So we only need to prove it is injective and surjective because of Proposition 2. According to Theorem 2, if M(T)=0 then Tvi=0 for any of the basis vector of V. This is only possible when T=0, i.e., nullM={0}. Thus, M is injective. To prove M is surjective, suppose AFm,n is an arbitrary matrix. we can always define a linear map T from V to W asTvi=j=1mAj,iwjfori=1,2,...,n.from which M(T)=A holds. In other words, rangeM=Fm,n, and M is surjective.2.1. Rank of a matrixSo far we only considered matrices that have each row or each column associated with a basis vector in vector spaces. What if we add to those matrices more rows or columns? Will the expanded matrices still be isomorphic to some linear maps? This is a weird question, but there is no limitation for expanding the size of matrices at all. Let's first add some rows to matrix A isomorphic to a linear map TL(V,W), and adding columns follows the same logic. Suppose v1,v2,v3 is a basis of V and w1,w2,w3 a basis of W. If we define the map asTvi=A1,iw1+A2,iw2+A3,iw3then we can write down A as a 3-by-3 matrix. Now, if we rewrite the definition above asTvi=A1,iw1+A2,iw2+A3,iw3+A4,i(2w2+w3)with A4,i0then it is obvious that A1,i=A1,i, A2,i=A2,i+2A4,i, A3,i=A3,i+A4,i. The two definitions will map arbitrary vector v=c1v1+c2v2+c3v3V to the same vector wW, but A is NOT isomorphic to T. To see this, we simply note that we can give different values to A2,i, A3,i, and A4,i and still make the relations A2,i=A2,i+2A4,i and A3,i=A3,i+A4,i hold. In other words, TL(V,W) corresponds to multiple AF4,3, resulting in the map not injective or surjective. There are matrices that have rows or columns linearly dependent. In those cases, we define the row rank and the column rank of the matrix as the following
Definition8Suppose AFm,n, then the row(column) rank of A is the dimension of the span of the rows(columns) of A.
The rank numbers of a matrix can be related to a linear map through the following proposition
Proposition4Dimension of rangeT equals column rank of M(T).
and
Proposition5Row rank of a matrix equals to column rank.
Thus, the rank of a matrix AFm,n is the column rank of A.2.2. Notations for matrix manipulationsWe end this section by introducing algebraic notations for matrix manipulations that will be used in the following sections. By following the notation in Axler's book, the transpose of a matrix is defined as the following
Definition9Transpose of a matrix A, denoted as At, has its element satisfying the relation below:Ai,jt=Aj,i
For people who have encountered linear algebra before, the following rule should not be alien:
Proposition6For two matrices A and B, we have (AB)t=BtAt.
Proof: To see this, suppose A is a m-by-n matrix and B is n-by-q. So the element in AB can be represented as(AB)i,j=knAi,kBk,jwhere Ai,k and Bk,j are elements in A and B, respectively. Definition 9 gives (AB)i,jt=(AB)j,i=knAj,kBk,i=knAk,jtBi,kt=(BtAt)i,jBecause of Proposition 3, we have invertibility of matrices well defined. Let A-1 to be the inverse of a m-by-n A, then A-1A=I, with I being n-by-n identity matrix. We can also use row and column indices to represent the ith row of a matrix A as Ai,, and the jth column as A,j. With this, the following identities hold:(AB)k,=Ak,B(AB),r=AB,r3. DualityFor quantum computing and quantum information theories, linear maps that map vectors to scaler field F play critical roles, as those maps are used to collapse wavefunctions into states with real physical properties. Because it is so special, we call linear map TL(V,F) linear functional. For an arbitrary vector space V, we call L(V,F) its dual space, V. With this, we can find that
Proposition7dimV=dimV
Proof: To see this, we first note that the isomorphism between L(V,W) and Fm,n results in dimL(V,W)=dimFm,n. Getting dimFm,n is easy. We can think of its basis vector as m-by-n matrices with only one nonzero element at ith row and jthe column. Thus, we have dimFm,n=m×n=dimL(V,W), and dimL(V,F)=n=dimV.If v1,...,vn is a basis of V, then we can define the dual basis of v1,...,vn as
Definition10The dual basis of v1,...,vnV is the list 𝜙1,...,𝜙n of elements of dual space V, where each 𝜙j is a linear functional on V such that𝜙j(vk)={1ifk=j0 if kj
With this definition, we can proof the following
Proposition8Dual basis is a basis of the dual space
Proof: we just need to prove that 𝜙1,...,𝜙n is linearly independent because every linearly independent list of lenth dimV is the basis of V. Suppose a1,...,anF gives thata1𝜙1++an𝜙n=0.So (a1𝜙1++an𝜙n)(vj)=aj for j=1,...,n, and the equation above implies that a1=a2==an=0. Hence the list 𝜙1,...,𝜙n is linearly independent We can take a step further to define the dual map,
Definition11If TL(V,W), then the dual map of T is the linear map TL(W,V) such that T(𝜙)=𝜙T for 𝜙W.
𝜙T means the composite application of two mapps. For arbitrary vector vV, (𝜙T) first maps it to a vector TvW, and then apply 𝜙 to Tv to map it to a scalar. The algebraic properties of dual maps listed below might be useful(S+T)=S+Tfor allS,TL(V,W)(𝜆T)=𝜆Tfor all𝜆F(ST)=TSwhereTL(U,V),SL(V,W)3.1. Null space and range of the dual of a linear mapWe can describe rangeT and nullT in terms of rangeT and nullT by first defining the annihilator of subspaces.
Definition12For UV, the annihilator of U, denoted as U0, is defined byU0={𝜙V:𝜙(u)=0uU}
The dimension of the annihilator is related to dimV through the folowing eqaution
Proposition9Suppose V if finite-dimensional and U is a subspace of V, then dimU+dimU0=dimV.
Suppose V and W are finite-dimensional and TL(V,W), then the following results hold. The author lists them here while have no clue how they are used in quantum-mechanical calculations.
nullT=(rangeT)0dimnullT=dimnullT+dimW-dimVdimrangeT=dimrangeTrangeT=(nullT)0
4. Things missed out in this noteLinear map is a rich branch of linear algebra. There are many interesting things left out here, among which products and quotients of space vectors are definitely worth diving. While quotients may not be pervasive in quantum computing literature, it seems to be a handy tool in proofs of quantum information theories.