Skip to content or footer

Course

Why does row rank equal column rank?

6Understanding the relation between row rank and column rank

We have seen that the kernel of a matrix is a measure for the uniqueness of solutions. But how to compute this kernel? A vector xRnx \in \R^n is in the kernel if and only if Ax=0Ax = 0. Explicitly this is the case if

for all i{1,,m}i \in \{1, \dots, m\}. This is the standard scalar product of the vector xx with the ii-th row vector

We thus have the description that the kernel of A is given by all vectors in Rn\R^n which are orthogonal to the vectors r1,,rmr_1, \dots, r_m. Formally

where y,x:=yTx\langle y, x\rangle:=y^Tx is the standard scalar product of x,yRnx, y \in \R^n.

Consider the linear system of equations

that we encountered before. It is represented by the matrix

with kernel

If we draw the kernel and the span of the rows

we see that they are indeed orthogonal to each other:

The span of the rows is orthogonal to the kernel.

If we have yim(A)y \in \operatorname{im}(A), by definition we find an xR2x \in \R^2 such that Ax=yAx = y. In the last section we defined the kernel and found that the difference of solutions is contained in the kernel. On the other hand, we can add any element from ker(A)\ker(A) to xx and obtain a new solution xx’ with Ax=yAx’ = y: If we have Ax=yAx = y and zker(A)z \in \ker(A), then A(x+z)=Ax+Az=y+0=yA(x+z) = Ax + Az = y + 0 = y. In the picture this corresponds to moving the point x parallel to the kernel.

The element x is transported parallel to ther kernel into the span of the rows.

If we pick the right element in the kernel, we can move xx into the span of the rows RR and get xRx’ \in R with Ax=yAx’ = y. We can do this with any element in the image of AA. Hence we find for every element bim(A)b \in \operatorname{im}( A), an element xRx \in R with Ax=yAx = y.

The picture suggests even more: Whenever we have another solution x’’Rx’’\in R, then the difference xx’’x’-x’’ has to lie in the kernel of AA. Since xx’’x’-x’’ is also in RR, it has to lie in the intersection of the kernel and RR. By the picture it has to be zero, which implies x=x’’x’ = x’’. If we look closely, we find a familiar reason for this: The kernel is orthogonal to the span of the rows.

Formally we can express this observation as follows. We know that xx’’x’ -x’’ is in RR, hence we can write it as a linear combination xx’’=λ1r1++λmrmx’ -x’’ = \lambda_1 r_1 + \dots + \lambda_m r_m of the rows r1,,rmr_1,\ldots,r_m of AA. Next we can exploit that elements in the kernel are orthogonal to the rows, by computing the scalar product of xx’’ker(A)x’-x’’ \in \ker(A) with xx’’x’-x’’ as the linear combination of the rows. That is

Hence we get xx’’=0x’-x’’ = 0 and thus x=x’’x’ = x’’.

All together, we showed that given some yy in the image of AA, i.e. some linear combination of the columns, there exists a unique solution xx with Ax=yAx = y, such that xx is a linear combination of the rows. So, the matrix AA yields a 1-1-correspondence between the space of linear combinations of the rows RR and the span of the columns im(A)\operatorname{im}( A). Formally, the restriction of AA to RR

gives a bijection (a 1-1 map or correspondence) between RR and im(A)\operatorname{im}(A). Furthermore, the correspondence preserves the structure of RR and im(A)\operatorname{im}(A). This means that we can transport all calculations and relations between elements from RR to im(A)\operatorname{im}(A) and back using the above restriction of AA to RR (and its inverse). Hence both spaces have the same properties. In particular RR and im(A)\operatorname{im}(A), i.e. the span of the rows and the span of the columns, have the same dimension. The dimension of RR is the row rank and the dimension of im(A)\operatorname{im}(A) is the column rank. Hence the row rank equals the column rank.


This content is licensed under
CC BY-SA 4.0Info