Appendix D
Exercise D.1-1
For all and we have
so and are also symmetric matrices.
Exercise D.1-2
Let be a matrix of size and a matrix of size .We have
Also, , so is always a symmetric matrix.
Exercise D.1-3
Let and be lower-triangular matrices and . For all we have
Therefore, the matrix is also lower-triangular.
Exercise D.1-4
As explained in the book, multiplying a vector by a permutation matrix has the effect of permuting (rearranging) the elements of . We can represent a matrix as a sequence of column vectors . Thus, is a matrix where each column vector of is permuted in the same way. In other words, rows of are permuted. Analogously, we can view a matrix as a sequence of row vectors , hence is a matrix where each row vector of is permuted in the same way. In other words, columns of are permuted.
The previous explanation proves the statement from the book at high level. Let us see where elements of land in . Row of contains only a single 1 at index , so and 0s elsewhere. We have
Notice that all elements from row of goes into row . In other words, rows of are permuted. This works since each distinct row of has a unique column index such that has 1 there. A similar observation applies to just that instead of permuting rows we permute columns.
Apparently, if we replace with , then is another permutation matrix, as permuting columns or rows of still maintains the property that it will have exactly one 1 in each row or column, and 0s elsewhere.
Exercise D.2-1
Suppose, for contradiction, that . We have
which contradicts our initial assumption of distinct inverses. Therefore, matrix inverses are unique.
Exercise D.2-2
Let be an lower-triangular matrix. We prove by induction on that is equal to the product of its diagonal elements. The base case holds; when then . For we have
According to the principles of mathematical induction, we can conclude that the statement holds for all . For an upper-triangular matrix the poof is nearly identical with a difference that we start at and work backwards.
Suppose, for contradiction, that is not lower-triangular, thus ith row of has additional nonzero entries for . We know that and the ith row of has only one nonzero element at column . Since is lower-triangular, the first nonzero element in the jth column is at row , hence ; , where , and for all . But this entails a cascading effect and so on. It turns out that for all , which contradicts our statement that is not lower-triangular. Observe that all diagonal elements of must be nonzero, otherwise would be singular.
Exercise D.2-3
Wikipedia has a proof here that is invertible and .
maintains the property that it has exactly one 1 in each row or column, and 0s elsewhere, since transposing a matrix swaps its columns and rows. Therefore, is also a permutation matrix.
Exercise D.2-4
We know that , where , and ; is a row vector of and is a column vector of . All row vectors of , except row , are the same in . All column vectors of , except column , are the same in . It is stated that and where . since for all distinct , and we still have
On the other hand, , hence has ones on the main diagonal and 0s elsewhere.
Exercise D.2-5
We only prove one direction, as the opposite is symmetrical due to . Assume every entry of is real. Let its inverse be represented as a sequence of complex column vectors for . Here, real -vectors and denote the real and imaginary part of , respectively. We know that , hence is a real -vector with single 1 at index and 0s elsewhere. Consequently, based on Theorem D.1 and Theorem D.2. The previous reasoning applies for all , thus is also a real matrix.
Exercise D.2-6
Exercise D.2-7
Let be a matrix represented as a sequence of column vectors and let be an -vector, thus . If has a full column rank, then all column vectors are linearly independent. Therefore, . In the opposite direction, assume, for contradiction, that the column rank of is less than even though does not have a null vector. Consequently, the column vectors are linearly dependent, so there is some such that . But this contradicts the initial assumption of . This concludes the proof of Theorem D.2.
Exercise D.2-8
Let and be matrices whose ranks are and , respectively. According to the alternate definition of the rank of a matrix, we have that and . Let be a matrix whose rank is , so . We can express in two different ways as follows:
Therefore, we can conclude that .
Now, assume that is a nonsingular matrix, so . If is also a nonsingular matrix then obviously and , so equality holds. Thus, let be a matrix whose rank is . We have
Consequently, . The mirror case, where only B is nonsingular, is proven analogously.
Problem D-1
We prove the statement from the book using induction on . The base case trivially holds, as .
For we rely on Theorem D.4, while applying the hint from the book, which says that a determinant is unchanged if entries in one row/column are added to those in another row/column. This can be extended to the case where entries are previously scaled by some factor . Theorem D.4 also states that a determinant is multiplied by such factor if all entries in some row/column are multiplied by it. So, we can simply multiply all entries by , add such scaled entries to another row/column, and divide the previously multiplied elements by .
We transform the Vandermonde matrix , by following the hint, to get
Each row is scaled by the factor for , so we have
According to the principles of mathematical induction, we can conclude that the statement from the book holds for all .
Problem D-2
a.
Let be a matrix represented as a sequence of column vectors for . WLOG assume that the first column vectors are linearly independent, otherwise we would explicitly work with a set of indices that denote those linearly independent objects. We prove the upper and lower bound separately.
For any -bit vector we can always find a matching -bit vector , with 0s on the last positions, such that . This is attainable, as the first linearly independent column vectors of can represent all possible combinations of the last dependent vectors. There are at most ways of choosing for all possible inputs , hence .
We claim that any unique combination of the first bits of an -bit vector produces a distinct image . Suppose, for contradiction, that there is another -bit vector such that . Assume that both and has 0s on the last bit positions. Therefore, , so is a null vector of . But this contradicts the assumption that has rank , since it turns out that neither the first column vectors are linearly independent (see also Exercise D.2-7). Thus, .
and . Therefore, must have a full rank to define a permutation on .
b.
Suppose, for contradiction, there is some such that . By the pigeonhole principle there would be at least one duplicated segment of bits linked to two different segments of bits associated with linearly independent columns of . But this would mean that , which is a contradiction (see part (a)). Therefore, . On the other hand, since no could have preimage of size less then without some other having more, which we have just shown is impossible. We know that and everything must add up. We can conclude that .
c.
At first glance, this task seems complicated, but it is essentially a corollary of parts (a) and (b). Let be a matrix represented as a sequence of column vectors for as before. An image of is a linear combination of these vectors . The identifier of the block where belongs spans its last bit positions, while the first bits specify the offset. Recall that every block is of size . This mirrors the structure of each column vector of , hence only the last rows of determine the block identifier. Inputs from separate blocks differ by a constant factor , so their images will be shifted by the same amount. WLOG we may assume that our source block is block 0. Consequently, all input -bit vectors have 0s in the last positions, thus only the leftmost columns of are relevant. Evidently, all the remaining columns of would be multiplied by 0s during the matrix-vector multiplication .
A fundamental property of any matrix is that its row rank always equals its column rank. The fact that is the rank of the lower-left submatrix of implies that there are linearly independent column vectors among the leftmost columns of . Every distinct combination of these vectors produces a unique block identifier. As a corollary of part (a) we have that . Furthermore, as a corollary of part (b), we can also conclude that for each block in , exactly numbers in map to that block.
d.
There are permutations of . The number of linear permutations is upper bounded by , where (total count of 0-1 matrices) and (total count of -bit vectors). For it holds that .
e.
We start by answering the hint from the book. For a given permutation, multiplying a matrix with a unit vector results in selecting the ith column vector for . By part (d) the smallest is 3, so choose with the following partial permutation : , , , . Now all parameters are fixed. Currently, 3 is mapped to 6, so just choose something else. Here is the rest of the permutation that cannot be achieved with a linear permutation model: , , , .
Last updated