A square matrix is block-diagonal if it has the form

[B1B2Bk]

where the blocks Bi are square sub-matrices centered on the main diagonal, and everything outside the blocks is 0. Here is an example of a 5×5 block-diagonal matrix with a 2×2 block and a 3×3 block:

[12000340000056700891000111213].

Block-diagonal matrices are extremely useful because they represent linear transformations that can be decomposed into simpler linear transformations. Consider again the generic block-diagonal matrix

B=[B1B2Bk].

For each 1ik, let i be the length of the block Bi (i.e. Bi is a i×i matrix). Then B is an n×n matrix where n=i=1ki. For each xFn (where F is R or C or your favourite field), we can write

x=[x1x2xk]

where x1F1 consists of the first 1 components of x, x2F2 consists of the next 2 components of x, and so on. The key fact is that

Bx=[B1B2Bk][x1x2xk]=[B1x1B2x2Bkxk].

Each block Bi independently affects its own chunk of components xi, and the final product Bx is equal to the direct sum of the vectors Bixi. Hence, we often say that B is the direct sum of the Bi.

Computing all the Bixi and then “concatenating” the results into one giant vector is easier than computing Bx directly without taking advantage of the blocks. The smaller that each block is, the fewer arithmetic operations are required to compute Bx. This is most obvious when each block is a 1×1 matrix (i.e. a single number), in which case B is just a diagonal matrix and Bx is very easy to compute.

Computing det(B) is also easier using the Bi—it turns out that det(B)=det(B1)det(B2)det(Bk). A short and easy proof can be found on Math StackExchange. This determinant formula also implies that the characteristic polynomial of B is the product of the characteristic polynomials of the Bi (because Bλ is also a block-diagonal matrix with blocks Biλ).

Finally, if

C=[C1C2Ck]

is a block-diagonal matrix with the same block sizes as B (i.e. Bi and Ci have the same dimensions for all 1ik), then the product BC is what you would probably expect:

BC=[B1C1B2C2BkCk].