A square matrix is block-diagonal if it has the form
where the blocks are square sub-matrices centered on the main diagonal, and everything outside the blocks is 0. Here is an example of a block-diagonal matrix with a block and a block:
Block-diagonal matrices are extremely useful because they represent linear transformations that can be decomposed into simpler linear transformations. Consider again the generic block-diagonal matrix
For each , let be the length of the block (i.e. is a matrix). Then is an matrix where . For each (where is or or your favourite field), we can write
where consists of the first components of , consists of the next components of , and so on. The key fact is that
Each block independently affects its own chunk of components , and the final product is equal to the direct sum of the vectors . Hence, we often say that is the direct sum of the .
Computing all the and then “concatenating” the results into one giant vector is easier than computing directly without taking advantage of the blocks. The smaller that each block is, the fewer arithmetic operations are required to compute . This is most obvious when each block is a matrix (i.e. a single number), in which case is just a diagonal matrix and is very easy to compute.
Computing is also easier using the —it turns out that
A short and easy proof can be found on Math StackExchange. This determinant formula also implies that the characteristic polynomial of is the product of the characteristic polynomials of the (because is also a block-diagonal matrix with blocks ).
Finally, if
is a block-diagonal matrix with the same block sizes as (i.e. and have the same dimensions for all ), then the product is what you would probably expect: