Marshall Hall's conjecture

From formulasearchengine
Revision as of 16:48, 18 December 2013 by en>ChrisGualtieri (Remove stub template(s). Page is start class or higher. Also check for and do General Fixes + Checkwiki fixes using AWB)
Jump to navigation Jump to search

In mathematics, especially in linear algebra and matrix theory, the commutation matrix is used for transforming the vectorized form of a matrix into the vectorized form of its transpose. Specifically, the commutation matrix K(m,n) is the mn × mn matrix which, for any m × n matrix A, transforms vec(A) into vec(AT):

K(m,n) vec(A) = vec(AT) .

Here vec(A) is the mn × 1 column vector obtain by stacking the columns of A on top of one another:

vec(A) = [ A1,1, ..., Am,1, A1,2, ..., Am,2, ..., A1,n, ..., Am,n ]T

where A = [Ai,j].

The commutation matrix is a special type of permutation matrix, and is therefore orthogonal. It is also an involution and symmetric.

The main use of the commutation matrix, and the source of its name, is to commute the Kronecker product: for every m × n matrix A and every r × q matrix B,

K(r,m)(A B)K(n,q) = B A.

An explicit form for the commutation matrix is as follows: if er,j denotes the j-th canonical vector of dimension r (i.e. the vector with 1 in the j-th coordinate and 0 elsewhere) then

K(r,m) = er,iem,jTem,jer,iT.

References

Jan R. Magnus and Heinz Neudecker (1988), Matrix Differential Calculus with Applications in Statistics and Econometrics, Wiley.


Template:Linear-algebra-stub