A similarity transformation of a matrix to a matrix is expressed as , where
-
- is an invertible matrix called the change of basis matrix.
- is a linear transformation matrix with respect to the basis .
- is the transformed representation of , such that performs the same linear transformation as but with respect to another basis .
Let a vector be with respect to the basis and with respect to the basis . Let another vector be with respect to the basis and with respect to the basis . Consider the transformation of these vectors as follows:
where the first two equations describe change of basis transformations and the last equation is a linear transformation of to in the same basis .
Combining the three equation, we have . If is invertible, we can multiply by on the left to give , where . Comparing and , is the transformed representation of , where performs the same linear transformation as but with respect to another basis . We say that is similar to because has properties that are similar to . For example, the trace of , which is defined as , is the same as the trace of .
To show that , we have,
where is the identity matrix and where we have used the identity for the second equality.
Question
Proof that .
Answer
The common properties of similar matrices are useful for explaining certain group theory concepts, such as why there are exactly 32 crystallographic point groups.
One of the most common applications of similarity transformations is to transform a matrix to a diagonal matrix . Consider the eigenvalue problem and let the eigenvectors of be the columns of :
Since , where , we have
If the eigenvectors of are linearly independent, then is non-singular (i.e. invertible). This allows us to multiply on the left by to give , with the diagonal entries of being the eigenvalues of .
Question
Why is non-singular if the eigenvectors of are linearly independent?
Answer
The eigenvectors are linearly independent if the only solution to is when for all . In other words,
or simply .
We need to show that the only solution to is and this is possible if is invertible such that
Using the same eigenvalue problem, we can show that Hermitian operators are diagonalisable, i.e. (see this article).
Question
Show that a Hermitian matrix can be diagonalised by , i.e. , where is a unitary matrix, and that is also Hermitian.
Answer
A unitary matrix has the property: . If a complete set of orthonormal eigenvectors of are the columns of , we have and because orthonormal eigenvectors are linearly independent. The remaining step is to show that .
For example, . Since is non-singular, multiplying on the right of gives . So, .
To show that is also Hermitian, we have
.
As mentioned above, . The -th column of is , while the -th column of is . So, , where is an eigenvector of and is the corresponding eigenvalue. Therefore, the order of the columns of the change of basis matrix corresponds to the order of the diagonal entries in the diagonal matrix.