A list of useful properties and their proofs in linear algebra. Some of them are also very useful in machine learning. Complex proofs are not in the scope of the post, but the references will be given if interested.
0. Basics
Proof 0.1
Let A and B be two square matrix, then
Proof: https://proofwiki.org/wiki/Determinant_of_Matrix_Product
Proof 0.2
The determinant of a orthogonal matrix must be
Because
Proof 0.3
Let be an matrix and let be its eigenvalues, then
Proof: https://yutsumura.com/determinant-trace-and-eigenvalues-of-a-matrix/
1. Symetric Matrix
Definition
Matrix is symmetric if
Proof 1.1
If is the eigen-value of , so is the conjugate of , denoted as
If is the eigen-vector of , so is the conjugate of , denoted as
Since is a real matrix,
Knowing that
Proof 1.2
has only real eigenvalues
Consider and are the eigne-value and eigen-vector of , repectively. Based on Proof 1.1,
Since is a vector and ,
Since ,
Proof 1.3
is diagonalizable by an orthogonal matrix.
Schur decomposition:
Every square matrix factors into where is upper triangular and . If has real eigenvalues then and can be chosen real: (a.k.a is an orthogonal matrix)
Based on Proof 1.2, all the eigen values of are real.
Based on Schur decomposition, .
Then,
Denote the diagonal matrix as , we have
Proof 1.4
If is nonsingular, is symmetric
Since is invertible,
Taking the transpose, we have
Hence,
2. Positive definite symmetric matrix
Definition
A real symmetric matrix is called positive definite if for all non-zero vectors .
Proof 2.1
The eigenvalues of a real symmetric positive-definite matrix are all positive.
Let be a (real) eigenvalue of and let be a corresponding real eigenvector. That is, we have
Then we multiply by on left and obtain,
The left hand side is positive as is positive definite and is a nonzero vector as it is an eigenvector.
Since the norm is positive, we must have is positive.
It follows that every eigenvalue of is real.
Proof 2.2
If eigenvalues of a real symmetric matrix are all positive, then is positive-definite.
Since Proof 1.3, where , we have
where
Putting , we can rewrite the above equation as
Let
Then we have
By assumption eigenvalues are positive.
Also, since is a nonzero vector and is invertible, is not a zero vector.
Thus the sum expression above is positive, hence is positive for any nonzero vector .
Therefore, the matrix is positive-definite.
Proof 2.3
A is invertible
Method 1
Since Proof 2.1, the matrix does not have 0 as an eigenvalue
We can prove this by contradiction:
If for some then by definition of eigenvalues (non-invertible), is an eigenvector with eigenvalue
Method 2
We can prove this by using determinent
Proof 2.4
the inverse of A is positive-definite
Based on Proof 2.1 and Proof 2.2, we know the fact that a symmetric matrix is positive-definite if and only if its eigenvalues are all positive.
All eigenvalues of are of the form , where is an eigenvalue of .
Since A is positive-definite, each eigenvalue is positive, hence is positive.
So all eigenvalues of are positive, and it yields that is positive-definite.
3. Matrix calculus
Definition
https://en.wikipedia.org/wiki/Matrix_calculus
Reference
[1] matrix cookbook
[2] matrix identities
Derivatives of Matrices, Vectors and Scalar Form
is scalar and is a column vector
and are column vectors
is matrix, is column vector and is scalar
If is symmetric, then
and are column vectors, is a matrix and is scalar
Derivatives of a Determinant
is a matrix
Derivatives of an Inverse
is a matrix and depends on
If is , then