A matrix is a bi-dimensional rectangular array of
expressions arranged in rows and columns
An
matrix has
rows and
columns. Given a matrix
,
the notation
refers to the element of
in the
th
row and the
th
column, counting from the top and from the left,
respectively.
Binary Operations
Addition
Substraction
Scaling (Scalar
Multiplication)
Given constant
:
Vector Dot Product
(or Inner Product)
Denoted
or
given two vectors
and
.
Note that the vectors must have the same number of rows, and
that the result of a dot product is a scalar.
Two vectors
and
are orthogonal if
.
Vector dot product of
and
is equivalent to the matrix product of
and
:
Vector Outer Product
Given two vectors
and
with the same number of elements, the outer product between them
is
,
where the result is always a square matrix:
Vector Product (or
Cross Product)
Note that the vectors must have the same number of rows, and
that the result of a cross product is another vector of the same
number of rows.
Cross product is not commutative:
.
Matrix-Vector
Multiplication
Given matrix
and vector
,
the number of columns in
must equal the number of rows in
:
The resulting matrix has the same number of rows as
,
but only 1 column.
Note that the following addition is a linear combination:
Notice that given matrices
and vectors
and
,
is equivalent to
.
Matrix Multiplication
Given matrices
and
,
the number of columns in
must match the number of rows in
.
Note that matrix multiplication is associative:
but its not commutative:
.
Multiplying a
matrix with a
matrix looks like this:
Scalar Division
Given constant
:
Matrix Division
Dividing
is the same as multiplying
by the inverse of
:
.
Unary Operations
Trace
The trace of a square matrix its the sum of its diagonal, and
its defined as
.
For example:
Given constant
,
then
.
The trace function is commutative and associative:
,
and
.
Also
.
Vector Norm (length)
Given vector
,
the norm of
is the absolute value:
,
which is also equal to the square root of the dot product of
with itself:
.
Unit Vector
The unit vector of vector
is
divided by its norm:
.
Minor
The minor of an entry
of a square matrix
is the determinant of the square submatrix of
when the
row and
column (indexed by 1) are removed, and is denoted
.
For example, given:
,
its minor
is
.
Cofactor
The cofactor of an entry
of a square matrix
is denoted
or
,
and is defined as the entry’s minor with alternating sign
depending on the indexes:
.
Adjugate
The adjugate matrix of
matrix
is another
where every entry of
is replaced by its cofactor.
For example,
as:
Determinant
The determinant of a square matrix
is a scalar denoted
or
.
The determinant of a
matrix is the element itself:
.
Given a
matrix:
.
For
and larger matrices
,
the determinant is defined recursively:
where
is the number of columns in
.
The following laws hold given two square matrices
and
:
-
where
is the number of rows in
The rows of a matrix
are linearly independent if
.
We can say
if any of the rows of
is all zeroes. Also, matrix
is not invertible if
.
If
then
is deficient, and full otherwise.
Given row operations:
- Adding a multiple of one row to another row doesn’t change
the determinant of the matrix
- Swapping rows changes the sign of the determinant
- Multiplying a row by a constant is equal to multiplying the
determinant by the same constant
Considering RREF, given square matrix
,
then
implies that
.
Also, if
,
then
,
and conversely, if
,
then
.
Inverse
A matrix
is the inverse of matrix
if either
or
.
The Invertible
Matrix Theorem states that for any square matrix
,
the following statements are either all true or all false:
-
is invertible
-
is invertible
-
has exactly one solution for any
dimensional vector
- The null space of
only contains the zero vector:
-
only has solution
- The rank of
is
- The determinant of
is non zero:
- The RREF of
is the
dimensional identity matrix
- The columns of
are linearly independent
- The rows of
are linearly independent
The following laws hold, given two invertible matrices
and
:
Using Adjugates
We can calculate the inverse of an
square matrix
using its adjugate and determinant as follows:
For example, given
,
we know its adjugate is
and its determinant is
,
so
.
Which we can check as:
Using Gauss-Jordan
Elimination
We can calculate the inverse of an
square matrix
by creating an
matrix that contains
at the left and
at the right:
Given
,
the matrix is then
.
Calculate the RREF of the matrix:
The left side of the RREF should be the identity matrix
(otherwise the matrix is not invertible) and the right side
contains the inverse:
Which we can check as:
Transpose
Matrix transpose flips a matrix by its diagonal, and its
denoted
for a matrix
.
- Given a
matrix:
- Given a
matrix:
- Given a
matrix:
- Given a square matrix:
- Given a
matrix:
The following laws hold, given
and
:
Rank
The rank of a matrix
,
denoted
is a scalar that equals the number of pivots in the RREF of
.
More formally, is the dimension of either the row or column
spaces of
:
.
Basically, the rank describes the number of linearly independent
rows or columns in a matrix.
Nullity
The nullity of a matrix
,
denoted
,
is the number of linearly independent vectors in the null space
of
:
.
The first non-zero element of a matrix row is the leading
coefficient or pivot of the row. A matrix is in
row echelon form (REF) if:
- The leading coefficients of all rows are at the right of the
leading coefficients of the rows above
- All rows containing all zeroes are below the rows with
leading coefficients
For example:
.
The process of bringing a matrix to row echelon form is
called Gaussian Elimination. Starting with the first
row:
- Obtain a leading coefficient 1 in the row by either:
- Swapping the current row with any of the rows below
- Dividing or multiplying the row vector by a constant
- Subtract or add the row one or more times to the rows below
to zero out the leading coefficient column in all the rows
below
- Repeat the process with the row below
For example, given
,
the leading coefficient of the first row is already 1, so we can
move on. The value below the first leading coefficient is 4, so
we can multiply the first vector by 4 and substract it from the
second row:
so the matrix is now
.
The leading coefficient of the third row is 7, so we can
multiply the first row by 7 and substract it from the third row:
so the matrix is now:
.
The entries below the first row’s leading coefficient are zero,
so we can move on to the second row, which we can divide by -3
to make its leading coefficient 1:
,
so the matrix is now:
.
The coefficient below the second row’s leading coefficient is
-6, so we can add the second row multiplied by 6 to it:
so the matrix is now:
and is in row echelon form as the third row is all zeroes.
A matrix is in reduced row echelon form (RREF) if:
- It is in row echelon form (REF)
- The leading coefficients of all non-zero rows are 1
- All the entries above and below a pivot are zero for that
column
The process of bringing a matrix to row echelon form is
called Gaussian-Jordan Elimination. Starting with the
last row with a pivot:
- Add or subtract the row one or more times to the rows above
it to zero out the entries above the pivot in that column
- Repeat the process with the row above
For example, given
,
the last row with a pivot is the second row. The entry above the
leading coefficient is 2, so we can multiply the second row by 2
and substract it from the first row:
,
so the matrix is now:
and is in reduced row echelon form. There is no pivot in the
third column, so the last elements of the first and second rows
don’t need to be zeroed out.
Vector Spaces
The following vector spaces are the fundamental vector
spaces of a matrix. Assume an
matrix
.
Left Space
The set of all vectors
that can multiply
from the left. Basically the vectors
where
is a valid operation. Given an
matrix
,
its left space is
dimensional.
Any element
from the left space can be written as the sum of a vector from
the column space and a vector from the left null space:
Right Space
The set of all vectors
that can multiply
from the right. Basically the vectors
where
is a valid operation. Given an
matrix
,
its right space is
dimensional.
Any element
from the right space can be written as the sum of a vector from
the row space and a vector from the null space:
Row Space
The span of the rows of matrix
:
.
Note that
.
Defined as
.
Column Space
The span of the columns of matrix
:
.
Defined as
.
(Right) Null Space
The set of vectors
where
is the zero vector:
.
It always contains the zero vector. Sometimes called the
kernel of the matrix.
Given matrix
with a null space containing more than the zero vector, then the
equation
has infinite solutions, as the rows in
would not be linearly independent, and given a solution
,
we can add any member of the null space and it would still be a
valid solution.
For example, consider
.
Its null space consists of
and any linear combination of such vector, including the zero
vector. Then consider the equation
.
A valid solution is
as
and
.
But then another valid solution is
as
and
.
Same for any
given any constant
.
If the null space of
only contains the zero vector, then
has exactly one solution, as that solution is
plus any member of the vector space, which is only the zero
vector, and
plus the zero vector is just
.
Left Null Space
The set of vectors
where
is the zero vector. It is denoted as the (right) null space of
the transpose of the input vector:
,
or similarly:
.
We say that matrices
and
are related by a similarity transformation if there
exists an invertible matrix
such that:
.
If the above holds, then the following statements hold as
well:
Special Matrices
Identity Matrix
The identity matrix
is a square matrix with 1’s in the diagonal and 0’s elsewhere.
The
identity matrix is:
.
Given a square and invertible matrix
,
then
.
The identity matrix is symmetric and positive
semidefinite.
Multiplying the
identity matrix with a
dimensional vector is equal to the same vector. Basically
for any vector
.
Elementary Matrices
Every row or column operation that can be performed on a
matrix, such as a row swap, can be expressed as left
multiplication by special matrices called elementary
matrices.
For example, given a
matrix
,
the elementary matrix to swap the first and second rows is
as:
In order to find elementary matrices, we can perform the
desired operation on the identity matrix. In the above case, we
can build a
identity matrix
and then swap the rows:
.
Some more
elementary matrices examples:
- Add
times the second row to the first row:
- Multiply the first row
times:
Diagonal Matrices
A diagonal matrix is a square matrix with values on
the diagonal and zeroes everywhere else, such as:
.
The values on the diagonal are the eigenvalues of
:
.
An
matrix is only diagonalizable if it has
eigenvalues. All normal matrices are diagonalizable.
Properties
Normal
A matrix
is normal if
Orthogonal
A matrix
is orthogonal if
,
which means that
.
All orthogonal matrices are normal. The determinant of an
orthogonal matrix is always -1 or 1.
Symmetric
A matrix
is symmetric if
.
All symmetric matrices are normal. Notice that given any
matrix
,
the matrix
is always symmetric.
Upper Triangular
A matrix
is upper triangular if it contains zeroes below the
diagonal, such as
.
Square
An
matrix
is a square matrix if
.
A trick to convert a non-square matrix into a square matrix is
multiply it by its transpose:
-
has the same column space as
:
-
has the same row space as
:
Positive Semidefinite
A matrix
is positive semidefinite if
.
For example, conside
and let
,
then:
Both
and
,
so
is positive semidefinite.
Positive Definite
A matrix
is positive definite if
.