Matrix Theory and the Structure of Linear Systems
Matrices are not merely rectangular arrays of scalars; they are the canonical representations of linear transformations between finite-dimensional vector spaces. Let and be vector spaces over a field (typically or ) with and . Given bases and , any linear map corresponds uniquely to a matrix .
1. Matrix Algebra and Non-Commutativity
Matrix multiplication is defined to correspond to the composition of linear maps. For and , the product is defined by: A critical property of matrix algebra is its non-commutativity: in general, , even for square matrices. This reflects the fact that the order of performing linear transformations matters (e.g., a rotation followed by a shear is not the same as a shear followed by a rotation).
Transposition and Adjoint
The transpose is defined by . An essential identity in matrix algebra is the reversal of order under transposition: Proof Sketch: The -entry of is the -entry of , which is . The -entry of is , which is identical.
2. Theory of Determinants
The determinant is a functional that characterizes the scaling factor of the -dimensional volume under the transformation. Rigorously, it is the unique alternating multilinear form on the columns of such that .
The Leibniz Formula
For a matrix , the determinant is given by: where is the symmetric group of all permutations of , and is the signature of the permutation (either or ).
Key Properties
- Linearity: The determinant is a linear function of any single row (or column) when others are held fixed.
- Alternating: Switching any two rows multiplies the determinant by .
- Singularity: if and only if is singular (non-invertible), which occurs if the rows are linearly dependent.
- Multiplicativity: .
3. Matrix Inversion and the Adjugate
A square matrix is invertible if there exists such that . The existence of is guaranteed if and only if .
The Adjugate Matrix is the transpose of the cofactor matrix , where ( being the minor). The relationship is: While computationally expensive for large , this formula provides profound theoretical insights into the continuity of the inverse as a function of the matrix entries.
4. Systems of Linear Equations
Consider the system , where , , and .
The Rouché–Capelli Theorem
The system is consistent (has at least one solution) if and only if the rank of the coefficient matrix is equal to the rank of the augmented matrix :
- If , the system has a unique solution.
- If , the system has infinitely many solutions (assuming an infinite field).
5. Rank-Nullity and Subspaces
The structure of a matrix is defined by four fundamental subspaces:
- Column Space : The span of the columns of (image of the map).
- Null Space : The set of all such that (kernel).
- Row Space : The span of the rows.
- Left Null Space : The set of all such that .
The Rank-Nullity Theorem
For any matrix : This theorem bridges the dimension of the input space with the dimensions of the image and the kernel. Note that row rank always equals column rank, even for non-square matrices.
6. Matrix Decompositions
Numerical linear algebra relies on decomposing matrices into simpler forms to facilitate computation.
LU Decomposition
A matrix (often after row permutations ) is factored as , where is lower triangular and is upper triangular. This allows solving via two stages of back-substitution: and , reducing complexity from to once the factorisaton is known.
Cholesky Decomposition
For a Hermitian, positive-definite matrix , there exists a unique lower triangular matrix such that . This is twice as efficient as LU and is numerically more stable.
7. Change of Basis and Similarity
If represents a linear operator in basis , and is the transition matrix to a new basis , the representation of in the new basis is given by the similarity transformation: Matrices and are called similar. Similarity is an equivalence relation that preserves the essential characteristics of the linear operator, known as invariants:
- (Trace)
- (Determinant)
- (Characteristic Polynomial and Eigenvalues)
Computational Implementation
Using scipy to solve a system and perform LU decomposition.
import numpy as np
from scipy.linalg import solve, lu
# Define a system Ax = b
A = np.array([[2, 1, 1],
[4, 3, 3],
[8, 7, 9]])
b = np.array([1, -1, 2])
# 1. Solve the system directly
x = solve(A, b)
print(f"Solution x: {x}")
# 2. LU Decomposition
# P: Permutation matrix, L: Lower triangular, U: Upper triangular
P, L, U = lu(A)
print("P:\n", P)
print("L:\n", L)
print("U:\n", U)
# Verification: P @ L @ U should equal A
assert np.allclose(P @ L @ U, A)