In mathematics, a square matrix is a matrix with the same number of rows and columns. An n-by-n matrix is known as a square matrix of order . Any two square matrices of the same order can be added and multiplied.

Square matrices are often used to represent simple linear transformations, such as shearing or rotation. For example, if is a square matrix representing a rotation (rotation matrix) and is a column vector describing the position of a point in space, the product yields another column vector describing the position of that point after that rotation. If is a row vector, the same transformation can be obtained using , where is the transpose of .
Main diagonal Edit
The entries (i = 1, ..., n) form the main diagonal of a square matrix. They lie on the imaginary line which runs from the top left corner to the bottom right corner of the matrix. For instance, the main diagonal of the 4×4 matrix above contains the elements a11 = 9, a22 = 11, a33 = 4, a44 = 10.
The diagonal of a square matrix from the top right to the bottom left corner is called antidiagonal or counterdiagonal.
Special kinds Edit
Name | Example with n = 3 |
---|---|
Diagonal matrix | |
Lower triangular matrix | |
Upper triangular matrix |
Diagonal or triangular matrix Edit
If all entries outside the main diagonal are zero, is called a diagonal matrix. If only all entries above (or below) the main diagonal are zero, is called an upper (or lower) triangular matrix.
Identity matrix Edit
The identity matrix of size is the matrix in which all the elements on the main diagonal are equal to 1 and all other elements are equal to 0, e.g.
Invertible matrix and its inverse Edit
A square matrix is called invertible or non-singular if there exists a matrix such that[1][2]
Symmetric or skew-symmetric matrix Edit
A square matrix that is equal to its transpose, i.e., , is a symmetric matrix. If instead , then is called a skew-symmetric matrix.
For a complex square matrix , often the appropriate analogue of the transpose is the conjugate transpose , defined as the transpose of the complex conjugate of . A complex square matrix satisfying is called a Hermitian matrix. If instead , then is called a skew-Hermitian matrix.
By the spectral theorem, real symmetric (or complex Hermitian) matrices have an orthogonal (or unitary) eigenbasis; i.e., every vector is expressible as a linear combination of eigenvectors. In both cases, all eigenvalues are real.[3]
Definite matrix Edit
Positive definite | Indefinite |
---|---|
Q(x,y) = 1/4 x2 + y2 | Q(x,y) = 1/4 x2 − 1/4 y2 |
Points such that Q(x, y) = 1 (Ellipse). |
Points such that Q(x, y) = 1 (Hyperbola). |
A symmetric n×n-matrix is called positive-definite (respectively negative-definite; indefinite), if for all nonzero vectors the associated quadratic form given by
A symmetric matrix is positive-definite if and only if all its eigenvalues are positive.[5] The table at the right shows two possibilities for 2×2 matrices.
Allowing as input two different vectors instead yields the bilinear form associated to A:[6]
Orthogonal matrix Edit
An orthogonal matrix is a square matrix with real entries whose columns and rows are orthogonal unit vectors (i.e., orthonormal vectors). Equivalently, a matrix A is orthogonal if its transpose is equal to its inverse:
An orthogonal matrix A is necessarily invertible (with inverse A−1 = AT), unitary (A−1 = A*), and normal (A*A = AA*). The determinant of any orthogonal matrix is either +1 or −1. The special orthogonal group consists of the n × n orthogonal matrices with determinant +1.
The complex analogue of an orthogonal matrix is a unitary matrix.
Normal matrix Edit
A real or complex square matrix is called normal if . If a real square matrix is symmetric, skew-symmetric, or orthogonal, then it is normal. If a complex square matrix is Hermitian, skew-Hermitian, or unitary, then it is normal. Normal matrices are of interest mainly because they include the types of matrices just listed and form the broadest class of matrices for which the spectral theorem holds.[7]
Operations Edit
Trace Edit
The trace, tr(A) of a square matrix A is the sum of its diagonal entries. While matrix multiplication is not commutative, the trace of the product of two matrices is independent of the order of the factors:
Determinant Edit
The determinant or of a square matrix is a number encoding certain properties of the matrix. A matrix is invertible if and only if its determinant is nonzero. Its absolute value equals the area (in ) or volume (in ) of the image of the unit square (or cube), while its sign corresponds to the orientation of the corresponding linear map: the determinant is positive if and only if the orientation is preserved.
The determinant of 2×2 matrices is given by
The determinant of a product of square matrices equals the product of their determinants:[9]
Eigenvalues and eigenvectors Edit
A number λ and a non-zero vector satisfying
See also Edit
Notes Edit
- ^ Brown 1991, Definition I.2.28
- ^ Brown 1991, Definition I.5.13
- ^ Horn & Johnson 1985, Theorem 2.5.6
- ^ Horn & Johnson 1985, Chapter 7
- ^ Horn & Johnson 1985, Theorem 7.2.1
- ^ Horn & Johnson 1985, Example 4.0.6, p. 169
- ^ Artin, Algebra, 2nd edition, Pearson, 2018, section 8.6.
- ^ Brown 1991, Definition III.2.1
- ^ Brown 1991, Theorem III.2.12
- ^ Brown 1991, Corollary III.2.16
- ^ Mirsky 1990, Theorem 1.4.1
- ^ Brown 1991, Theorem III.3.18
- ^ Eigen means "own" in German and in Dutch.
- ^ Brown 1991, Definition III.4.1
- ^ Brown 1991, Definition III.4.9
- ^ Brown 1991, Corollary III.4.10
References Edit
- Brown, William C. (1991), Matrices and vector spaces, New York, NY: Marcel Dekker, ISBN 978-0-8247-8419-5
- Horn, Roger A.; Johnson, Charles R. (1985), Matrix Analysis, Cambridge University Press, ISBN 978-0-521-38632-6
- Mirsky, Leonid (1990), An Introduction to Linear Algebra, Courier Dover Publications, ISBN 978-0-486-66434-7
External links Edit
- Media related to Square matrices at Wikimedia Commons