Algebra Math Notes โข Study Guide
Linear Algebra 1
Vector Spaces
1-1
Vector Spaces A vector space (or linear space) V over a field F is a set on which the operations addition (+) and scalar multiplication, are defined so that for all ๐ฅ, ๐ฆ, ๐ง โ ๐ and all ๐, ๐ โ ๐น, 0. ๐ฅ + ๐ฆ and ๐๐ฅ are unique elements in V. Closure 1. ๐ฅ + ๐ฆ = ๐ฆ + ๐ฅ Commutativity of Addition 2. Associativity of Addition ๐ฅ + ๐ฆ + ๐ง = ๐ฅ + (๐ฆ + ๐ง) 3. There exists 0 โ ๐ such that for every ๐ฅ โ ๐, ๐ฅ + 0 = ๐ฅ. Existence of Additive Identity (Zero Vector) 4. There exists an element โ ๐ฅ such that ๐ฅ + โ๐ฅ = 0. Existence of Additive Inverse 5. 1๐ฅ = ๐ฅ Multiplicative Identity 6. Associativity of Scalar ๐๐ ๐ฅ = ๐(๐๐ฅ) Multiplication 7. ๐ ๐ฅ + ๐ฆ = ๐๐ฅ + ๐๐ฆ Left Distributive Property 8. Right Distributive Property ๐ + ๐ ๐ฅ = ๐๐ฅ + ๐๐ฅ Elements of F, V are scalars, vectors, respectively. F can be โ, โ, โค/๐, etc. Examples: ๐น๐ ๐นโ ๐๐ ร๐ (๐น) or ๐น ๐ ร๐ โฑ(๐, ๐น) ๐ ๐น or ๐น[๐ฅ] ๐ถ ๐, ๐ , ๐ถ โ
n-tuples with entries from F sequences with entries from F mxn matrices with entries from F functions from set S to F polynomials with coefficients from F continuous functions on ๐, ๐ , (โโ, โ)
Cancellation Law for Vector Addition: If ๐ฅ, ๐ฆ, ๐ง โ ๐ and ๐ฅ + ๐ง = ๐ฆ + ๐ง, then ๐ฅ = ๐ฆ. Corollary: 0 and -x are unique. For all ๐ฅ โ ๐, ๐ โ ๐น, ๏ท 0๐ฅ = 0 ๏ท ๐ฅ0 = 0 ๏ท โ๐ ๐ฅ = โ ๐๐ฅ = ๐(โ๐ฅ)
1-2
Subspaces A subset W of V over F is a subspace of V if W is a vector space over F with the operations of addition and scalar multiplication defined on V. ๐ โ ๐ is a subspace of V if and only if 1. ๐ฅ + ๐ฆ โ ๐ whenever ๐ฅ โ ๐, ๐ฆ โ ๐. 2. ๐๐ฅ โ ๐ whenever ๐ โ ๐น, ๐ฅ โ ๐. A subspace must contain 0.
Any intersection of subspaces of V is a subspace of V. If S1, S2 are nonempty subsets of V, their sum is ๐1 + ๐2 = {๐ฅ + ๐ฆ|๐ฅ โ ๐1 , ๐ฆ โ ๐2 }. V is the direct sum of W1 and W 2 (๐ = ๐1 โ ๐2 ) if W 1 and W 2 are subspaces of V such that ๐1 โฉ ๐2 = {0} and ๐1 + ๐2 = ๐. Then each element in V can be written uniquely as ๐ค1 + ๐ค2 where ๐ค1 โ ๐1 , ๐ค2 โ ๐2 . ๐1 , ๐2 are complementary. ๐1 + ๐2 (๐1 โง ๐2 ) is the smallest subspace of V containing W 1 and W 2, i.e. any subspace containing W 1 and W 2 contains ๐1 + ๐2 . For a subspace W of V, ๐ฃ + ๐ = {๐ฃ + ๐ค|๐ค โ ๐} is the coset of W containing v. ๏ท ๐ฃ1 + ๐ = ๐ฃ2 + ๐ iff ๐ฃ1 โ ๐ฃ2 โ ๐. ๏ท The collection of cosets ๐ ๐ = {๐ฃ + ๐|๐ฃ โ ๐} is called the quotient (factor) space of V modulo W. It is a vector space with the operations o (๐ฃ1 + ๐) + ๐ฃ2 + ๐ = ๐ฃ1 + ๐ฃ2 + ๐ o ๐ ๐ฃ + ๐ = ๐๐ฃ + ๐
1-3
Linear Combinations and Dependence A vector ๐ฃ โ ๐ is a linear combination of vectors of ๐ โ ๐ if there exist a finite number of vectors ๐ข1 , ๐ข2 , โฆ ๐ข๐ โ ๐ and scalars ๐1 , ๐2 , โฆ ๐๐ โ ๐น such that ๐ฃ = ๐1 ๐ข1 + โฏ + ๐๐ ๐ข๐ . v is a linear combination of ๐ข1 , ๐ข2 , โฆ ๐ข๐ . The span of S, span(S), is the set consisting of all linear combinations of the vectors in S. By definition, span ๐ = {0}. S generates (spans) V if span(S)=V. The span of S is the smallest subspace containing S, i.e. any subspace of V containing S contains span(S). A subset ๐ โ ๐ is linearly (in)dependent if there (do not) exist a finite number of distinct vectors ๐ข1 , ๐ข2 , โฆ ๐ข๐ โ ๐ and scalars ๐1 , ๐2 , โฆ ๐๐ , not all 0, such that ๐1 ๐ข1 + โฏ + ๐๐ ๐ข๐ = 0. Let S be a linearly independent subset of V. For ๐ฃ โ ๐ โ ๐, ๐ โช {๐ฃ} is linearly dependent iff ๐ฃ โ span(๐).
1-4
Bases and Dimension A (ordered) basis ฮฒ for V is a (ordered) linearly independent subset of V that generates V. Ex. ๐1 = 1,0, โฆ 0 , ๐2 = 0,1, โฆ 0 , โฆ ๐๐ = (0,0, โฆ 1) is the standard ordered basis for ๐น ๐ . A subset ฮฒ of V is a basis for V iff each ๐ฃ โ ๐ can be uniquely expressed as a linear combination of vectors of ฮฒ. Any finite spanning set S for V can be reduced to a basis for V (i.e. some subset of S is a basis). Replacement Theorem: (Steinitz) Suppose V is generated by a set G with n vectors, and let L be a linearly independent subset of V with m vectors. Then ๐ โค ๐ and there exists a
subset H of G containing ๐ โ ๐ vectors such that ๐ฟ โช ๐ป generates V. Pf. Induct on m. Use induction hypothesis for {๐ฃ1 , โฆ ๐ฃ๐ }; remove a ๐ข1 and replace by ๐ฃ๐ +1 . Corollaries: ๏ท If V has a finite basis, every basis for V contains the same number of vectors. The unique number of vectors in each basis is the dimension of V (dim(V)). ๏ท Suppose dim(V)=n. Any finite generating set/ linearly independent subset contains โฅn/โคn elements, can be reduced/ extended to a basis, and if the set contains n elements, it is a basis. Subsets of V, dim(V)=n
Basis (n elements)
Linearly Independent Sets (โคn elements)
Generating Sets (โฅn elements)
Let W be a subspace of a finite-dimensional vector space V. Then dim(W)โคdim(V). If dim(W)=dim(V), then W=V. dim ๐1 + ๐2 = dim ๐1 + dim ๐2 โ dimโก (๐1 โฉ ๐2 ) dim ๐ = dim ๐ + dimโก (๐ โ ๐) The dimension of V/W is called the codimension of V in W.
1-5
Infinite-Dimensional Vector Spaces Let โฑ be a family of sets. A member M of โฑ is maximal with respect to set inclusion if M is contained in no member of โฑ other than M. (โฑ is partially ordered by โ.) A collection of sets ๐ is a chain (nest, tower) if for each A, B in ๐, either ๐ด โ ๐ต or ๐ต โ ๐ด. (โฑ is totally ordered by โ.) Maximal Principle: [equivalent to Axiom of Choice] If for each chain ๐ โ โฑ, there exists a member of โฑ containing each member of ๐, then โฑ contains a maximal member. A maximal linearly independent subset of ๐ โ ๐ is a subset B of S satisfying (a) B is linearly independent. (b) The only linearly independent subset of S containing B is B. Any basis is a maximal linearly independent subset, and a maximal linearly independent
subset of a generating set is a basis for V. Let S be a linearly independent subset of V. There exists a maximal linearly independent subset (basis) of V that contains S. Hence, every vector space has a basis. Pf. โฑ = linearly independent subsets of V. For a chain ๐, take the union of sets in ๐, and apply the Maximal Principle. Every basis for a vector space has the same cardinality. Suppose ๐1 โ ๐2 โ ๐, S1 is linearly independent and S2 generates V. Then there exists a basis such that ๐1 โ ๐ฝ โ ๐2 . Let ฮฒ be a basis for V, and S a linearly independent subset of V. There exists ๐1 โ ๐ฝ so ๐ โช ๐1 is a basis for V.
1-6
Modules A left/right R-module ๐
๐/๐๐
over the ring R is an abelian group (M,+) with addition and scalar multiplication (๐
ร ๐ โ ๐ or ๐ ร ๐
โ ๐) defined so that for all ๐, ๐ โ ๐
and ๐ฅ, ๐ฆ โ ๐, Left Right 1. Distributive ๐ ๐ฅ + ๐ฆ = ๐๐ฅ + ๐๐ฆ ๐ฅ + ๐ฆ ๐ = ๐ฅ๐ + ๐ฆ๐ 2. Distributive ๐ + ๐ ๐ฅ = ๐๐ฅ + ๐ ๐ฅ ๐ฅ ๐ + ๐ = ๐ฅ๐ + ๐ฅ๐ 3. Associative ๐ฅ๐ ๐ = ๐ฅ(๐๐ ) ๐ ๐ ๐ฅ = ๐๐ ๐ฅ 4. Identity 1๐ฅ = ๐ฅ ๐ฅ1 = ๐ฅ Modules are generalizations of vector spaces. All results for vector spaces hold except ones depending on division (existence of inverse in R). Again, a basis is a linearly independent set that generates the module. Note that if elements are linearly independent, it is not necessary that one element is a linear combination of the others, and bases do not always exist. A free module with n generators has a basis with n elements. V is finitely generated if it contains a finite subset spanning V. The rank is the size of the smallest generating set. Every basis for V (if it exists) contains the same number of elements.
1-7
Algebras A linear algebra over a field F is a vector space ๐ over F with multiplication of vectors defined so that for all ๐ฅ, ๐ฆ, ๐ง โ ๐, ๐ โ ๐น, 1. Associative ๐ฅ ๐ฆ๐ง = ๐ฅ๐ฆ ๐ง 2. Distributive ๐ฅ ๐ฆ + ๐ง = ๐ฅ๐ฆ + ๐ฅ๐ง, ๐ฅ + ๐ฆ ๐ง = ๐ฅ๐ง + ๐ฆ๐ง 3. ๐ ๐ฅ๐ฆ = ๐๐ฅ ๐ฆ = ๐ฅ(๐๐ฆ) If there is an element 1 โ ๐ so that 1๐ฅ = ๐ฅ1 = ๐ฅ, then 1 is the identity element. ๐ is commutative if ๐ฅ๐ฆ = ๐ฆ๐ฅ. Polynomials made from vectors (with multiplication defined as above), linear transformations, and ๐ ร ๐ matrices (see Chapters 2-3) all form linear algebras.
2
Matrices
2-1
Matrices A ๐ ร ๐ matrix has m rows and n columns arranged filled with entries from a field F (or ring R). ๐ด๐๐ = ๐ด(๐, ๐) denotes the entry in the ith column and jth row of A. Addition and scalar multiplication is defined component-wise: ๐ด + ๐ต ๐๐ = ๐ด๐๐ + ๐ต๐๐ ๐๐ด ๐๐ = ๐๐ด๐๐ The ๐ ร ๐ matrix of all zeros is denoted ๐ช๐ or just O.
2-2
Matrix Multiplication and Inverses Matrix product: Let A be a ๐ ร ๐ and B be a ๐ ร ๐ matrix. The product AB is the ๐ ร ๐ matrix with entries ๐
๐ด๐ต
๐๐
=
๐ด๐๐ ๐ต๐๐ , 1 โค ๐ โค ๐, 1 โค ๐ โค ๐ ๐=1
Interpretation of the product AB: 1. Row picture: Each row of A multiplies the whole matrix B. 2. Column picture: A is multiplied by each column of B. Each column of AB is a linear combination of the columns of A, with the coefficients of the linear combination being the entries in the column of B. 3. Row-column picture: (AB)ij is the dot product of row I of A and column j of B. 4. Column-row picture: Corresponding columns of A multiply corresponding rows of B and add to AB. Block multiplication: Matrices can be divided into a rectangular grid of smaller matrices, or blocks. If the cuts between columns of A match the cuts between rows of B, then you can multiply the matrices by replacing the entries in the product formula with blocks (entry i,j is replaced with block i,j, blocks being labeled the same way as entries). The identity matrix In is a nxn square matrix with ones down the diagonal, i.e. 1 if ๐ = ๐ ๐ผ๐ ๐๐ = ๐ฟ๐๐ = 0 if ๐ โ ๐ A is invertible if there exists a matrix A-1 such that ๐ด๐ดโ1 = ๐ดโ1 ๐ด = ๐ผ. The inverse is unique, and for square matrices, any inverse on one side is also an inverse on the other side. Properties of Matrix Multiplication (A is mxn): 1. ๐ด ๐ต + ๐ถ = ๐ด๐ต + ๐ด๐ถ Left distributive 2. Right distributive ๐ด + ๐ต ๐ถ = ๐ด๐ถ + ๐ต๐ถ 3. ๐ผ๐ ๐ด = ๐ด = ๐ด๐ผ๐ Left/ right identity 4. ๐ด ๐ต๐ถ = ๐ด๐ต ๐ถ Associative 5. ๐ ๐ด๐ต = ๐๐ด ๐ต = ๐ด(๐๐ต) 6. ๐ด๐ต โ1 = ๐ต โ1 ๐ดโ1 (A, B invertible) ๐ด๐ต โ ๐ต๐ด: Not commutative Note that any 2 polynomials of the same matrix commute. A nxn matrix A is either a zero divisor (there exist nonzero matrices B, C such that ๐ด๐ต = ๐ถ๐ด = ๐ช) or it is invertible.
The Kronecker (tensor) product of pxq matrix A and rxs matrix B is ๐11 ๐ต โฏ ๐1๐ ๐ต โฎ โฑ โฎ . If v and w are column vectors with q, s elements, ๐ดโจ๐ต = ๐๐1 ๐ต โฏ ๐๐๐ ๐ต ๐ดโจ๐ต ๐ฃโจ๐ค = (๐ด๐ฃ)โจ(๐ต๐ค). Kronecker products give nice eigenvalue relations- for example the eigenvalues are the products of those of A and B. [AMM 107-6, 6/2000]
2-3
Other Operations, Classification The transpose of a mxn matrix A, At, is defined by ๐ด๐ ๐๐ = ๐ด๐๐ . The adjoint or Hermitian of a matrix A is its conjugate transpose: ๐ดโ = ๐ด๐ป = ๐ด๐ Name Definition Properties ๐ Symmetric ๐ด=๐ด Self-adjoint/ Hermitian ๐ด = ๐ดโ ๐ง โ ๐ด๐ง is real for any complex z. Skew-symmetric โ๐ด = ๐ด๐ Skew-self-adjoint/ Skew-Hermitian โ๐ด = ๐ดโ Upper triangular ๐ด๐๐ = 0 for ๐ > ๐ Lower triangular ๐ด๐๐ = 0 for ๐ < ๐ Diagonal ๐ด๐๐ = 0 for ๐ โ ๐ Properties of Transpose/ Adjoint 1. ๐ด๐ต ๐ = ๐ต ๐ ๐ด๐ , ๐ด๐ต โ = ๐ต โ ๐ดโ (For more matrices, reverse the order.) 2. (๐ดโ1 )๐ = ๐ด๐ โ1 3. ๐ด๐ฅ ๐ ๐ฆ = ๐ฅ ๐ ๐ด๐ ๐ฆ = ๐ฅ ๐ (๐ด๐ ๐ฆ), ๐ด๐ฅ โ ๐ฆ = ๐ฅ โ ๐ดโ ๐ฆ = ๐ฅ โ (๐ดโ ๐ฆ) 4. ๐ด๐ ๐ด is symmetric. The trace of a ๐ ร ๐ matrix A is the sum of its diagonal entries: ๐
tr ๐ด =
๐ด๐๐ ๐=1
The trace is a linear operator, and tr ๐ด๐ต = tr ๐ด tr(๐ต). The direct sum ๐ด โ ๐ต of ๐ ร ๐ and ๐ ร ๐ matrices A and B is the ๐ + ๐ ร (๐ + ๐) ๐ด ๐ matrix C given by ๐ถ = , ๐ ๐ต ๐ด๐๐ for 1 โค ๐, ๐ โค ๐ ๐ถ๐๐ = ๐ต๐โ๐ ,๐ โ๐ for ๐ + 1 โค ๐, ๐ โค ๐ + ๐ 0, else
3
Linear Transformations
3-1
Linear Transformations For vector spaces V and W over F, a function ๐: ๐ โ ๐ is a linear transformation (homomorphism) if for all ๐ฅ, ๐ฆ โ ๐ and ๐ โ ๐น, (a) ๐(๐ฅ + ๐ฆ) = ๐(๐ฅ) + ๐(๐ฆ) (b) ๐(๐๐ฅ) = ๐๐(๐ฅ) It suffices to verify ๐(๐๐ฅ + ๐ฆ) = ๐๐(๐ฅ) + ๐(๐ฆ). ๐(0) = 0 is automatic. ๐
๐
๐
๐๐ ๐ฅ๐ = ๐=1
๐๐ ๐(๐ฅ๐ ) ๐=1
Ex. Rotation, reflection, projection, rescaling, derivative, definite integral Identity Iv and zero transformation T0 An endomorphism (or linear operator) is a linear transformation from V into itself. T is invertible if it has an inverse T-1 satisfying ๐๐ โ1 = ๐ผ๐ , ๐ โ1 ๐ = ๐ผ๐ . If T is invertible, V and W have the same dimension (possibly infinite). Vector spaces V and W are isomorphic if there exists a invertible linear transformation (an isomorphism, or automorphism if V=W) ๐: ๐ โ ๐. If V and W are finite-dimensional, they are isomorphic iff dim(V)=dim(W). V is isomorphic to ๐น dim V . The space of all linear transformations โ ๐, ๐ = Hom(๐, ๐) from V to W is a vector space over F. The inverse of a linear transformation and the composite of two linear transformations are both linear transformations. The null space or kernel is the set of all vectors x in V such that T(x)=0. ๐ ๐ = {๐ฅ โ ๐|๐ ๐ฅ = 0} The range or image is the subset of W consisting of all images of vectors in V. ๐
๐ = {๐(๐ฅ)|๐ฅ โ ๐} Both are subspaces. nullity(T) and rank(T) denote the dimensions of N(T) and R(T), respectively. If ๐ฝ = {๐ฃ1 , ๐ฃ2 , โฆ ๐ฃ๐ } is a basis for V, then ๐
๐ = span({๐ ๐ฃ1 , ๐ ๐ฃ2 , โฆ ๐(๐ฃ๐ )}). Dimension Theorem: If V is finite-dimensional, nullity(T)+rank(T)=dim(V). Pf. Extend a basis for N(T) to a basis for V by adding {๐ฃ๐+1 , โฆ , ๐ฃ๐ }. Show {๐(๐ฃ๐+1 ), โฆ , ๐(๐ฃ๐ )} is a basis for R(T) by using linearity and linear independence. T is one-to-one iff N(T)={0}. If V and W have equal finite dimension, the following are equivalent: (a) T is one-to-one. (b) T is onto. (c) rank(T)=dim(V) (a) and (b) imply T is invertible.
A linear transformation is uniquely determined by its action on a basis, i.e., if ๐ฝ = {๐ฃ1 , ๐ฃ2 , โฆ ๐ฃ๐ } is a basis for V and ๐ค1 , ๐ค2 , โฆ ๐ค๐ โ ๐, there exists a unique linear transformation ๐: ๐ โ ๐ such that ๐ ๐ฃ๐ = ๐ค๐ , ๐ = 1,2, โฆ ๐. A subspace W of V is T-invariant if ๐(๐ฅ) โ ๐ for every ๐ฅ โ ๐. TW denotes the restriction of T on W.
3-2
Matrix Representation of Linear Transformation Matrix Representation: Let ๐ฝ = ๐ฃ1 , ๐ฃ2 , โฆ ๐ฃ๐ be an ordered basis for V and ๐พ = ๐ค1 , ๐ค2 , โฆ ๐ค๐ be an ordered basis for W. For ๐ฅ โ ๐, define ๐1 , ๐2 , โฆ ๐๐ so that ๐
๐ฅ= The coordinate vector of x relative to ฮฒ is
๐๐ ๐ข๐ ๐=1
๐1 ๐2 ๐๐ฝ ๐ฅ = ๐ฅ ๐ฝ = โฎ ๐๐ n Note ฯฮฒ is an isomorphism from V to F . The ith coordinate is ๐๐ ๐ฅ = ๐๐ . Suppose ๐: ๐ โ ๐ is a linear transformation satisfying ๐
๐ ๐ฃ๐ =
๐๐๐ ๐ค๐ for 1 โค ๐ โค ๐ ๐=1
๐พ
๐พ
The matrix representation of T in ฮฒ and ฮณ is ๐ด = [๐]๐ฝ = โณ๐ฝ (๐) with entries as defined above. (i.e. load the coordinate representation of ๐ ๐ฃ๐ into the jth column of A.) Properties of Linear Transformations (Composition) 1. ๐ ๐1 + ๐2 = ๐๐1 + ๐๐2 Left distributive 2. Right distributive ๐1 + ๐2 ๐ = ๐1 ๐ + ๐2 ๐ 3. ๐ผ๐ ๐ = ๐ = ๐๐ผ๐ Left/ right identity 4. ๐ ๐๐ = ๐๐ ๐ Associative (holds for any functions) 5. ๐ ๐๐ = ๐๐ ๐ = ๐(๐๐) 6. ๐๐ โ1 = ๐ โ1 ๐ โ1 (T, U invertible) Linear transformations [over finite-dimensional vector spaces] can be viewed as leftmultiplication by matrices, so linear transformations under composition and their corresponding matrices under multiplication follow the same laws. This is a motivating factor for the definition of matrix multiplication. Facts about matrices, such as associativity of matrix multiplication, can be proved by using the fact that linear transformations are associative, or directly using matrices. Note: From now on, definitions applying to matrices can also apply to the linear transformations they are associated with, and vice versa. The left-multiplication transformation ๐ฟ๐ด : ๐น ๐ โ ๐น ๐ is defined by ๐ฟ๐ด ๐ฅ = ๐ด๐ฅ (A is a mxn matrix). Relationships between linear transformations and their matrices: 1. To find the image of a vector ๐ข โ ๐ under T, multiply the matrix corresponding to T
๐พ
๐พ
on the left: ๐ ๐ข ๐พ = [๐]๐ฝ ๐ข ๐ฝ i.e. ๐ฟ๐ด ๐๐ฝ = ๐๐พ ๐ where ๐ด = [๐]๐ฝ . 2. Let V, W be finite-dimensional vector spaces with bases ฮฒ, ฮณ. The function ๐พ ฮฆ: โ ๐, ๐ โ ๐๐ ร๐ (๐น) defined by ฮฆ ๐ = [๐]๐ฝ is an isomorphism. So, for linear transformations ๐, ๐: ๐ โ ๐, ๐พ ๐พ ๐พ a. [๐ + ๐]๐ฝ = [๐]๐ฝ + [๐]๐ฝ ๐พ ๐พ b. [๐๐]๐ฝ = ๐[๐]๐ฝ for all scalars a. c. โ ๐, ๐ has dimension mn. 3. For vector spaces V, W, Z with bases ฮฑ, ฮฒ, ฮณ and linear transformations ๐: ๐ โ ๐, ๐ฝ ๐พ ๐พ ๐: ๐ โ ๐, [๐๐]๐ผ = [๐]๐ฝ [๐]๐ผ . ๐ฝ
๐พ
๐พ
4. T is invertible iff [๐]๐ฝ is invertible. Then [๐ โ1 ]๐พ = ( ๐ ๐ฝ )โ1 .
3-3
Change of Coordinates Let ฮฒ and ฮณ be two ordered bases for finite-dimensional vector space V. The change of ๐พ coordinate matrix (from ฮฒ-coordinates to ฮณ-coordinates) is ๐ = [๐ผ๐ ]๐ฝ . Write vector j of ฮฒ in terms of the vectors of ฮณ, take the coefficients and load them in the jth column of Q. (This is so (0,โฆ1,โฆ0) gets transformed into the jth column.) 1. ๐ โ1 changes ฮณ-coordinates into ฮฒ-coordinates. 2. ๐ ๐พ = ๐ ๐ ๐ฝ ๐ โ1 Two nxn matrices are similar if there exists an invertible matrix Q such that ๐ต = ๐ โ1 ๐ด๐. Similarity is an equivalence relation. Similar matrices are manifestations of the same linear transformation in different bases.
3-4
Dual Spaces A linear functional is a linear transformation from V to a field of scalars F. The dual space is the vector space of all linear functionals on V: ๐ โ = โ(๐, ๐น). V** is the double dual. If V has ordered basis ๐ฝ = {๐ฅ1 , ๐ฅ2 , โฆ ๐ฅ๐ }, then ๐ฝ โ = ๐1 , ๐2 , โฆ ๐๐ (coordinate functionsโthe dual basis) is an ordered basis for V*, and for any ๐ โ ๐ โ , ๐
๐=
๐ ๐ฅ๐ ๐๐ ๐=1
To find the coordinate representations of the vectors of the dual bases in terms of the standard coordinate functions: 1. Load the coordinate representations of the vectors in ฮฒ into the columns of W. 2. The desired representation are the rows of ๐ โ1 . 3. The two bases are biorthogonal. For an orthonormal basis (see section 5-5), the coordinate representations of the basis and dual bases are the same. Let V, W have ordered bases ฮฒ, ฮณ. For a linear transformation ๐: ๐ โ ๐, define its transpose (or dual) ๐ ๐ก : ๐ โ โ ๐ โ by ๐ ๐ก g = g๐. Tt is a linear transformation satisfying ๐ฝโ
[๐ ๐ก ]๐พ โ =
๐
๐พ ๐ฝ
๐ก
.
Define ๐ฅ: ๐ โ โ ๐น by ๐ฅ f = f(๐ฅ), and ๐: ๐ โ ๐ โโ by ๐ ๐ฅ = ๐ฅ. (The input is a function; the output is a function evaluated at a fixed point.) If V is finite-dimensional, ฯ is an
isomorphism. Additionally, every ordered basis for V* is the dual basis for some basis for V. The annihilator of a subset S of V is a subspace of ๐ โ : ๐ 0 = Ann(๐) = {๐ โ ๐ โ |๐ ๐ฅ = 0 โ ๐ฅ โ ๐}
4
Systems of Linear Equations
4-1
Systems of Linear Equations The system of equations
๐11 ๐ฅ1 + โฏ +๐๐1 ๐ฅ๐ = ๐1 โฎ ๐๐1 ๐ฅ1 + โฏ ๐๐๐ ๐ฅ๐ = ๐๐ ๐11 โฏ ๐๐1 ๐1 โฑ โฎ and ๐ = โฎ . The can be written in matrix form as Ax=b, where ๐ด = โฎ ๐๐1 โฏ ๐๐๐ ๐๐ augmented matrix is ๐ด ๐ (the entries of b placed to the right of A). The system is consistent if it has solution(s). It is singular if it has zero or infinitely many solutions. If b=0, the system is homogeneous. 1. Row picture: Each equation gives a line/ plane/ hyperplane. They meet at the solution set. 2. Column picture: The columns of A combine (with the coefficients ๐ฅ1 , โฆ ๐ฅ๐ ) to produce b.
4-2
Elimination There are three types of elementary row/ column operations: (1) Interchanging 2 rows/ columns (2) Multiplying any row/ column by a nonzero scalar (3) Adding any multiple of a row/ column to another row/ column An elementary matrix is the matrix obtained by performing an elementary operation on I n. Any two matrices related by elementary operations are (row/column-)equivalent. Performing an elementary row/ column operation is the same as multiplying by the corresponding elementary matrix on the left/ right. The inverse of an elementary matrix is an elementary matrix of the same type. When an elementary row operation is performed on an augmented matrix or the equation ๐ด๐ฅ = ๐, the solution set to the corresponding system of equations does not change. Gaussian elimination- Reduce a system of equations (line up the variables, the equations are the rows), a matrix, or an augmented matrix by using elementary row operations. Forward pass 1. Start with the first row. 2. Excluding all rows before the current row (row j), in the leftmost nonzero column (column k), make the entry in the current row nonzero by switching rows as necessary. (Type 1 operation) The pivot di is the first nonzero in the current row, the row that does the elimination. [Optional: divide the current row by the pivot to make the entry 1. (2)] 3. Make all numbers below the pivot zero. To make the entry a ik in the ith row 0, subtract row j times the multiplier ๐๐๐ = ๐๐๐ /๐๐ from row i. This corresponds to multiplication by a type 3 elementary matrix ๐๐๐ . 4. Move on to the next row, and repeat until only zero rows remain (or rows are exhausted). Backward pass (Back-substitution) 5. Work upward, beginning with the last nonzero row, and add multiples of each row to
the rows above to create zeros in the pivot column. When working with equations, this is essentially substituting the value of the variable into earlier equations. 6. Repeat for each preceding row except the first. A free variable is any variable corresponding to a column without a pivot. Free variables can be arbitrary, leading to infinitely many solutions. Express the solution in terms of free variables. If elimination produces a contradiction (in A|b, a row with only the last entry a nonzero, corresponding to 0=a), there is no solution. Gaussian elimination produces the reduced row echelon form of the matrix: (Forward/ backward pass accomplished 1, (2), 3/ 4.) 1. Any row containing a nonzero entry precedes any zero row. 2. The first nonzero entry in each row is 1. 3. It occurs in a column to the right of the first nonzero entry in the preceding row. 4. The first nonzero entry in each row is the only nonzero entry in its column. The reduced row echelon of a matrix is unique.
4-3
Factorization Elimination = Factorization Performing Gaussian elimination on a matrix A is equivalent to multiplying A by a sequence of elementary row matrices. If no row exchanges are made, ๐ = ( ๐ธ๐๐ )๐ด, so A can be factored in the form ๐ด=
๐ธ๐๐โ1 ๐ = ๐ฟ๐
where L is a lower triangular matrix with 1โs on the diagonal and U is an upper triangular matrix (note the factors are in opposite order). Note ๐ธ๐๐ and ๐ธ๐๐โ1 differ only in the sign of entry (i,j), and the multipliers go directly into the entries of L. U can be factored into a diagonal matrix D containing the pivots and Uโ an upper triangular matrix with 1โs on the diagonal: ๐ด = ๐ฟ๐ท๐โฒ The first factorization corresponds to the forward pass, the second corresponds to completing the back substitution. If A is symmetric, ๐ โฒ = ๐ฟ๐ . Using ๐ด = ๐ฟ๐, ๐ฟ๐ ๐ฅ = ๐ด๐ฅ = ๐ can be split into two triangular systems: 1. Solve ๐ฟ๐ = ๐ for c. 2. Solve ๐๐ฅ = ๐ for x. A permutation matrix P has the rows of I in any order; it switches rows. If row exchanges are required, doing row exchanges 1. in advance gives ๐๐ด = ๐ฟ๐. 2. after elimination gives ๐ด = ๐ฟ1 ๐1 ๐1 .
4-4
The Complete Solution to Ax=b, the Four Subspaces The rank of a matrix A is the rank of the linear transformation LA, and the number of pivots after elimination.
Properties: 1. Multiplying by invertible matrices does not change the rank of a matrix, so elementary row and column matrices are rank-preserving. 2. rank(At)=rank(A) 3. Ax=b is consistent iff rank(A)=rank(A|b). 4. Rank inequalities Linear transformations T, U Matrices A, B rank(TU) โค min(rank(T), rank(U)) rank(AB) โค min(rank(A), rank(B)) Four Fundamental Subspaces of A 1. The row space C(AT) is the subspace generated by rows of A, i.e. it consists of all linear combinations of rows of A. a. Eliminate to find the nonzero rows. These rows are a basis for the row space. 2. The column space C(A) is the subspace generated by columns of A. a. Eliminate to find the pivot columns. These columns of A (the original matrix) are a basis for the column space. The free columns are combinations of earlier columns, with the entries of F the coefficients. (See below) b. This gives a technique for extending a linearly independent set to a basis: Put the vectors in the set, then the vectors in a basis down the columns of A. 3. The nullspace N(A) consists of all solutions to ๐ด๐ฅ = 0. a. Finding the Nullspace (after elimination) i. Repeat for each free variable x: Set x=1 and all other free variables to 0, and solve the resultant system. This gives a special solution for each free variable. ii. The special solutions found in (1) generate the nullspace. b. Alternatively, the nullspace matrix (containing the special solutions in its โ๐น ๐ผ ๐น columns) is ๐ = when the row reduced echelon form is ๐
= . If ๐ผ 0 0 columns are switched in R, corresponding rows are switched in N. 4. The left nullspace N(AT) consists of all solutions to ๐ด๐ ๐ฅ = 0 or ๐ฅ ๐ ๐ด = 0. Fundamental Theorem of Linear Algebra (Part 1): Dimensions of the Four Subspaces: A is mxn, rank(A)=r (If the field is complex, replace ๐ด๐ by ๐ดโ .)
Row space ๐ถ ๐ด๐ โข {๐ด๐ ๐ฆ} โข Dimension r
Row rank = column rank
Column space ๐ถ(๐ด) โข {๐ด๐ฅ} โข Dimension r
๐น ๐ = ๐ถ(๐ด)โจ๐(๐ด๐ ) ๐น ๐ = ๐ถ ๐ด ๐ โจ๐(๐ด)
Nullspace ๐(๐ด) โข {๐ฅ|๐ด๐ฅ = 0} โข Dimension n-r
Left nullspace ๐(๐ด๐ ) โข {๐ฆ|๐ด๐ ๐ฆ = 0} โข Dimension m-r
The relationships between the dimensions can be shown using pivots or the dimension theorem. The Complete Solution to Ax=b 1. Find the nullspace N, i.e. solve Ax=0. 2. Find any particular solution xp to Ax=b (there may be no solution). Set free variables to 0. 3. The solution set is ๐ + ๐ฅ๐ ; i.e. all solutions are in the form ๐ฅ๐ + ๐ฅ๐ , where ๐ฅ๐ is in the nullspace and ๐ฅ๐ is a particular solution.
4-5
Inverse Matrices A is invertible iff it is square (nxn) and any one of the following is true: 1. ๐ด has rank n, i.e. ๐ด has n pivots. 2. ๐ด๐ฅ = ๐ has exactly 1 solution. 3. Its columns/ rows are a basis for ๐น ๐ . Gauss-Jordan Elimination: If A is an invertible nxn matrix, it is possible to transform (A|In) into (In|A-1) by elementary row operations. Follow the same steps as in Gaussian elimination, but on (A|In). If A is not invertible, then such transformation leads to a row whose first n entries are zeros.
5
Inner Product Spaces
5-1
Inner Products An inner product on a vector space V over F (โ or โ) is a function that assigns each ordered pair (๐ฅ, ๐ฆ) โ ๐ a scalar ๐ฅ, ๐ฆ , such that for all ๐ฅ, ๐ฆ, ๐ง โ ๐ and ๐ โ ๐น, 1. ๐ฅ + ๐ง, ๐ฆ = ๐ฅ, ๐ฆ + ๐ง, ๐ฆ 2. ๐๐ฅ, ๐ฆ = ๐ ๐ฅ, ๐ฆ (The inner product is linear in its first component.) 3. ๐ฅ, ๐ฆ = ๐ฆ, ๐ฅ (Hermitian) 4. ๐ฅ, ๐ฅ > 0 for ๐ฅ > 0. (Positive) V is called an inner product space, also an Euclidean/ unitary space if F is โ/ โ. The inner product is conjugate linear in the second component: 1. ๐ฅ, ๐ฆ + ๐ง = ๐ฅ, ๐ฆ + ๐ฅ, ๐ง 2. ๐๐ฅ, ๐ฆ = ๐ ๐ฅ, ๐ฆ If ๐ฅ, ๐ฆ = ๐ฅ, ๐ง for all ๐ฅ โ ๐ then ๐ฆ = ๐ง. The standard inner product (dot product) of ๐ฅ = (๐1 , โฆ , ๐๐ ) and ๐ฆ = (๐1 , โฆ , ๐๐ ) is ๐
๐ฅ โ
๐ฆ = ๐ฅ, ๐ฆ =
๐๐ ๐๐ ๐=1
The standard inner product for the space of continuous complex functions H on [0,2๐] is 1 2๐ ๐, ๐ = ๐ ๐ก ๐(๐ก) ๐๐ก 2๐ 0 A norm of a vector space is a real-valued function โ
satisfying 1. ๐๐ฅ = ๐ ๐ฅ , ๐ โฅ 0 2. ๐ฅ โฅ 0, equality iff ๐ฅ = 0. 3. Triangle Inequality: ๐ฅ + ๐ฆ โค ๐ฅ + ๐ฆ The distance between two vectors x, y is ๐ฅ โ ๐ฆ . In an inner product space, the norm (length) of a vector is ๐ฅ = Cauchy-Schwarz Inequality: ๐ฅ, ๐ฆ โค ๐ฅ
5-2
๐ฅ, ๐ฅ .
๐ฆ
Orthogonality Two vectors are orthogonal (perpendicular) when their inner product is 0. A subset S is orthogonal if any two distinct vectors in S are orthogonal, orthonormal if additionally all vectors have length 1. Subspaces V and W are orthogonal if each ๐ฃ โ ๐ is orthogonal to each ๐ค โ ๐. The orthogonal complement ๐ โฅ (V perp) of V is the subspace containing all vectors orthogonal to V. (Warning: ๐ โฅโฅ = ๐ holds when V is finite-dimensional, not necessarily when V is infinite-dimensional.) When an orthonormal basis is chosen, every inner product on finite-dimensional V is similar to the standard inner product. The conditions effectively determine what the inner product has to be. Pythagorean Theorem: If x and y are orthogonal, ๐ฅ + ๐ฆ
2
= ๐ฅ
2
+ ๐ฆ
Fundamental Theorem of Linear Algebra (Part 2): The nullspace is the orthogonal complement of the row space. The left nullspace is the orthogonal complement of the column space.
2
.
5-3
Projections Take 1: Matrix and geometric viewpoint The [orthogonal] projection of ๐ onto ๐ is ๐, ๐ ๐โ
๐ ๐โ ๐ ๐= ๐ = ๐ = ๐ ๐ 2 ๐โ
๐ ๐โ ๐ ๐ฅ
The last two expressions are for (row) vectors in โ๐ , using the dot product. (Note: this shows that ๐ โ
๐ = ๐ ๐ cos ๐ for 2 and 3 dimensions.) Let ๐ be a finite orthogonal basis. A vector y is the sum of its projections onto the vectors of S: ๐ฆ, ๐ฃ ๐ฆ= ๐ฃ ๐ฃ 2 ๐ฃโ๐
Pf. Write y as a linear combination and take the inner product of y with a vector in the basis; use orthogonality to cancel all but one term. As a corollary, any orthogonal subset is linearly independent. To find the projection of ๐ onto a finite-dimensional subspace W, first find an orthonormal basis for W (see section 5-5), ๐ฝ. The projection is ๐=
๐, ๐ฃ ๐ฃ ๐ฃโ๐ฝ
and the error is ๐ = ๐ โ ๐. ๐ is perpendicular to ๐, and ๐ is the vector in W so that ๐ โ ๐ is minimal. (Proof uses Pythagorean theorem) Besselโs Inequality: (ฮฒ a basis for a subspace) ๐ฃโ๐ฝ
๐ฆ,๐ฃ 2 ๐ฃ 2
2
โค ๐ฆ
, equality iff ๐ฆ =
๐ฆ,๐ฃ ๐ฃโ๐ฝ ๐ฃ 2
๐ฃ
If ๐ฝ = ๐ฃ1 , โฆ , ๐ฃ๐ is an orthonormal basis, then for any linear transformation T,
๐
๐ฝ ๐๐
=
๐ ๐ฃ๐ , ๐ฃ๐ . Alternatively: Let W be a subspace of โ๐ generated by the linearly independent set {๐1 , โฆ ๐๐ }. Solving ๐ดโ ๐ โ ๐ด๐ฅ = 0 โ ๐ดโ ๐ด๐ฅ = ๐ดโ ๐, the projection of ๐ onto W is ๐ = ๐ด๐ฅ = ๐ด ๐ดโ ๐ด โ1 ๐ดโ ๐ ๐
where P is the projection matrix. In the special case that the set is orthonormal, ๐๐ฅ โ ๐ โ ๐ฅ = ๐ ๐ ๐, ๐ = ๐๐ ๐ ๐ ๐
A matrix P is a projection matrix iff ๐2 = ๐. Take 2: Linear transformation viewpoint If ๐ = ๐1 โ ๐2 then the projection on W 1 along W 2 is defined by ๐ ๐ฅ = ๐ฅ1 when ๐ฅ = ๐ฅ1 + ๐ฅ2 ; ๐ฅ1 โ ๐1 , ๐ฅ2 โ ๐2 T is an orthogonal projection if ๐
๐ โฅ = ๐(๐) and ๐ ๐ โฅ = ๐
(๐). A linear operator T is an orthogonal projection iff ๐ 2 = ๐ = ๐ โ .
5-4
Minimal Solutions and Least Squares Approximations When ๐ด๐ฅ = ๐ is consistent, the minimal solution is the one with least absolute value.
1. There exists exactly one minimal solution s, and ๐ โ ๐ถ(๐ดโ ). 2. s is the only solution to ๐ด๐ฅ = ๐ in ๐ถ(๐ดโ ): ๐ด๐ดโ ๐ข = ๐ โ ๐ = ๐ดโ ๐ข = ๐ดโ ๐ด๐ดโ
โ1
๐.
The least squares solution ๐ฅ makes ๐ธ = ๐ด๐ฅ โ ๐ 2 as small as possible. (Generally, ๐ด๐ฅ = ๐ is inconsistent.) Project b onto the column space of A. To find the real function in the form ๐ฆ(๐ก) =
๐ ๐=1 ๐ถ๐ ๐๐ (๐ก)
for fixed functions ๐๐ that is closest to 2
the points ๐ก1 , ๐ฆ1 , โฆ ๐ก๐ , ๐ฆ๐ , i.e. such that the error ๐ = ๐๐=1 ๐๐2 = ๐๐=1 ๐ฆ๐ โ ๐ฆ ๐ก๐ is least, ๐ฆ1 let A be the matrix with ๐ด๐๐ = ๐๐ (๐ก๐ ), ๐ = โฎ . Then ๐ด๐ฅ = ๐ is equivalent to the system ๐ฆ๐ ๐ฆ ๐ก๐ = ๐ฆ๐ . Now find the projection of ๐ onto the columns of ๐ด, by multiplying by ๐ด๐ and solving ๐ด๐ ๐ด๐ฅ = ๐ด๐ ๐. Here, p is the values estimated by the best-fit curve and e gives the errors in the estimates. Ex. Linear functions ๐ฆ = ๐ถ + ๐ท๐ก: 1 ๐ก1 ๐ ๐ก๐ ๐ถ ๐ฆ๐ ๐ด = โฎ โฎ .The equation ๐ด๐ ๐ด๐ฅ = ๐ด๐ ๐ becomes = . 2 ๐ก๐ ๐ฆ๐ ๐ท ๐ก ๐ก ๐ ๐ 1 ๐ก๐ A has orthogonal columns when ๐ก๐ = 0. To produce orthogonal columns, shift the times by ๐ก +โฏ+๐ก ๐ฆ ๐ฆ ๐ก letting ๐๐ = ๐ก๐ โ ๐ก = ๐ก๐ โ 1 ๐ ๐ . Then ๐ด๐ ๐ด is diagonal and ๐ถ = ๐ ๐ , ๐ท = ๐ก๐ 2 ๐ . The least ๐
squares line is ๐ฆ = ๐ถ + ๐ท(๐ก โ ๐ก). Row space ๐ถ ๐ด๐ โข {๐ด๐ ๐ฆ} โข Dimension r
Column space ๐ถ(๐ด) โข {๐ด๐ฅ} โข Dimension r
Least squares solution Minimal solution to ๐ด๐ฅ๐ = ๐ ๐ฅ๐
๐ด๐ฅ๐ = ๐
๐ด+ ๐ = ๐ฅ๐
๐ ๐
๐ด๐ฅ = ๐
๐ถ ๐ด
๐ โฅ
๐ฅ = ๐ฅ๐ + ๐ฅ๐
= ๐(๐ด)
๐ด+ ๐ = ๐ฅ๐
๐ =๐+๐ ๐ถ(๐ด)
โฅ
= ๐(๐ด๐ )
๐ด๐ฅ๐ = 0 ๐ฅ๐ Nullspace ๐(๐ด) โข {๐ฅ|๐ด๐ฅ = 0} โข Dimension n-r
5-5
๐ด+ ๐ = 0
๐
Left nullspace ๐(๐ด๐ ) โข {๐ฆ|๐ด๐ ๐ฆ = 0} โข Dimension m-r
Orthogonal Bases Gram-Schmidt Orthogonalization Process: Let ๐ = ๐ฃ1 , โฆ ๐ฃ๐ be a linearly independent subset of V. Define ๐โฒ = ๐ค1 , โฆ ๐ค๐ by ๐ฃ1 = ๐ค1 and
๐โ1
๐ฃ๐ = ๐ค๐ โ ๐ =1
๐ฆ, ๐ฃ๐ ๐ฃ๐2
๐ฃ๐
Then Sโ is an orthogonal set having the same span as S. To make Sโ orthonormal, divide every vector by its length. (It may be easier to subtract the projections of ๐ค๐ on ๐ค๐ for all ๐ > ๐ at step ๐, like in elimination.) Ex. Legendre polynomials
1
, 2
3
๐ฅ, 2
5 8
3๐ฅ 2 โ 1 , โฆ are an orthonormal basis for โ[๐ฅ]
(integration from -1 to 1). Factorization A=QR From ๐1 , โฆ ๐๐ , Gram-Schmidt constructs orthonormal vectors ๐1 , โฆ ๐๐ . Then ๐ด = ๐๐
๐1โ ๐1 ๐1โ ๐2 โฏ ๐1โ ๐๐ 0 ๐2โ ๐2 โฑ ๐2โ ๐๐ ๐1 โฏ ๐๐ = ๐1 โฏ ๐๐ โฑ โฎ โฎ โฑ โฏ ๐๐โ ๐๐ 0 0 Note R is upper triangular. Suppose ๐ = ๐ฃ1 , โฆ ๐ฃ๐ is an orthonormal set in n-dimensional inner product space V. Then (a) S can be extended to an orthonormal basis {๐ฃ1 , โฆ ๐ฃ๐ } for V. (b) If W=span(S), ๐1 = {๐ฃ๐+1 , โฆ ๐ฃ๐ } is an orthonormal basis for ๐ โฅ . (c) Hence, ๐ = ๐ โ ๐ โฅ and dim ๐ = dim ๐ + dimโก (๐ โฅ ).
5-6
Adjoints and Orthogonal Matrices Let V be a finite-dimensional inner product space over F, and let g: ๐ โ ๐น be a linear transformation. The unique vector ๐ฆ โ ๐ such that g ๐ฅ, ๐ฆ = ๐ฅ, ๐ฆ for all ๐ฅ โ ๐ is given by ๐
๐ฆ=
g(๐ฃ๐ )๐ฃ๐ ๐=1
Let ๐: ๐ โ ๐ be a linear transformation, and ฮฒ and ฮณ be bases for inner product spaces V, ๐ฝ W. Define the adjoint of T to be the linear transformation ๐ โ : ๐ โ ๐ such that ๐ โ ๐พ = ๐พ ( ๐ ๐ฝ )โ. (See section 2.3) Then ๐ โ is the unique (linear) function such that ๐ ๐ฅ , ๐ฆ ๐ = ๐ฅ, ๐ โ ๐ฆ ๐ for all ๐ฅ โ ๐, ๐ฆ โ ๐ and ๐ โ ๐น. A linear operator T on V is an isometry if ๐(๐ฅ) = ๐ฅ for all ๐ฅ โ ๐. If V is finitedimensional, T is orthogonal for V real and unitary for V complex. The corresponding matrix representations, as well as properties of T, are described below. Commutative property Normal ๐ด๐ด๐ = ๐ด๐ ๐ด Complex Normal ๐ด๐ดโ = ๐ดโ ๐ด Linear ๐๐ฃ, ๐๐ค = ๐ โ ๐ฃ, ๐ โ ๐ค Transformation ๐๐ฃ = ๐ โ ๐ฅ Real
Inverse property Orthogonal ๐ด๐ ๐ด = ๐ผ Unitary ๐ดโ ๐ด = ๐ผ ๐๐ฃ, ๐๐ค = ๐ฃ, ๐ค ๐๐ฃ = ๐ฃ (๐๐ฅ)๐ ๐๐ฆ = ๐ฅ ๐ ๐ฆ
Symmetry property Symmetric ๐ด๐ = ๐ด Self-adjoint/ Hermitian ๐ดโ = ๐ด ๐๐ฃ, ๐ค = ๐ฃ, ๐๐ค
A real matrix ๐ has orthonormal columns iff ๐ ๐ ๐ = ๐ผ. If ๐ is square it is called an orthogonal matrix, and its inverse is its transpose. A complex matrix ๐ has orthonormal columns iff ๐ โ ๐ = ๐ผ. If ๐ is square it is a unitary matrix, and its inverse is its adjoint. If ๐ has orthonormal columns it leaves lengths unchanged ( ๐๐ฅ = ๐ฅ for every x) and preserves dot products (๐๐ฅ)๐ ๐๐ฆ = ๐ฅ ๐ ๐ฆ. ๐ดโ ๐ด is invertible iff A has linearly independent columns. More generally, ๐ดโ ๐ด has the same rank as A.
5-7
Geometry of Orthogonal Operators A rigid motion is a function ๐: ๐ โ ๐ satisfying ๐ ๐ฅ โ ๐(๐ฆ) = ๐ฅ โ ๐ฆ for all ๐ฅ, ๐ฆ โ ๐. Each rigid motion is the composition of a translation and an orthogonal operator. A (orthogonal) linear operator is a 1. rotation (around ๐ โฅ ) if there exists a 2-dimensional subspace ๐ โ ๐ and an orthonormal basis ๐ฝ = {๐ฅ1 , ๐ฅ2 } for W, and ๐ such that ๐ฅ1 cos ๐ sin ๐ ๐ฅ1 ๐ ๐ฅ = . 2 โ sin ๐ cos ๐ ๐ฅ2 โฅ and ๐ ๐ฆ = ๐ฆ for ๐ฆ โ ๐ . 2. reflection (about ๐ โฅ ) if W is a one-dimensional subspace of V such that ๐ ๐ฅ = โ๐ฅ for all ๐ฅ โ ๐ and ๐ ๐ฆ = ๐ฆ for all ๐ฆ โ ๐ โฅ . Structural Theorem for Orthogonal Operators: 1. Let T be an orthogonal operator on finite-dimensional real inner product space V. There exists a collection of pairwise orthogonal T-invariant subspaces {๐1 , โฆ , ๐๐ } of V of dimension 1 or 2 such that ๐ = ๐1 โ โฏ โ ๐๐ . Each ๐๐๐ is a rotation or reflection; the number of reflections is even/ odd when det ๐ = 1/ det ๐ = โ1. It is possible to choose the subspaces so there is 0 or 1 reflection. 2. If A is orthogonal there exists orthogonal Q such that ๐ผ๐ โ๐ผ๐ โ1 ๐๐๐ = where p, q are the dimensions of N(T-I), N(T+I) ๐
๐1 โฑ ๐
๐๐ cos ๐ โ sin ๐ and ๐
๐ = . sin ๐ cos ๐ Alternate method to factor QR: Q is a product of reflection matrices ๐ผ โ 2๐ข๐ข๐ and plane rotation matrices (Givens rotation) in the form (1s on diagonal. Shown are rows/ columns i, j). โฑ cosโก (๐) โsinโก (๐) ๐๐๐ = โฑ sinโก (๐) cosโก (๐) โฑ Multiply by ๐๐๐ to produce 0 in the (i,j) position, as in elimination. ๐๐๐ ๐ด = ๐
โ ๐ด =
๐๐๐โ1 ๐
๐
where the factors are reversed in the second product.
6
Determinants
6-1
Characterization The determinant (denoted ๐ด or detโก (๐ด)) is a function from the set of square matrices to the field F, satisfying the following conditions: 1. The determinant of the nxn identity matrix is 1, i.e. det ๐ผ = 1. 2. If two rows of A are equal, then det ๐ด = 0, i.e. the determinant is alternating. 3. The determinant is a linear function of each row separately, i.e. it is n-linear. That is, if ๐1 , โฆ ๐๐ , ๐ข, ๐ฃ are rows with n elements, ๐1 ๐1 ๐1 โฎ โฎ โฎ ๐๐โ1 ๐๐โ1 ๐๐โ1 ๐ข det ๐ข + ๐๐ฃ = det + ๐ detโก ๐ฃ ๐๐+1 ๐๐+1 ๐๐+1 โฎ โฎ โฎ ๐ ๐ ๐๐ ๐ ๐ These properties completely characterize the determinant. 4. The determinant changes sign when two rows are exchanged. 5. Adding a multiple of one row to another row leaves det ๐ด unchanged. 6. A matrix with a row of zeros has det ๐ด = 0. 7. If A is triangular then det ๐ด = ๐11 ๐22 โฏ ๐๐๐ is the product of diagonal entries. 8. A is singular iff det ๐ด = 0. 9. det ๐ด๐ต = det ๐ด detโก (๐ต) ๐ 10. ๐ด has the same determinant as A. Therefore the preceding properties are true if โrowโ is replaced by โcolumnโ.
6-2
Calculation 1. The Big Formula: Use n-linearity and expand everything. det ๐ด =
sgn(๐)๐ด1,๐
1
๐ด2,๐
2
โฏ ๐ด๐ ,๐
๐
๐โ๐๐
1, if ๐ is even . โ1, if ๐ is odd 2. Cofactor Expansion: Recursive, useful with many zeros, perhaps with induction. (Row) where the sum is over all ๐! permutations of {1,โฆn} and sgn ๐ =
๐
det ๐ด = (Column)
๐
๐๐๐ ๐ถ๐๐ = ๐ =1
๐ =1
๐
๐
det ๐ด =
๐๐๐ ๐ถ๐๐ = ๐=1
๐=1
๐๐๐ โ1
๐+๐
det ๐๐๐
๐๐๐ โ1
๐+๐
det ๐๐๐
where ๐๐๐ is A with the ith row and jth column removed. 3. Pivots: If the pivots are ๐1 , ๐2 , โฆ ๐๐ , and ๐๐ด = ๐ฟ๐, (P a permutation matrix, L is lower triangular, U is upper triangular) det ๐ด = det ๐ (๐1 ๐2 โฏ ๐๐ ) where det(P)=1/ -1 if P corresponds to an even/ odd permutation. a. Let ๐ด๐ denote the matrix consisting of the first k rows and columns of A. If
there are no row exchanges in elimination, det ๐ด๐ ๐๐ = det ๐ด๐โ1 4. By Blocks: ๐ด ๐ต a. = A ๐ถ ๐ ๐ถ ๐ด ๐ต ๐ด ๐ต b. = = ๐ด ๐ท โ ๐ถ๐ดโ1 ๐ต ๐ถ ๐ท ๐ ๐ท โ ๐ถ๐ดโ1 ๐ต Tips and Tricks Vandermonde determinant (look at when the determinant is 0, gives factors of polynomial) 1 1 โฏ 1 ๐ฅ1 ๐ฅ2 โฏ ๐ฅ๐ = (๐ฅ๐ โ ๐ฅ๐ ) โฎ โฎ โฑ โฎ ๐>๐ ๐ฅ1๐ โ1 ๐ฅ2๐ โ1 โฏ ๐ฅ๐๐ โ1 Circulant Matrix (find eigenvectors, determinant is product of eigenvalues) ๐0 ๐1 โฏ ๐๐โ1 ๐โ1 ๐โ1 2๐๐ ๐๐ ๐๐โ1 ๐0 โฏ ๐๐โ2 ๐ = ๐ ๐๐ โฎ โฎ โฑ โฎ ๐ =0 ๐=0 โฏ ๐0 ๐1 ๐2 ๐1 ๐ฅ โฏ ๐ฅ ๐ ๐ฅ ๐2 โฏ ๐ฅ ๐๐ โ ๐ฅ โฑ โฎ = ๐1 โ ๐ฅ โฏ ๐๐ โ ๐ฅ + ๐ฅ โฎ โฎ ๐=1 ๐ โ ๐ ๐ฅ ๐ฅ โฏ ๐๐ 1 1 1 = ๐ฅ ๐1 โ ๐ฅ โฏ ๐๐ โ ๐ฅ + + โฏ+ ๐ฅ ๐1 โ ๐ฅ ๐๐ โ ๐ฅ For a real matrix A, det ๐ผ + ๐ด2 = det ๐ผ + ๐๐ด 2 โฅ 0 If A has eigenvalues ๐1 , โฆ , ๐๐ , then det ๐ด + ๐๐ผ = ๐1 + ๐ โฏ (๐๐ + ๐) In particular, if M has rank 1, det ๐ผ + ๐ = 1 + tr ๐
6-3
Properties and Applications Cramerโs Rule: If A is a nxn matrix and det ๐ด โ 0 then ๐ด๐ฅ = ๐ has the unique solution given by detโก (๐ต๐ ) ๐ฅ๐ = ,1 โค ๐ โค ๐ detโก (๐ด) Where ๐ต๐ is A with the ith column replaced by b. Inverses: Let C be the cofactor matrix of A. Then ๐ดโ1 =
๐ถ๐ detโก (๐ด)
The cross product of ๐ข = ๐ข1 , ๐ข2 , ๐ข3 and ๐ฃ = (๐ฃ1 , ๐ฃ2 , ๐ฃ3 ) is ๐ ๐ ๐ ๐ข ร ๐ฃ = ๐ข1 ๐ข2 ๐ข3 ๐ฃ1 ๐ฃ2 ๐ฃ3 a vector perpendicular to u and v (direction determined by the right-hand rule) with length
๐ข
๐ฃ sin ๐ .
Geometry:
๐ฅ1 ๐ฆ1 The area of a parallelogram with vertices sides ๐ฅ1 , ๐ฆ1 , ๐ฅ2 , ๐ฆ2 is ๐ฅ ๐ฆ . (Oriented areas 2 2 satisfy the same properties as determinants.) The area of a parallelepiped with sides ๐ข = ๐ข1 , ๐ข2 , ๐ข3 , ๐ฃ = (๐ฃ1 , ๐ฃ2 , ๐ฃ3 ), and ๐ข = ๐ค1 , ๐ค2 , ๐ค3 ๐ข1 ๐ข2 ๐ข3 is ๐ข ร ๐ฃ โ
๐ค = ๐ฃ1 ๐ฃ2 ๐ฃ3 ๐ค1 ๐ค2 ๐ค3 The Jacobian used to change coordinate systems in integrals is
๐๐ฅ
๐๐ฅ
๐๐ฅ
๐๐ข ๐๐ฆ
๐๐ฃ ๐๐ฆ
๐๐ค ๐๐ฆ
๐๐ข ๐๐ง
๐๐ฃ ๐๐ง
๐๐ค ๐๐ง
๐๐ข
๐๐ฃ
๐๐ค
.
7
Eigenvalues and Eigenvectors, Diagonalization
7-1
Eigenvalues and Eigenvectors Let T be a linear operator (or matrix) on V. A nonzero vector ๐ฃ โ ๐ is an (right) eigenvector of T if there exists a scalar ๐, called the eigenvalue, such that ๐ ๐ฃ = ๐๐ฃ. The eigenspace of ฮป is the set of all eigenvectors corresponding to ฮป: ๐ธ๐ = {๐ฅ โ ๐|๐ ๐ฅ = ๐๐ฅ}. The characteristic polynomial of a matrix A is detโก (๐ด โ ๐๐ผ). The zeros of the polynomial are the eigenvalues of A. For each eigenvalue solve ๐ด๐ฃ = ๐๐ฃ to find linearly independent eigenvalues that span the eigenspace. Multiplicity of an eigenvalue ฮป: 1. Algebraic (๐๐๐๐ )- the multiplicity of the root ฮป in the characteristic polynomial of A. 2. Geometric (๐๐๐๐๐ )- the dimension of the eigenspace of ฮป. 1 โค dim ๐ธ๐ โค ๐๐๐๐ (๐). dim ๐ธ๐ = dim ๐ ๐ด โ ๐๐ผ = ๐ โ rank(๐ด โ ๐๐ผ). For real matrices, complex eigenvalues come in conjugate pairs. The product of the eigenvalues (counted by algebraic multiplicity) equals detโก (๐ด). The sum of the eigenvalues equals the trace of A. An eigenvalue of 0 implies that A is singular. Spectral Mapping Theorem: Let A be a nxn matrix with eigenvalues ๐1 , โฆ , ๐๐ (not necessarily distinct, counted according to algebraic multiplicity), and P be a polynomial. Then the eigenvalues of ๐(๐ด) are ๐ ๐1 , โฆ , ๐ ๐๐ . Gerschgorinโs Disk Theorem: Every eigenvalue of A is strictly in a circle in the complex plane centered at some diagonal entry ๐ด๐๐ with radius ๐๐ = ๐ โ ๐ ๐๐๐ (because ๐ โ ๐ด๐๐ ๐ฅ๐ = ๐ โ ๐ ๐๐๐ ๐ฅ๐ ). Perron-Frobenius Theorem: Any square matrix with positive entries has a unique eigenvector with positive entries (up to multiplication by a positive factor), and the corresponding eigenvalue has multiplicity one and has strictly greater absolute value than any other eigenvalue. Generalization: Holds for any irreducible matrix with nonnegative entries, i.e. there is no reordering of rows and columns that makes it block upper triangular. A left eigenvalue of A satisfies ๐ฃ ๐ ๐ด = ๐๐ฃ instead. Biorthogonality says that any right eigenvector of A associated with ฮป is orthogonal to all left eigenvectors of A associated with eigenvalues other than ฮป.
7-2
Invariant and T-Cyclic Subspaces The subspace ๐ถ๐ฅ = ๐(๐ฅ; ๐) = ๐ = span( ๐ฅ. ๐ ๐ฅ , ๐ 2 ๐ฅ , โฆ ) is the T-cyclic subspace generated by x. W is the smallest T-invariant subspace containing x. 1. If W is a T-invariant subspace, the characteristic polynomial of TW divides that of T. 2. If k=dim(W) then ๐ฝ๐ฅ = {๐ฅ, ๐ ๐ฅ , โฆ , ๐ ๐โ1 ๐ฅ } is a basis for W, called the T-cyclic basis
generated by x. If ๐๐=0 ๐๐ ๐ ๐ (๐ฅ) = 0 with ๐๐ = 1, the characteristic polynomial of TW is โ1 ๐ ๐๐=0 ๐๐ ๐ก ๐ . 3. If ๐ = ๐1 โจ๐2 โฏ ๐๐ , each ๐๐ is a T-invariant subspace, and the characteristic polynomial of ๐๐๐ is ๐๐ (๐ก), then the characteristic polynomial of T is ๐๐=1 ๐๐ (๐ก). Cayley-Hamilton Theorem: A satisfies its own characteristic equation: if ๐(๐ก) is the characteristic polynomial of A, then ๐ ๐ด = ๐ช.
7-3
Triangulation A matrix is triangulable if it is similar to an upper triangular matrix. (Schur) A matrix is triangulable iff the characteristic polynomial splits over F. A real/ complex matrix A is unitarily/ orthogonally equivalent to a real/ complex upper triangular matrix. (i.e. ๐ด = ๐๐๐ โ1 , Q is orthogonal/ unitary) Pf. T=LA has an eigenvalue iff T* has. Induct on dimension n. Choose an eigenvector z of T*, and apply the induction hypothesis to the T-invariant subspace span ๐ง โฅ .
7-4
Diagonalization T is diagonalizable if there exists an ordered basis ๐ฝ for V such that ๐ ๐ฝ is diagonal. A is diagonalizable if there exists an invertible matrix S such that ๐ โ1 ๐ด๐ = ฮ is a diagonal matrix. Let ๐1 , โฆ , ๐๐ be the eigenvalues of A. Let ๐๐ be a linearly independent subset of ๐ธ๐ ๐ for 1 โค ๐ โค ๐. Then ๐๐ is linearly independent. (Loosely, eigenvectors corresponding to different eigenvalues are linearly independent.) T is diagonalizable iff both of the following are true: 1. The characteristic polynomial of T splits (into linear factors). 2. For each eigenvalue, the algebraic and geometric multiplicities are equal. Hence there are n linearly independent eigenvectors T is diagonalizable iff V is the direct sum of eigenspaces of T. To diagonalize A, put the ๐ linearly independent eigenvectors into the columns of A. Put the corresponding eigenvalues into the diagonal entries of ฮ. Then ๐ด = ๐ฮ๐ โ1 or ๐๐ท๐ โ1 For a linear transformation, this corresponds to ๐ฝ ๐พ ๐๐ฝ= ๐ผ๐พ ๐๐พ ๐ผ๐ฝ Simultaneous Triangulation and Diagonalization Commuting matrices share eigenvectors, i.e. given that A and B can be diagonalized, there exists a matrix S that is an eigenvector matrix for both of them iff ๐ด๐ต = ๐ต๐ด. Regardless, AB and BA have the same set of eigenvalues, with the same multiplicities. More generally, let ๐ be a commuting family of triangulable/ diagonalizable linear operators on V. There exists an ordered basis for V such that every operator in ๐ is simultaneously represented by a triangular/ diagonal matrix in that basis.
7-5
Normal Matrices (For review see 5-6)
A nxn [real] symmetric matrix: 1. Has only real eigenvalues. 2. Has eigenvalues that can be chosen to be orthonormal. (๐ = ๐, ๐ โ1 = ๐ ๐ ) (See below.) 3. Has n linearly independent eigenvectors so can be diagonalized. 4. The number of positive/ negative eigenvalues equals the number of positive/ negative pivots. For real/ complex finite-dimensional inner product spaces, T is symmetric/ normal iff there exists an orthonormal basis for V consisting of eigenvectors of T. Spectral Theorem (Linear Transformations) Suppose T is a normal linear operator (๐ โ ๐ = ๐๐ โ ) on a finite-dimensional real/ complex inner product space V with distinct eigenvalues ๐1 , โฆ , ๐๐ (its spectrum). Let ๐๐ be the eigenspace of T corresponding to ๐๐ and ๐๐ the orthogonal projection of ๐ on ๐๐ . 1. T is diagonalizable and ๐ = ๐1 โ โฏ โ ๐๐ . 2. ๐๐ is orthogonal to the direct sum of ๐๐ with ๐ โ ๐. 3. There is an orthonormal basis of eigenvectors. 4. Resolution of the identity operator: ๐ผ = ๐1 + โฏ + ๐๐ 5. Spectral decomposition: ๐ = ๐1 ๐1 + โฏ + ๐๐ ๐๐ Pf. The triangular matrix in the proof of Schurโs Theorem is actually diagonal. 1. If ๐ด๐ฅ = ๐๐ฅ then ๐ดโ ๐ฅ = ๐๐ฅ. 2. W is T-invariant iff ๐ โฅ is ๐ โ -invariant. 3. Take a eigenvector v; let ๐ = span ๐ฃ . From (1) v is an eigenvector of ๐ โ ; from (2) ๐ โฅ is T-invariant. 4. Write ๐ = ๐ โ ๐ โฅ . Use induction hypothesis on ๐ โฅ . (Matrices) Let A be a normal matrix (๐ดโ ๐ด = ๐ด๐ดโ ). Then A is diagonalizable with an orthonormal basis of eigenvectors: ๐ด = ๐ฮ๐ โ where ฮ is diagonal and U in unitary. Type of Matrix Hermitian (Self-adjoint)
Condition ๐ดโ = ๐ด
Unitary
๐ดโ ๐ด = ๐ผ
Symmetric (real)
๐ด๐ = ๐ด
Orthogonal (real)
๐ด๐ ๐ด = ๐ผ
Factorization ๐ด = ๐ฮ๐ โ1 U unitary, ฮ real diagonal Real eigenvalues (because ๐๐ฃ โ ๐ฃ = ๐ฃ โ ๐ด๐ฃ = ๐๐ฃ โ ๐ฃ) ๐ด = ๐ฮ๐ โ1 U unitary, ฮ diagonal Eigenvalues have absolute value 1 ๐ด = ๐ฮ๐โ1 Q orthogonal, ฮ real diagonal Real eigenvalues ๐ด = ๐ฮ๐โ1 Q unitary, ฮ diagonal Eigenvalues have absolute value 1
7-6
Positive Definite Matrices and Operators A real matrix A is positive (semi)definite if ๐ฅ โ ๐ด๐ฅ > 0 (๐ฅ โ ๐ด๐ฅ โฅ 0) for every nonzero vector x. A linear operator T on a finite-dimensional inner product space is positive (semi)definite if T is self-adjoint and ๐ ๐ฅ , ๐ฅ > 0 ( ๐ ๐ฅ , ๐ฅ โฅ 0) for all ๐ฅ โ 0. The following are equivalent: 1. A is positive definite. 2. All eigenvalues are positive. 3. All upper left determinants are positive. 4. All pivots are positive. Every positive definite matrix factors into ๐ด = ๐ฟ๐ท๐ โฒ = ๐ฟ๐ท๐ฟ๐ with positive pivots in D. The Cholesky factorization is ๐ด= ๐ฟ ๐ท ๐ฟ ๐ท
7-7
๐
Singular Value Decomposition Every ๐ ร ๐ matrix A has a singular value decomposition in the form ๐ด๐ = ๐ฮฃ โ ๐ด = ๐ฮฃ๐ โ1 = ๐ฮฃ๐ โ ๐1 where U and V are unitary matrices and ๐ด = is diagonal. The singular values โฑ ๐๐ ๐1 , โฆ ๐๐ (๐๐ = 0 for ๐ > ๐ = rank(๐ด)) are positive and are in decreasing order, with zeros at the end (not considered singular values). If A corresponds to the linear transformation ๐: ๐ โ ๐, then this says there are orthonormal bases ๐ฝ = {๐ฃ1 , โฆ , ๐ฃ๐ } and ๐พ = {๐ข1 , โฆ , ๐ข๐ } such that ๐ ๐ข if 1 โค ๐ โค ๐ ๐ ๐ฃ๐ = ๐ ๐ 0 if ๐ > ๐ Letting ๐ฝ โฒ , ๐พโฒ be the standard ordered bases for V, W, ๐ฝโฒ ๐พโฒ ๐พโฒ ๐พ ๐ด๐ = ๐ฮฃ โ ๐ ๐ฝ โฒ ๐ผ ๐ฝ = ๐ผ ๐พ ๐ ๐ฝ Orthogonal elements in the basis are sent to orthogonal elements; the singular values give the factors the lengths are multiplied by. To find the SVD: 1. Diagonalize ๐ดโ ๐ด, choosing orthonormal eigenvectors. The eigenvalues are the squares of the singular values and the eigenvector matrix is V. ๐12 ๐ดโ ๐ด = ๐ฮฃ 2 ๐ โ = ๐ ๐โ โฑ ๐๐2 2. Similarly, ๐ด๐ดโ = ๐ฮฃ 2 ๐ โ If V and the singular values have already been found, the columns of U are just the images of ๐ฃ1 , โฆ , ๐ฃ๐ under left multiplication by A: ๐ข๐ = ๐ด๐ฃ๐ , unless this gives 0. 3. If A is a mxn matrix: a. The first r columns of V generate the row space of A. b. The last n-r columns generate the nullspace of A. c. The first r columns of U generate the column space of A. d. The last m-r columns of U generate the left nullspace of A.
The pseudoinverse of a matrix A is the matrix ๐ด+ such that for ๐ฆ โ ๐ถ(๐ด), ๐ด+๐ฆ is the vector x in the row space such that ๐ด๐ฅ = ๐ฆ, and for ๐ฆ โ ๐(๐ด๐ ), ๐ด+๐ฆ = 0. For a linear transformation, replace ๐ถ(๐ด) with ๐
(๐) and ๐(๐ด๐ ) with ๐
๐ โฅ . In other words, 1. ๐ด๐ด+ is the projection matrix onto the column space of A. 2. ๐ด+๐ด is the projection matrix onto the row space of A. Finding the pseudoinverse:
๐1โ1
๐ด+ = ๐ฮฃ +๐ โ = ๐
โฑ
๐๐โ1
๐โ
The shortest least squares solution to ๐ด๐ฅ = ๐ is ๐ฅ + = ๐ด+๐. See Section 5-4 for a picture. The polar decomposition of a complex (real) matrix A is ๐ด = ๐๐ป where Q is unitary (orthogonal) and H is semi-positive definite Hermitian (symmetric). Use the SVD: ๐ด = ๐๐ โ (๐ฮฃ๐ โ ) If A is invertible, Q is positive definite and the decomposition is unique.
Summary Type of matrix Real symmetric Orthogonal Skew-symmetric Self-adjoint Positive definite
Eigenvalues Real Absolute value 1 (Pure) Imaginary Real Positive
Eigenvectors (can be chosenโฆ)
Orthogonal
8
Canonical Forms A canonical form is a standard way of presenting and grouping linear transformations or matrices. Matrices sharing the same canonical form are similar; each canonical form determines an equivalence class. Similar matrices shareโฆ ๏ท Eigenvalues ๏ท Trace and determinant ๏ท Rank ๏ท Number of independent eigenvectors ๏ท Jordan/ Rational canonical form
8-1
Decomposition Theorems A minimal polynomial of T is the (unique) monic polynomial ๐(๐ก) of least positive degree such that ๐ ๐ = ๐0 . If ๐ ๐ = ๐0 then ๐ ๐ก |๐(๐ก); in particular, ๐(๐ก) divides the characteristic polynomial of T. Let W be an invariant subspace for T and let ๐ฅ โ ๐. The T-conductor (โT-stufferโ) of x into W is the set ๐๐ (๐ฅ; ๐) which consists of all polynomials g over F such that (๐ ๐ )(๐ฅ) โ ๐. (It may also refer to the monic polynomial of least degree satisfying the condition.) If ๐ = {0}, T is called the T-annihilator of x, i.e. it is the (unique) monic polynomial ๐(๐ก) of least degree for which ๐ ๐ ๐ฅ = 0. The T-conductor/ annihilator divides any other polynomial with the same property. The T-annihilator ๐ ๐ก is the minimal polynomial of TW, where W is the T-cyclic subspace generated by x. The characteristic polynomial and minimal polynomial of TW are equal or negatives. Let L be a linear operator on V, and W a subspace of V. W is T-admissible if 1. W is invariant under T. 2. If ๐ ๐ ๐ฅ โ ๐, there exists ๐ฆ โ ๐ such that ๐ ๐ (๐ฅ) = ๐ ๐ (๐ฆ). Let T be a linear operator on finite-dimensional V. Primary Decomposition Theorem (leads to Jordan form): Suppose the minimal polynomial of T is ๐
๐๐ ๐ ๐
๐ ๐ก = ๐=1
where ๐๐ are distinct irreducible monic polynomials and ๐๐ are positive integers. Let ๐๐ be the null space of ๐๐ ๐ ๐ ๐ . Then 1. ๐ = ๐1 โ โฏ โ ๐๐ . 2. Each ๐๐ is invariant under T. ๐ 3. The minimal polynomial of ๐๐๐ is ๐๐ ๐ . ๐ Pf. Let ๐๐ = ๐ ๐ . Find ๐๐ so that ๐๐=1 ๐๐ ๐๐ = 1. ๐ธ๐ = ๐๐ ๐ ๐๐ (๐) is the projection onto ๐๐ . ๐๐
Cyclic Decomposition Theorem (leads to rational canonical form): Let T be a linear operator on finite-dimensional V and ๐0 (often taken to be {0}) a proper Tadmissible subspace of V. There exist nonzero ๐ฅ1 , โฆ ๐ฅ๐ with (unique) T-annihilators ๐1 , โฆ , ๐๐ , called invariant factors such that
1. ๐ = ๐0 โ ๐ ๐ฅ1 ; ๐ โ โฏ โ ๐(๐ฅ๐ ; ๐) 2. ๐๐ |๐๐โ1 for 2 โค ๐ โค ๐. Pf. 1. There exist nonzero vectors ๐ฝ1 , โฆ , ๐ฝ๐ in V such that a. ๐ = ๐0 + ๐ ๐ฝ1 ; ๐ + โฏ + ๐ ๐ฝ๐ ; ๐ b. If 1 โค ๐ โค ๐ and ๐๐ = ๐0 + ๐ ๐ฝ1 ; ๐ + โฏ + ๐(๐ฝ๐ ; ๐) then ๐๐ has maximum degree among all T-conductors into ๐๐โ1 . 2. Let ๐ = ๐ (๐ฝ; ๐๐โ1 ). If ๐ ๐ (๐ฝ) = ๐ฝ0 + 1โค๐