+ All Categories
Home > Documents > Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2....

Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2....

Date post: 16-Jul-2020
Category:
Upload: others
View: 7 times
Download: 0 times
Share this document with a friend
48
Chapter 2 Dramatis Personae 1 Vector Spaces In four words, the realm of linear algebra can be described as geometry of vector spaces. Axioms By definition, a vector space is a set, equipped with operations of addition and multiplication by scalars which are required to satisfy certain axioms: The set will be denoted here by V , and its elements referred to as vectors. Here are the axioms. (i) The sum of two vectors u and v is a vector (denoted u + v); the result of multiplication of a vector v by a scalar λ is a vector (denoted λv). (ii) Addition of vectors is commutative and associative: u + v = v + u, (u + v)+ w = u +(v + w) for all u, v, w ∈V . (iii) There exists the zero vector (denoted by 0) such that v + 0 = 0 + v = v for every v ∈V . (iv) For every vector u there exists the opposite vector, denoted by u, such that u + u = 0. 21
Transcript
Page 1: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

Chapter 2

Dramatis Personae

1 Vector Spaces

In four words, the realm of linear algebra can be described asgeometry of vector spaces.

Axioms

By definition, a vector space is a set, equipped with operationsof addition and multiplication by scalars which are required tosatisfy certain axioms:

The set will be denoted here by V, and its elements referred toas vectors. Here are the axioms.

(i) The sum of two vectors u and v is a vector (denoted u+ v);the result of multiplication of a vector v by a scalar λ is a vector(denoted λv).

(ii) Addition of vectors is commutative and associative:

u+ v = v + u, (u+ v) +w = u+ (v +w) for all u,v,w ∈ V.

(iii) There exists the zero vector (denoted by 0) such that

v + 0 = 0+ v = v for every v ∈ V.

(iv) For every vector u there exists the opposite vector, denotedby −u, such that

−u+ u = 0.

21

Page 2: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

22 Chapter 2. DRAMATIS PERSONAE

(v) Multiplication by scalars is distributive: For all vectors u,vand scalars λ, µ we have

(λ+ µ)(u+ v) = λu+ λv + µu+ µv.

(vi) Multiplication by scalars is associative in the following sense:For every vector u and all scalars λ, µ we have:

(λµ)u = λ(µu).

(vii) Multiplication by scalars 0 and 1 acts on every vector u as

0u = 0, 1u = u.

We have to add to this definition the following comment aboutscalars. Taking one of the sets R or C of real or complex numberson the role of scalars one obtains the definition of real vector spacesor complex vector spaces. In fact these two choices will suffice forall our major goals. The reader may assume that we use the symbolK to cover both cases K = R and K = C in one go.

On the other hand, any field K would qualify on the role ofscalars, and this way one arrives at the notion of K-vector spaces.We refer to Supplement “Fields” for the definition and examples.

From the axioms, all further properties of vectors are derived. Asan illustration, let us prove that

−u = (−1)u for all u.

Indeed, 1 + (−1) = 0. By the axion (vii), 0u = 0, and hence 0 =(1 + (−1))u = u + (−1)u, where we applied the distributive law(v) and 1u = u from (vii). Now we add −u to both parts, to find−u+ 0) = −u (by (iii)) on the left, and

−u+ (u+ (−1)u) = (−u+ u) + (−1)u = 0+ (−u) = −u

(consecutively by (ii), (iv), and (iii)) on the right. Thus −u = (−1)u.

Such derivations might be a fun and useful game, but this isnot the purpose of introducing the notion of a vector space, nor arethe axioms “absolute truth accepted without proof.” The abstractaxiomatic definition of vector spaces is in fact a unification device:it allows one to study many examples at once without the need toreiterate the same derivations. Yet, to apply general conclusions, thevalidity of the axioms needs to be verified in each example separately.The axioms are modeleled, of course, on the properties of geometricvectors. However, to justify the axiomatic approach, at least onemore example is needed.

Page 3: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 23

Examples

The above definition of vector spaces is doubly abstract: not only itneglects to specify the set V of vectors, but it does not even tell usanything explicit about the nature of the operations of addition ofvectors and multiplication of vectors by scalars. To find various ex-amples of vector spaces we should figure out which operations wouldbe good candidates to satisfy the axioms (i–vii). It turns out that inthe majority of useful examples, the operations are pointwise additionof functions and multiplication of functions by scalars.

Example 1. Let S be any set, and V be the set of all functionson S with values in K. We will denote this set by KS . The sum andmultiplication by scalars are defined on KS as pointwise operationswith functions. Namely, given two functions f, g and a scalar λ, thevalues of the sum f + g and the product λf at a point s ∈ S are

(f + g)(s) = f(s) + g(s), (λf)(s) = λ(f(s)).

It is immediate to check that V = KS equipped with these operationssatisfies the axioms (i–vii). Thus KS is a K-vector space.

Example 1a. Let S be the set of n elements 1, 2, ..., n. Thenthe space KS is the space Kn (e.g. RS = Rn and CS = Cn) ofcoordinate vectors. Namely, each function on the set {1, . . . , n} isspecified by a string x = (x1, ..., xn) of its values, called coordinatesor components of the coordinate vector. By tradition, the string iswritten as a column, and the pointwise operations with functionsturn into termwise operations with the columns:

λ

x1...xn

=

λx1...

λxn

,

x1...xn

+

y1...yn

=

x1 + y1...

xn + yn

.

Example 1b. Let S be the set of all ordered pairs (i, j), wherei = 1, ...,m, j = 1, ..., n. Then the vector space KS is the space ofm×n-matrices. By tradition, a matrix is denoted by an upper-caseletter, e.g. A, and is represented by a rectangular array whose entryat the intersection of the ith row and jth column is an element of Kdenoted by aij (the same lower-case letter with subscripts):

A =

a11 . . . a1n... aij

...am1 . . . amn

.

Page 4: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

24 Chapter 2. DRAMATIS PERSONAE

The poinwise operations with matrices as functions turn into elemen-twise addition of the arrays and their multiplication by scalars.

Example 2. Let V be a K-vector space. Consider the set VS ofall functions on a given set S with values in V. Elements of V canbe added and multiplied by scalars. Respectively the vector-valuedfunctions can be added and multiplied by scalars in the pointwisefashion. Thus, VS is an example of a K-vector space.

Example 3. A non-empty subset W in a vector space V is calleda linear subspace (or simply subspace) if linear combinationsλu + µv of vectors from W with arbitrary coefficients lie in W. Inparticular, for every u ∈ W, −u = (−1)u ∈ W, and 0 = 0u ∈ W.A subspace of a vector space satisfies the axioms of a vector spaceon its own (since the operations are the same as in V). Thus everysubspace of a K-vector space is an example of a K-vector space.

Example 3a. An m × n-matrix is called square, if m = n. Asquare matrix A is called diagonal respectively, upper-triangular,or lower-triangular) if aij = 0 whenever i 6= j (respectively, i >j, or i < j). Diagonal (respectively, upper-triangular, or lower-triangular) matrices form a subspace in the space of all n×n-matrices,and therefore provide an example of a vector space.

Example 3b. The set of all polynomials (say, in one variable),1

form a subspace in the space RR of all real-valued functions on thenumber line and therefore provide examples of real vector spaces.More generally, polynomials with coefficients in K (as well as suchpolynomials of degree not exceeding 7) form examples of K-vectorspaces.

Morphisms

The modern ideology requires that mathematical entities be orga-nized into categories. This means that in addition to specifying ob-jects of interest, one should also specify morphisms, i.e. maps be-tween them. The category of vector spaces is obtained by takingvector spaces for objects, and linear maps for morphisms.

By definition, a function A : V → W from a vector space V to avector space W is called a linear map if it respects the operationswith vectors, i.e. if it maps linear combinations of vectors to linear

1As well as sets of all continuous, differentiable, 5 times continuously differen-tiable, infinitely differentiable, Riemann-integrable, measurable, etc. functions,introduced in mathematical analysis.

Page 5: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 25

combinations of their images with the same coefficients:

A(λu+ µv) = λAu+ µAv for all u,v ∈ V and λ, µ ∈ K.

With a linear map A : V → W, one associates two subspaces, onein V, called the null space, or the kernel of A and denoted KerA,and the other in W, called the range of A and denoted A(V):

KerA := {v ∈ V : Av = 0},A(V) := {w ∈ W : w = Av for some v ∈ V}.

A linear map is injective, i.e. maps different vectors to differentones, exactly when its kernel is trivial. Indeed, if KerA 6= {0}, thenit contains non-zero vectors mapped to the same point 0 in W as 0from V. This makes the map A non-injective. Vice versa, if A isnon-injective, i.e. if Av = Av′ for some v 6= v′, then u = v− v′ 6= 0lies in KerA. This makes the kernel nontrivial.

When the range of a map is the whole target space, A(V) = W,the map is called surjective. If a linear map A : V → W is bi-jective, i.e. both injective and surjective, it establishes a one-to-onecorrespondence between V and W in a way that respects vector oper-ations. Then one says that A establishes an isomorphism betweenthe vector spaces V and W. Two vector spaces are called isomor-phic (written V ∼= W) if there exists an isomorphism between them.

Example 4. Let V = W be the space K[x] of all polynomialsin one indeterminate x with coefficients from K = R or C. Thedifferentiation d/dx : K[x] → K[x] is a linear map defined by

d

dx(a0 + a1x+ · · ·+ anx

n) = a1 + 2a1x+ · · · + nanxn−1.

It is surjective, with the kernel consisting of constant polynomials.

Example 5. Linear combinations λA+ µB of linear maps A,B :V → W are linear. Therefore all linear maps from V toW form a sub-space in the space of all vector-valued functions V → W. The vectorspace of linear maps from V to W is usually denoted byHom(V,W)(from the word homomorphism, synonimous in our context to theterm linear map.

Example 5a. A linear map from V to K is called a linear form(or linear function) on V. Linear forms on V form a subspace inthe space of all functions from V to K, and thus form a vector spaceon their own. This space is usually denoted V∗ and called the dualspace of V.

Page 6: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

26 Chapter 2. DRAMATIS PERSONAE

The following formal construction indicates that every vectorspace can be identified with a subspace in a space of functions withpointwise operations of addition and multiplication by scalars.

Example 5b. Given a vector v ∈ V and a linear function f ∈ V∗,the value f(v) ∈ K is defined. We can consider it not as a functionf of v, but as a function of f defined by v. This way, to a vectorv we associate the function Ev : V∗ → K defined by evaluating alllinear functions V → K on the vector v. The function Ev is linear,since (λf + µg)(v) = λf(v) + µg(v). The linear function Ev is anelement of the second dual space (V∗)∗. The formula f(λv + µw) =λf(v) + µf(w), expressing linearity of linear functions, shows thatEv depends linearly on v. Thus the evaluation map E : v 7→Ev is a linear map V → (V∗)∗. One can showthat E is injectiveand thus provides an isomorphism between V and its rangeE(V) ⊂ (V∗)∗.

The previous result and examples suggest that vector spaces neednot be described abstractly, and raises the suspicion that the ax-iomatic definition is misleading as it obscures the actual nature ofvectors as functions subject to the pointwise algebraic operations.Here are however some examples where vectors do not come natu-rally as functions.

Perhaps, the most important example of this kind is given bygeometric vectors, (as well as by forces and velocities in physics). Itprovides the opportunity to use geometric intuition in the contextunrelated to geometry. That is, one can “visualize” functions as geo-metric vectors. Furthermore, taking the field Q of rational numbers,or the field Z2 = {0, 1} on the role of scalars, one can apply geometricintuition to number theory or computer science. Later we will havesee how this works.

(v,w) vv’

W0

Figure 23Figure 22

v

w

Another justification for introducing vector spaces abstractly isthat this approach provides great flexibility in constructing new vec-tor spaces from given ones. Such constructions are used regularly,and it would be very awkward to constantly express the resultingvector spaces as spaces of functions, even when the given spaces areexpressed this way. Here are two of such constructions.

Page 7: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 27

Direct Sums and Quotients

Given two vector spaces V and W, their direct sum V ⊕W (Figure22) is defined as the set of all ordered pairs (v,w), where v ∈ V,w ∈ W, equipped with the component-wise operations:

λ(v,w) = (λv, λw), (v,w) + (v′,w′) = (v + v′,w +w′).

Of course, one can similarly define the direct sum of several vectorspaces. E.g. Kn = K⊕ · · · ⊕K (n times).

Example 7. Given a linear map A : V → W, its graph is definedas a subspace in the direct sum V ⊕W:

Graph A = {(v,w) ∈ V ⊕W : w = Av}.

The quotient space of a vector space V by a subspace W isdefined as follows. Two vectors v and v′ (Figure 23) are calledequivalent modulo W, if v − v′ ∈ W. This way, all vectors fromV become partitioned into equivalence classes. These equivalenceclasses form the quotient vector space V/W.

In more detail, denote by π : V → V/W the canonical projec-tion, which assigns to a vector v its equivalence class modulo W.This class can be symbolically written as v+W, a notation empha-sizing that the class consists of all vectors obtained from v by addingarbitrary vectors from W. Alternatively, one may think of v + Was a “plane” obtained from W as translation by the vector v. Whenv ∈ W, we have v +W = W. When v /∈ W, v +W is not a linearsubspace in V. We will call it an affine subspace parallel to W.

The set V/W of all affine subspaces in V parallel toW is equippedwith algebraic operations of addition and multiplication by scalarsin such a way that the canonical projection π : V → V/W becomes alinear map. In fact this condition leaves no choices, since it requiresthat for every u,v ∈ V and λ, µ ∈ K,

λπ(u) + µπ(v) = π(λu+ µv).

In other words, the linear combination of given equivalence classesmust coincide with the equivalence class containing the linear com-bination λu+µv of arbitrary representatives u,v of these classes. Itis important here that picking different representatives u′ and v′ willresult in a new linear combination λu′+µv′ which is however equiv-alent to the previous one. Indeed, the difference λ(u−u′)+µ(v−v′)

Page 8: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

28 Chapter 2. DRAMATIS PERSONAE

lies in W since u − u′ and v − v′ do. Thus linear combinations inV/W are well-defined.

The construction of the quotient space is admittedly one of themost abstract ones so far. Here is a hint to how one could think ofelements of V/W and the projection π.

Example 8. Projecting 3D images to a 2-dimensional screen isdescribed in geometry by the canonical projection π from the 3Dspace V to the plane V/W of the screen along the line W of the eyesight (Figure 24).

W0 0

V/WV

Figure 24

Example 9. The direct sum V ⊕ W contains V and W as sub-spaces consisting of the pairs (v,0) and (0,w) respectively. The quo-tient of V⊕W byW is canonically identified with V, because each pair(v,w) is equivalent moduloW to (v,0). Likewise, (V ⊕W) /V = W.

Example 10. Let V = R[x] be the space of polynomials with realcoefficients, and W the subspace of polynomials divisible by x2 + 1.Then the quotient space V/W can be identified with the plane C

of complex numbers, and the projection π : R[x] → C with the mapP 7→ P (i) of evaluating a polynomial P at x = i. Indeed, polynomialsP and P ′ are equivalent modulo W if and only if P − P ′ is divisibleby x2 + 1, in which case P (i) = P ′(i). Vice versa, if P (i) = P ′(i),then P (−i) = P ′(−i) (since the polynomials are real), and henceP − P ′ is divisible by (x− i)(x + i) = x2 + 1.

For every linear map A : V → V ′, there is a canonical isomor-phism A : V/KerA → A(V) between the quotient by the kernel ofA, and its range. Namely, Au = Av if and only if u − v ∈ KerA,i.e. whenever u is equivalent to v modulo the kernel. Thus, onecan think of every linear map as the projection of the source space

Page 9: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 29

onto the range along the null space. This is a manifestation of a gen-eral homomorphism theorem in algebra, which in the context ofvector spaces can be formally stated this way:

Theorem. Every linear map A : V → V ′ is uniquely repre-sented as the composition A = iAπ of the canonical projec-tion π : V → V/KerA with the isomorphism A : V/KerA →A(V) followed by the inclusion i : A(V) ⊂ V ′:

V A−→ V ′

π ↓ ∪ i

V/KerA∼=−→ A(V)A

.

This result, although called a theorem, is merely a rephrasing ofthe definitions (of vector spaces, subspaces, quotient spaces, linearmaps, isomorphisms, etc.) and is in this sense tautological,2 void ofnew knowledge. The results of the next subsection are quite differentin this regard.

Bases and Dimension

Let V be a K-vector space. A subset V ⊂ V (finite or infinite) iscalled a basis of V if every vector of V can be uniquely written as a(finite!) linear combination of vectors from V .

Example 11. Monomials xk, k = 0, 1, 2, . . . , form a basis in thespace K[x] since every polynomial is uniquely written as a linearcombination of monomials.

Example 12. In Kn, every vector is uniquely written as the linearcombination of unit coordinate vectors e1, . . . , en:

x1x2...xn

= x1

10...0

+ · · ·+ xn

0...01

.

Thus, the vectors e1, . . . , en form a basis. It is called the standardbasis of Kn.

2Dictionaries define tautology as “a representation of anything as the cause,condition, or consequence of itself.”

Page 10: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

30 Chapter 2. DRAMATIS PERSONAE

The notion of basis has two aspects which can be consideredseparately.

Let V ⊂ V be any set of vectors. Linear combinations λ1v1+· · ·+λkvk, where the vectors vi are taken from the subset V , and λi arearbitrary scalars, form a subspace in V. (Indeed, sums and scalarmultiples of linear combinations of vectors from V are also linearcombinations of vectors from V .) This subspace is often denoted asSpanV . One says that the set V spans the subspace SpanV , or thatSpanV is spanned by V .

A set V of vectors is called linearly independent, if no vectorfrom SpanV can be represented as a linear combination of vectorsfrom V in more than one way. To familiarize ourselves with thisnotion, let us give several reformulations of the definition. Here isone: no two distinct linear combinations of vectors from V are equalto each other. Yet another one: if two linear combinations of vectorsfrom V are equal: α1v1 + · · · + αkvk = β1v1 + · + βkvk, then theircoefficients must be the same: α1 = β1, . . . , αk = βk. Subtractingone linear combination from the other, we arrive at one more refor-mulation: if γ1v1 + · · · + γkvk = 0 for some vectors vi ∈ V , thennecessarily γ1 = · · · = γk = 0. In other words, V is linearly inde-pendent, if the vector 0 can be represented as a linear combinationof vectors from V only in the trivial fashion: 0 = 0v1 + · · · + 0vk.Equivalently, every nontrivial linear combination of vectors fromV is not equal to zero: γ1v1 + · · ·+ γkvk 6= 0 whenever at least oneof γi 6= 0.

Of course, a set V ⊂ V is called linearly dependent if it is notlinearly independent. Yet, it is useful to have an affirmative reformu-lation: V is linearly dependent if and only if some nontrivial linearcombination of vectors from V vanishes, i.e. γ1v1 + · · · + γkvk = 0,where at least one of the coefficients (say, γk) is non-zero. Dividingby this coefficient, and moving all other terms to the other side ofthe equality, we obtain one more reformulation: a set V is linearlydependent if one of its vectors can be represented as a linear combina-tion of the others: vk = −γ−1

k γ1v1−· · ·− γ−1k γk−1vk−1.

3 Obviously,every set containing the vector 0 is linearly dependent; every set con-taining two proportional vectors is linearly dependent; adding newvectors to a linearly dependent set leaves it linearly dependent.

Thus, a basis of V is a linearly independent set of vectors thatspans the whole space.

3It is essential here that division by all non-zero scalars is well-defined.

Page 11: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 31

In this course, we will be primarily concerned with finite dimen-sional vector spaces, i.e. spaces which can be spanned by finitelymany vectors. If such a set of vectors is linearly dependent, thenone of its vectors is a linear combination of the others. Removingthis vector from the set, we obtain a smaller set that still spans V.Continuing this way, we arrive at a finite linearly independent setthat spans V. Thus, a finite dimensional vector space has a basis,consisting of finitely many elements. The number of elements in abasis does not depend (as we will see shortly) on the choice of thebasis. This number is called the dimension of the vector space Vand is denoted dimV.

Let v1, . . . ,vn be a basis of V. Then every vector x ∈ V isuniquely written as x = x1v1 + · · · + xnvn. We call (x1, . . . , xn)coordinates of the vector x with respect to the basis v1, . . . ,vn.For y = y1v1 + · · ·+ ynvn ∈ V and λ ∈ K, we have:

x+ y = (x1 + y1)v1 + · · ·+ (xn + yn)vn,

λx = (λx1)v1 + · · ·+ (λnxn)vn.

This means that operations of addition of vectors and multiplicationby scalars are performed coordinate-wise. In other words, the map:

Kn → V : x1e1 + · · ·+ xnen 7→ x1v1 + · · · + xnvn

defines an isomorphism of the coordinate space Kn onto thevector space V with a basis {v1, . . . ,vn}.

Lemma. A set of n+1 vectors in Kn is linearly dependent.

Proof. Any two vectors in K1 are proportional and thereforelinearly dependent. We intend to prove the lemma by deducing fromthis that any 3 vectors in K2 are linearly dependent, then deducingfrom this that any 4 vectors in K3 are linearly dependent, and so on.Thus we only need to prove that if every set of n vectors in Kn−1 islinearly dependent then every set of n + 1 vectors in Kn is linearlydependent too.4

To this end, consider n + 1 column vectors v1, ...,vn+1 of size neach. If the bottom entry in each column is 0, then v1, ...,vn+1 areeffectively n− 1-columns. Hence some nontrivial linear combination

4This way of reasoning is called mathematical induction. Put abstractly,it establishes a sequence Pn of propositions in two stages called respectively thebase and step of induction: (i) P1 is true; (ii) for all n = 2, 3, 4, . . . , if Pn−1 istrue (the induction hypothesis) then Pn is true.

Page 12: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

32 Chapter 2. DRAMATIS PERSONAE

of v1, ...,vn is equal to 0 (by the induction hypothesis), and thus thewhole set is linearly dependent. Now consider the case when at leastone column has the bottom entry non-zero. Reordering the vectorswe may assume that it is the column vn+1. Subtracting the columnvn+1 with suitable coefficients α1, ..., αn from v1, ...,vn we can formn new columns u1 = v1 − α1vn+1, ...,un = vn − αnvn+1 so thatall of them have the bottom entries equal to zero. Thus u1, ...,un

are effectively n − 1-vectors and are therefore linearly dependent:β1u1 + ...+ βnun = 0 for some β1, ..., βn not all equal to 0. Thus

β1v1 + ...+ βnvn − (α1β1 + ...+ αnβn)vn+1 = 0.

Here at least one of βi 6= 0, and hence v1, ...,vn+1 are linearly de-pendent. �

Corollaries. (1) Any set of m > n vectors in Kn is lin-early dependent.

(2) Kn and Km are not isomorphic unless n = m.

(3) Every finite dimensional vector space is isomorphicto exactly one of the spaces Kn.

(4) In a finite dimensional vector space, all bases havethe same number of elements. In particular, dimension iswell-defined.

(5) Two finite dimensional vector spaces are isomorphicif and only if their dimensions are equal.

Indeed, (1) is obvious because adding new vectors to a linearlydependent set leaves it linearly dependent. Since the standard basisin Km consists of m linearly independent vectors, Km cannot beisomorphic to Kn if m > n. This implies (2) and hence (3), becausetwo spaces isomorphic to a third one are isomorphic to each other.Now (4) follows, since the choice of a basis of n elements establishesan isomorphism of the space with Kn. Rephrasing (3) in terms ofdimensions yields (5).

Example 13. Let V be a vector space of dimension n, and letv1, . . . ,vn be a basis of it. Then every vector x ∈ V can be uniquelywritten as x = x1v1 + · · ·+ xnvn. Here x1, . . . , xn can be consideredas linear functions from V to K. Namely, the function xi takes thevalue 1 on the vector vi and the value 0 on all vj with j 6= i. Everylinear function f : V → K takes on a vector x the value f(x) =x1f(v1) + · · · + xnf(vn). Therefore f is the linear combination ofx1, . . . , xn with the coefficients f(v1), . . . , f(vn), i.e. x1, . . . , xn spanthe dual space V∗. In fact, they are linearly independent, and thus

Page 13: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 33

form a basis of V∗. Indeed, if a linear combination γ1x1 + · · ·+ γnxncoincides with the identically zero function, then its values γi on thevectors vi must be all zeroes. We conclude that the dual space V∗

has the same dimension n as V and is isomorphic to it. Thebasis x1, . . . , xn is called the dual basis of V∗ with respect to thebasis v1, . . . ,vn of V.

Remark. Corollaries 3, 5, and Example 3 suggest that in a sensethere is “only one” K-vector space in each dimension n = 0, 1, 2, . . . ,namely Kn. The role of this fact, which is literally true if the unique-ness is understood up to isomorphism, should not be overestimated.An isomorphism Kn → V is determined by the choice of a basis inV, and is therefore not unique. For example, the space of polyno-mials of degree < n in one indeterminate x has dimension n and isisomorphic to Kn. However, different isomorphisms may be usefulfor different purposes. In elementary algebra one would use the basis1, x, x2, . . . , xn−1. In Calculus 1, x, x2/2, . . . , xn−1/(n − 1)! may bemore common. In the theory of interpolation the basis of Lagrangepolynomials is used:

Li(x) =

j 6=i(x− xj)∏

j 6=i(xi − xj).

Here x1, . . . , xn are given distinct points on the number line, andLi(xj) = 0 for j 6= i and Li(xi) = 1. The theory of orthogonalpolynomials leads to many other important bases, e.g. those formedby Chebyshev polynomials5 Tk, or Hermite polynomials Hk:

Tk(x) = cos(

k cos−1(x))

, Hk(x) = ex2 dk

dxke−x2

.

There is no preferred basis in an n-dimensional vector space V (andhence no preferred isomorphism between V and Kn).

The lack of the preferred isomorphism becomes really importantwhen continuous families of vector spaces get involved. For instance,consider on the plane R2, all subspaces of dimension 1 (Figure 25).When subspaces rotate, one can pick a basis vector in each of them,which would vary continuously, but when the angle of rotation ap-proaches π, the direction of the vector disagrees with the initial di-rection of the same line. In fact it is impossible to choose bases inall the lines in a continuous fashion. The reason is shown on Figure

5After Pafnuty Chebyshev (1821– 1984).

Page 14: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

34 Chapter 2. DRAMATIS PERSONAE

26: The surface formed by the continuous family of 1-dimensionalsubspaces in R2 has the topology of a Mobius band (rather thana cylinder). The Mobius band is a first example of nontrivial vec-tor bundles. Vector bundles are studied in Homotopic Topology. Itturns out that among all k-dimensional vector bundles (i.e. continu-ous families of k-dimensional vector spaces) the most complicated arethe bundles formed by all k-dimensional subspaces in the coordinatespace of dimension n ≫ k.

Figure 26Figure 25

Corollary 6. A finite dimensional space V is canonicallyisomorphic to its second dual V∗∗.

Here “canonically” means that there is a preferred isomorphism.Namely, the isomorphism is established by the map E : V → V∗∗

from Example 5b. Recall that to a vector v ∈ V, it assigns the linearfunction Ev : V ∗ → K, defined by evaluation of linear functions onthe vector v. The kernel of this map is trivial (e.g. because one canpoint a linear function that takes a non-zero value on a given non-zero vector). The range E(V) must be a subspace in V∗∗ isomorphicto V. But dimV∗∗ = dimV∗ = dimV. Thus the range must be thewhole space V∗∗.

EXERCISES

52. Prove that in a vector space, the zero vector is unique. �

53. Prove that the opposite of each vector is unique. �

54. Give an example of a “vector space” that satisfies all axioms exceptthe last one: 1u = u for all u. �

55. Prove that the axiom: 0u = 0 for all u, in the definition of vectorspaces is redundant, i.e. can be derived from the remaining axioms. �

56. Verify that KS and VS are vector spaces.

57. Show that in K[x], polynomials of degree n do not form a subspace,but polynomials of degree ≤ n do.

58. Prove that intersection of subspaces is a subspace.

Page 15: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

1. Vector Spaces 35

59. Show that the union of two subspaces is a subspace if and only if oneof the subspaces contains the other.

60. Show that a map A : V → W is linear if and only if A(u+v) = Au+Avand A(λu) = λAu for all u,v in V and all λ ∈ K.

61. Prove that rotation about the origin is a linear map from the plane toitself. �

62. Show that every linear form on the coordinate (x, y)-plane is a linearcombination of the coordinates: αx+ βy.

63. Show that the map f 7→∫ b

af(x) dx defined by integration of (say)

polynomial functions is a linear form R[x] → R.

64. For V = K2, verify that the map E : V → V∗∗ defined in Example 5bis an isomorphism.

65. Establish an isomorphism between Km ⊕Kn and Km+n.

66. Show that (V ⊕W)∗ = V∗ ⊕W∗. �

67. Prove that for a linear map A : V → W , KerA = (GraphA) ∩ V .68. Describe all affine subspaces in geometric 3D space. �

69.⋆ Prove that the intersection of two affine subspaces, parallel to givenlinear ones, if non-empty, is an affine subspace parallel to the intersectionof the given linear subspaces. �

70.⋆ Let V = R[x]], and W ⊂ V be the subspace of all polynomials divisibleby x2 − 1. Establish an isomorphism between V/W and R{1,−1}, the spaceof all functions from {1,−1} to R. �

71. Describe all bases in K1.

72. Prove that the set consisting of one vector is linearly dependent if andonly if the vector is 0.

73. Prove that a subset of a linearly independent set of vectors is linearlyindependent.

74. Prove that a set of vectors containing a linearly dependent subset islinearly dependent.

75. Prove that the set of four polynomials x2, (x− 1)2, (x− 2)2, (x− 3)2 islinearly dependent, but any proper subset of it is linearly independent.

76. Find a basis of the subspace inR3 given by the equation x1+x2+x3 = 0.

77. The same, for the subspace in C4 given by the equation x1 + x2 + x3+x4 = 0.

78. Find a basis in and the dimension of the subspace in K4 given by twoequations: x1 + x2 + x3 = 0 and x2 + x3 + x4 = 0. �

79. Find the dimensions of spaces of: (a) diagonal n × n-matrices; (b)upper-triangular n× n-matrices. �

Page 16: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

36 Chapter 2. DRAMATIS PERSONAE

80. Find the dimension of the subspace in RR spanned by functionscos(x+ θ1), . . . , cos(x+ θn), where θ1, . . . , θn are given distinct angles. �

81.⋆ In the space of polynomials of degree < n, express the basis xk, k =0, 1, . . . , n−1 of monomials in terms of the basis Li, i = 1, . . . , n, of Lagrangepolynomials. �

82. In R[x], find coordinates of the Chebyshev polynomials T4 in the basisof monomials xk, k = 0, 1, 2, . . . �

83. For a finite-dimensional vector space V , show that the canonical mapE : V → V∗∗ (defined by evaluation of linear forms on vectors) is injective. �

Page 17: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 37

2 Matrices

Matrices are rectangular arrays of numbers. An m× n-matrix

A =

a11 . . . a1n. . . aij . . .am1 . . . amn

has m rows and n columns. The matrix entry aij is positioned inrow i and column j.

In linear algebra, matrices are found all over the place. Yet, themain point of this section is that matrices per se are not objectsof linear algebra. Namely, the same matrix can represent differentmathematical objects, and may behave differently depending on whatkind of object is meant.

More precisely, various geometric objects, such as vectors, linearfunctions, quadratic forms, linear maps, etc., when expressed in co-ordinates, are represented by matrices. It is one of our goals in thisbook to demonstrate advantages of the geometric (i.e. “conceptual,”coordinate-less) way of thinking over the algebraic (i.e. “computa-tional,”, coordinate-wise) attitude. So, we begin here with going asfar as possible in the oppositite direction by examining the matrix,coordinate expressions of all geometric objects of our interest. When-ever suitable, we derive properties of operations with matrices fromthe properties of the geometric objects they represent (and not theother way around).

Vectors and Linear Functions

Let V be a finite dimensional vector space. Picking a basis {e1, . . . , en}identifies each vectors x = x1e1 + · · · + xnen with the n× 1-matrix,the column of its coordinates x1, . . . , xn. As we saw in the previoussection, the results of vector operations: the multiplication by scalarsλx and the sum x + y, are expressed by componentwise operationswith the columns:

λ

x1...xn

=

λx1...

λxn

,

x1...xn

+

y1...yn

=

x1 + y1...

xn + yn

.

A linear function (or linear form) a : V → K is determinedby its values on the basis vectors:

a(x) = x1a(e1) + · · · xna(en) = a1x1 + · · ·+ anxn, ai := a(ei).

Page 18: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

38 Chapter 2. DRAMATIS PERSONAE

Conversely, each function of the form a1x1 + · · · + anxn is linear(for it is a linear combination of the coordinate functions xi, whichare linear). Linear functions are traditionally represented by 1 × n-matrices, the rows of their coefficients: [a1, . . . , an].

Linear combinations λa+ µb of linear functions are linear func-tions. Their coefficients are expressed by linear combinations of therows:

λ[a1, . . . , an] + µ[b1, . . . , bn] = [λa1 + µb1, . . . , λan + µbn].

The operation of evaluation, i.e. taking the value a(x) of a linearfunction on a vector, is expressed in coordinates by the simplestinstance of matrix product, namely the product of a 1 × n rowwith an n× 1 column:

a(x) = [a1, . . . , an]

x1. . .xn

= a1x1 + · · ·+ anxn.

Note that by the very definition of linear functions,

a(λx+ µy) = λa(x) + µa(y) for all vectors x,y and scalars λ, µ.

This can be viewed as a manifestation of the distributive law formatrix product.

Linear Maps

Let A : V → W be a linear map between two finite dimensional vectorspaces. Picking a basis {e1, . . . , en} in V and a basis {f1, . . . , fm} inW, one can express Aej as linear combinations of fi:

Aej = a1jf1 + · · ·+ amjfm, j = 1, . . . , n.

The whole map A is uniquely determined by the coefficients aij ;namely, if x = x1e1 + · · · + xnen, then by the very definition oflinearity, we have:

Ax = x1Ae1 + · · ·+ xnAen = x1

a11...

am1

+ · · ·+ xn

a1n...

amn

.

In other words, the map y = Ax is given by m linear functions

yi = ai(x) = a11x1 + · · ·+ ainxn, i = 1, . . . ,m.

Page 19: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 39

Whereas each yi is given by the product of a row with a column, thewhole linear map can be described as the product of them×n-matrixA = [aij ] with the column:

y =

y1...ym

=

a11 . . . a1n...

...am1 . . . amn

x1...xn

= Ax.

Thus, the rows of the matrix A represent the m linear functionsa1, . . . ,am, the columns represent the images Ae1, . . . , Aen in W ofthe basis vectors of V, and the matrix entries aij = ai(ej) the valuesof the linear functions on the basis vectors.

Conversely, in coordinates, the rows of any m × n-matrix deter-mine m linear functions yi = ai(x), which altogether determine amap from V = Kn → Km = W, which is automatically linear (sincethe functions ai are).

Composition

Given linear maps B : U → V and A : V → W, one can formtheir composition C : U → W by substituting v = Bu into Av:Cu := A(B(u)). When the spaces are finite dimensional, dimU = l,dimV = n, dimW = m, one can pick their bases (thereby identify-ing the spaces with Kl,Kn, and Km respectively), and express theoperation in terms of the corresponding matrices. This leads to thegeneral notion of matrix product, C = AB, defined whenever thenumber of columns of A coincides with the number of rows of B:

c11 . . . c1l. . . cij . . .cm1 . . . cml

=

a11 . . . . a1n. . . . . .am1 . . . . amn

b11 . . . b1l. . . . . .. . . . . .bn1 . . . bnl

.

By definition of the composition, the entry cij located at the inter-section of the ith row and jth column of C is the value of the linearfunction ai on the image Bej in Kn of the standard basis vector ejfrom Kl. Since ai and Bej are represented by the ith row and jthcolumn respectively of the matrices A and B, we find:

cij = [ai1, . . . , ain]

b1j. . .bnj

= ai1b1j + · · · + ainbnj.

Page 20: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

40 Chapter 2. DRAMATIS PERSONAE

In other words, cij is the product of the ith row of A with the jthcolumn of B.

Based on this formula, it is not hard to verify that the matrixproduct is associative, i.e. (AB)C = A(BC), and satisfies theleft and right distributive laws: P (λQ + µR) = λPQ + µPR and(λX+µY )Z = λXZ+µY Z, whenever the sizes of the matrices allowto form the expressions. However, our point is that there is no pointin making such verifications. The operations with matrices encode inthe coordinate form meaningful operations with linear maps, and theproperties of matrices simply reflect those of the maps. For instance,the matrix product is associative because composition of arbitrary(and not only linear) maps is associative.

Changes of Coordinates

Let us examine here how the components of vectors, the coefficientsof linear forms, and the matrices of linear maps are transformedunder linear changes of coordinates.

Suppose that a finite dimensional vector space V is identified withKn by the choice of a basis {e1, . . . , en}, and then identified with itin another way by the choice of a new basis, {e′1, . . . , e′n}. We canexpress the vectors of the new basis as linear combinations of the oldbasis vectors, and substitute these expressions into two coordinaterepresentations of the same vector in the old and new coordinatesystems:

x = x1e1 + · · ·+ xnen = x′1e′1 + · · ·+ x′ne

′n.

This will result in expressing the old coordinate functions xi in termsof the new coordinate functions x′j :

x1. . .xn

=

c11 . . . c1n. . . . . .cn1 . . . cnn

x′1. . .x′n

.

In matrix notation, this can be written as x = Cx′ where C is asquare matrix of size n.

Cnversely, x′ can be expressed by linear functions of x. In otherwords, there exists a square matrix D such that the substitutionsx′ = Dx and x = Cx′ are inverse to each other, i.e. x = CDx andx′ = DCx′ for all x and x′. It is immediate to see that this happens

Page 21: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 41

exactly when the matrices C and D satisfy CD = I = DC where Iis the identity matrix of size n:

I =

1 0 . . . 00 1 . . . 0. . . . . .0 . . . 0 1

.

When this happens, the square matrices C and D are called inverseto each other, and one writes: D = C−1, C = D−1.

To reiterate: the rows of C express old coordinates xi as linearfunctions of new coordinates, x′j. The columns of C represent, inthe old coordinate system, the vectors which in the new coordinatesystem serve as standard basis vectors: e′j = c1je1 + · · · + cnjen.The matrix C is often called the transition matrix between thecoordinate systems.

In spite of apparent simplicity of this notion, it is easy to getlost here, as a change of coordinates is easy to confuse with a linearmap from the space Kn to itself. For linear maps from a vector spaceto itself, we will reserve the term linear transformation. Thus,the same square matrix C defines the linear transformation x′ = Cxwhich associates to a vector x ∈ Kn a new vector x′ written inthe same coordinate system. The inverse transformation, whenexists, is given by the formula x = C−1x′.

Returning to changes of coordinates, examine how they affectlinear functions and maps.

Making the change of variables x = Cx′ in a linear function aresults in the values a(x) of this function being expressed as a′(x′)in terms of new coordinates of the same vectors. Using the matrixproduct notation, we find: a′x′ = ax = a(Cx′) = (aC)x′. Thus,coordinates of vectors and coefficients of linear functions are trans-formed differently:

x = Cx′, or x′ = C−1x, but a′ = aC, or a = a′C−1.

Next, let y = Ax be a linear map from Kn to Km, and let x = Cx′

and y = Dy′ be changes of coordinates in Kn and Km respectively.Then in new coordinates the same linear map is given by the newformula y′ = A′x′. We compute A′ in terms of A, C and D: Dy′ =y = Ax = ACx′, i.e. y′ = D−1ACx′, and hence

A′ = D−1AC.

Page 22: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

42 Chapter 2. DRAMATIS PERSONAE

In particular, suppose that x 7→ Ax is a linear transformation on Kn

(i.e. — we remind — a linear map from Kn to itself), and a changeof coordinates x = Cx′ is made. Then in new coordinates the samelinear transformation is x′ 7→ A′x′, where

A′ = C−1AC,

i.e. in the the previous rule we need to take D = C. This is becausethe same change applies to both: the input vector x and its imageAx. The operation A 7→ C−1AC over a square matrix A is oftencalled the similarity transformation by the invertible matrix C.

Transposition

Let A : V → W be a linear map, and a : W → K a linear function.Composing the linear fucntion with the map, we obtain a linear func-tion on V. This construction defines a map from W ∗, the space oflinear finctions on W, to V ∗. The map is called dual (or adjoint, ortransposed) to A, and denoted At. Thus At : W∗ → V∗ is definedby the formula

(Ata)(v) := a(Av).

It is a linear map. Indeed, if c = λa+ µb, then c(Av) = λa(Av) +µb(Av) for all v, and hence Atc = λAta+ µAtb. Note that At actsin the direction “opposite to A”.

In coordinates, if b = Ata, then

[b1, . . . , bn] = [a1, . . . , am]

a11 . . . a1n. . . aij . . .am1 . . . amn

,

that is, the row of the coefficients of the composite linear functionb(x) = a(Ax) is obtained from the row of the coefffiients of a by themultiplication on the right by the matrix A. This is a special case ofthe matrix expression for composition of linear maps. If however, wethink of elements of W∗ as “vectors,” i.e. represent them by columnsof their coordinates, then we have no other choice but to flip all therows into columns here:

b1...bn

=

a11 . . . am1... aji

...a1n . . . anm

a1...am

.

Page 23: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 43

Note that the matrix [aji] is not A but is the matrix At transposedto A. By definition, this means that it is an n × m-matrix, whoseentry atij on the intersection of row i and column j coincides with theentry aji of the m × n-matrix A, located at the intersection of therow j and column i. Thus, the equality can be written as bt = Atat.

Given an m× n-matrix A and n× l-matrix B, we have

(AB)t = BtAt.

One can check this matrix identity directly, but in fact there is noneed to do so. Indeed, given three vector spaces and two linear maps:B : U → V, and A : V → W, the adjoint map to their compositionAB : U → W acts on linear forms W → K as the composition:At : W∗ → V∗, followed by Bt : V∗ → U∗. Indeed, for all a ∈ W∗

and all u ∈ U ,

((AB)ta)(u) := a(ABu) = (Ata)(Bu) = (Bt(Ata))(u).

Bilinear Forms

The dot-product of two coordinate vectors x,y ∈ Rn is defined bythe formula

〈x,y〉 = x1y1 + · · ·+ xnyn.

It is an example of a bilinear form, i.e. a function of an orderedpair (x,y) of vector variables, which is a linear function of each ofthem.

In general, a bilinear form is a function B : V ×W → K, whichto each ordered pair (v,w), where v ∈ V and w ∈ W, associatesa scalar B(v,w) in such a way that for a fixed w this is a linearfunction of v, and for a fixed v a linear function of w:

B(λv + λ′v′, w) = λB(v,w) + λ′B(v′,w),

B(v, λw + λ′w′) = λB(v,w) + λ′B(v,w′)

for all v,v′ ∈ V, w,w′ ∈ W, and λ, λ′ ∈ K.

Assuming that V and W are finite dimensional vector spaces, onecan pick bases {e1 . . . , em} in V and {f1, . . . , fn} in W, and writingv = x1e1 + · · ·+ xmem, w = y1f1 + · · ·+ ynfn, find:

B(v,w) =

m∑

i=1

n∑

j=1

xiB(ei, fj)yj .

Page 24: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

44 Chapter 2. DRAMATIS PERSONAE

Thus, a bilinear form is uniquely determined by the m×n-matrix ofthe coefficients B(ei, fj). Conversely, every m × n-matrix B = [bij ]defines a bilinear form of two coordinate vectors, x ∈ Km and y ∈ Kn,by the formula:

xtBy =

m∑

i=1

n∑

j=1

xibijyj.

It is easy to see now how the coefficient matrix of a bilinearform behaves under changes of coordinate systems. If y = Cy′, andx = Dx′, where C and D are invertible n× n- and m×m-matrices,we have: xt = (x′)tDt and xtBy = (x′)tDtBCy′ = (x′)tB′y′. Thatis, the coefficient matrix B is transformed into6

B′ = DtBC.

With a bilinear form B : V × W → K, one can associate thetransposed bilinear form Bt : W × V → K by simply changingthe order of the arguments: Bt(w,v) := B(v,w). In coordinates,Bt(fi, ej) = B(ej, fi), that is the coefficient matrix of Bt is transposedto that of B: btij = bji.

To a bilinear form B : V × W → K, one can associate a linearmap (denoted by the same letter) B : W → V∗ from W to the dual ofV. Namely, to a vector w ∈ W, it associates the linear function on Vwhich on each vector v ∈ V assumes the value B(v,w). Conversely,given a linear map B : W → V∗ it defines a bilinear form, whosevalue on the pair (v,w) is equal to the value of the linear functionBw ∈ V∗ on the vector v ∈ V.

Choosing coordinate systems v =∑

xiei in V, and w =∑

j yjfjin W, one can describe the map B : W → V∗ by its matrix withrespect to the basis f1, . . . , fn in W, and x1, . . . , xm in V∗. We leaveit as an exerciste for the reader to check that [bij ] is the matrix(i.e. bij = B(ei, fj)). The map Bt, adjoint to B : W → V∗, acts fromV∗∗ = V to W∗, and of course, corresponds to the transposed bilinearform, and is desribed in coordinates by the transposed matrix.

In paractice, we will be interested in the case when V = W, i.e.in bilinear forms of two vector arguments from the same space. A bi-linear form B : V ×V → K is called symmetric (anti-symmetric),if B = Bt (B = −Bt), or more explicitly: B(v,w) = B(w,v) andB(u,v) = −B(v,u) for all vectors u,v from V. Every bilinear form

6Note that if B were the matrix of alinear map Kn → Km, the transformationrule would heve been different: B 7→ D−1BC.

Page 25: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 45

on a vector space V can be uniquely written as the sum of a sym-metric and anti-symmetric form:7

B(u,v) =B(u,v) +B(v,u)

2+

B(u,v) −B(v,u)

2.

Using a basis {e1, . . . , en} in V (and the same basis fj = ej inW = V), one obtains the coefficient matrix [bij] of the bilinear form,which is a square matrix with the entries bij = B(ei, ej). The matrixof a bilinear form is symmetric, i.e. bij = bij , or anti-symmetric,i.e. bij = −bji, whenever the bilinear form is.

Under a change of coordinates, x = Cx′, y = Cy′, the coefficientmatrix B is transformed into

B′ = CtBC,

and remains symmetric (anti-symmetric), whenever B is.

Quadratic Forms

From a bilinear form B : V ×V → K, one obtains a quadratic formQB : V → K by “restricting to the diagonal” V ⊂ V × V:

QB(v) := B(v,v).

IfB is written as the sumB = S+A of symmetric and anti-symmetricbilinear forms, we find that A(v,v) = −A(v,v) = 0, and so QB(v) =QS(v) = S(v,v).

A symmetric bilinear form can be reconstructed from the corre-sponding quadratic form:

S(u,v) =S(u+ v,u+ v) − S(u,u) − S(v,v)

2.

In fact, one can define a quadratic form as a function Q : V → K

which is homogeneous of degree 2 (i.e. Q(λv) = λ2Q(v)), and suchthat S(u,v) := (Q(u+ v)−Q(u)−Q(v)) /2 is bilinear.

In coordinates, any n × n-matrix Q = [qij ], which is symmet-ric, defines a symmetric bilinear form xtQy and the correspondingquadratic form

Q(x) = xtQx =n∑

i=1

n∑

j=1

xiqijxj .

7Note that division by 2 is used here, which will not work over the field con-taining Z2 = {0, 1}, where 1 + 1 = 0.

Page 26: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

46 Chapter 2. DRAMATIS PERSONAE

Hermitian Forms

Here we need to assume that K = C, the field of complex numbers,since the operation λ 7→ λ of complex conjugation of scalars will beinviolved.

A sesquilinear form on a complex vector space V is defined as afunction T : V ×V → C of two vector argumens which is anti-linear(or half-linear) in the first argument and linear in the second. Bydefinition, this means that for all λ, µ ∈ C and u,v,w ∈ V

T (λu+ µv,w) = λT (u,w) + µT (v,w),

T (w, λu+ µv) = λT (w,u) + µT (w,v).

In coordinates, let v = z1e1+ · · ·+ znen and w = w1e1+ · · ·+wnen,where zi, wj ∈ C. Then

T (v,w) =

n∑

i=1

n∑

j=1

zi tijwj , where tij = T (ei, ej).

Conversely, given an arbitrary complex n× n-matrix [tij ], the aboveformula defines a sesquilinear form on the coordinate space V = Cn.

Given a sesquilinear form T , its Hermitian adjoint8 (or simplyadjoint form T † is defined by

T †(u,v) := T (v,u) for all u,v ∈ V.Note that it is also sesquilinear, that is, anti-linear with respect to uand linear with respect to v. In coordinates, the coefficient matrix

[t†ij ] of the adjoint form and the matrix [tij ] are Hermitian conju-gate, i.e., are obtained from each other by the operations of complexconjugation of all entries and transposition. Indeed,

t†ij = T †(ei, ej) = T (ej, ei) = tji.

A sesquilinear form T is called Hermitian-symmetric if T † = Tand Hermitian-anti-symmetric if T † = −T . In fact, T is Hermi-tian symmetric whenever

√−1T is Hermitian anti-symmetric (check

this). Moreover, every sesquilinear form can be uniquely writtenas the sum of Hermitian-symmetric and Hermitian-anti-symmetricones:

T =T + T †

2+ i

T − T †

2i, i =

√−1.

8After French mathematician Charles Hermite (1822–1901).

Page 27: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 47

Similarly to the case of quadratic forms, a sesquilinear form Tcan be reconstructed from its restriction v 7→ T (v,v) to the diag-onal (see Exercises). This restriction is a complex-valued functionof one vector variable. It is real-homogeneous of degre 2 in thefollowing sense: T (λv, λv) = |λ|2T (v,v). When T is Hermitian-symmetric, this function takes on purely real values, and when T isHermitian-anti-symmetric, purely imaginary. We will call such func-tionsHermitian and anti-Hermitian quadratic forms, or simplyan Hermitian and anti-Hermitian forms.

In coordinates, an Hermitian quadratic form looks this way:

H(z) =

n∑

i=1

n∑

j=1

zi hijzj , where hji = hij .

When [hij ] is the identity matrix, the Hermitian form is |z1|2 + · · ·+|zn|n. The space Cn equipped with this Hermitian form is called thestandard Hermitian space, and plays in complex geometry therole analogous to the standard Euclidean space of real geometry.

EXERCISES

84. Are the functions 3x, x3, x+1, 0, sinx, (1+x)2−(1−x)2, tan arctan x,arctan tanx linear? �

85. Verify the linearity property of the coordinate functions xi on Kn.

86. Let v ∈ Km, and let a : Kn → K be a linear function. Define a linearmap E : Kn → Km by E(x) = a(x)v, and compute the matrix of E. �

87. Write down the matrix of rotation through the angle θ in R2. �

88. Let

A =

[

1 2 31 −1 1

]

, B =

[

1 23 4

]

, C =

[

1 2−2 10 −1

]

.

Compute those of the products ABC,BAC,BCA,CBA,CAB,ACB whichare defined. �

89. Let∑n

j=1aijxj = bi, i = 1, . . . ,m, be a system ofm linear equations

in n unknowns (x1, . . . , xn). Show that it can be written in the matrix formAx = b, where A is a linear map from Kn to Km.

90. LetX,Y, Z,W be four sets, and h : X → Y , g : Y → Z, and f : Z → Wbe three functions. Show that the compositions f ◦ (g ◦ h) and (f ◦ g) ◦ hcoincide. �

91. Prove that products of upper triangular matrices are upper triangularand products of lower triangular matrices are lower triangular.

Page 28: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

48 Chapter 2. DRAMATIS PERSONAE

92.⋆ Prove that a matrix of a linear transformation from Kn to itself is up-per triangular if and only if the linear transformation maps each of the sub-spaces Span(e1) ⊂ Span(e1, e2) ⊂ Span(e1, e2, e3) ⊂ ... to itself. Derivefrom this that products of upper triangular matrices is upper triangular.

93. For an identity matrix I, prove AI = A and IB = B for all allowedsizes of A and B.

94. For a square matrix A, define its powers Ak for k > 0 as A · · ·A (ktimes), for k = 0 as I, and for k < 0 as A−1 · · ·A−1 (k times), assuming Ainvertible. Prove that A that AkAl = Ak+l for all integer k, l.

95.⋆ Compute

[

cos 19◦ − sin19◦

sin 19◦ cos 19◦

]19

. �

96. Compute powers Ak, k = 0, 1, 2, . . . , of the square matrix A all of whoseentries are zeroes, except that ai,i+1 = 1 for all i. �

97.⋆ Let N be a linear transformation from Kn to itself, determined byNe1 = 0, Ne2 = e1, Ne3 = e2, . . . , Nen = en−1. Desctibe the maps Nk,k = 1, 2, . . . , and find their matrices. �

98. For which sizes of matrices A and B, both products AB and BA:(a) are defined, (b) have the same size? �

99.⋆ Give examples of matrices A and B which do not commute, i.e. AB 6=BA, even though both products are defined and have the same size. �

100. When does (A+B)2 = A2 + 2AB +B2? �

101. Which diagonal matrices are invertible?

102. Prove that an inverse of a given matrix is unique when exists. �

103. Let A,B be invertible n×n-matrices. Prove that AB is also invertible,and (AB)−1 = B−1A−1. �

104.⋆ If AB is invertible, does it imply that A, B, and BA are invertible? �

105.⋆ Give an example of matrices A and B such that AB = I, but BA 6= I.

106. Let A : Kn → Km be an isomorphism. (Thus, m = n, but we considerKn and Km as two different copies of the coordinate space.) Prove thatafter suitable changes of coordinates in Kn and Km, the matrix of thistransformation becomes the identity matrix I. �

107. Is the function xy: linear? bilinear? quadratic? �

108. Find the coefficient matrix of the dot product. �

109. Prove that all anti-symmetric bilinear forms in R2 are proportionalto each other. �

110. Represent the bilinear form B = 2x1(y1 + y2) in R2 as the sum S+Aof symmetric and anti-symmetric ones. �

111. Find the symmetric bilinear forms corresponding to the quadraticforms (x1 + · · ·+ xn)

2 and∑

i<j xixj . �

Page 29: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

2. Matrices 49

112. Is AB necessarily symmetric if A and B are? �

113. Prove that for any matrix A, both AtA and AAt are symmetric.

114. Find a square matrix A such that AtA 6= AAt. �

115. Which of the function zw, zw, zw, zw of two complex variables are:(a) linear with respect to z? (b) anti-linear with respect to w?, (c) bilinear?(d) sesquilinear?

116. Let T (z1, z2;w1, w2) = z1 w2. Compute T †, represent T as the sumS1+iS2, where S1, S2 are Hermitian symmetric, and ompute the Hermitianquadratic forms H1 and H2 corresponding to S1 and S2. �

117. Verify that the real part Re(z1 z2 + · · · + zn−1 zn) is an Hermitianquadratic form and find the corresponding Hermitian-symmetric sesquilin-ear form. �

118. Show that all sesquilinear forms in Cn form a complex vector space,and find its dimension.

119. Do the Hermitian symmetric and Hermitian-anti-symmetric sesquiu-linear forms in Cn form a complex vector space? a real vector space? Findtheir dimensions. �

120. Let S be an Hermitian-symmetric sesquilinear form. Show that

S(z+w, z+w) = S(z, z) + S(w,w) + 2ReS(z,w)

S(z+ iw, z+ iw) = S(z, z) + S(w,w)− 2 ImS(z,w)

and derive from this, that S is uniquely determined by the correspondingHermitian quadratic form QS(z) := S(z, z).

Page 30: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise
Page 31: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 51

3 Determinants

Definition

Let A be a square matrix of size n:

A =

a11 ... a1n...

an1 ... ann

Its determinant is a scalar detA defined by the formula

detA =∑

σ

ε(σ)a1σ(1)a2σ(2)...anσ(n).

Here σ is a permutation of the indices 1, 2, ..., n. A permuta-tion σ can be considered as an invertible function i 7→ σ(i) fromthe set of n elements {1, ..., n} to itself. We use the functionalnotation σ(i) in order to specify the i-th term in the permutation

σ =(

1 . . . nσ(1) . . . σ(n)

)

. Thus, each elementary product in the

determinant formula contains exactly one matrix entry from eachrow, and these entries are chosen from n different columns. The sumis taken over all n! ways of making such choices. The coefficient ε(σ)in front of the elementary product equals 1 or −1 and is called thesign of the permutation σ.

We will explain the general rule of the signs after a few examples.In these examples, we begin using one more conventional notation fordeterminants. According to it, a square array of matrix entries placedbetween two vertical bars denotes the determinant of the matrix.

Thus,

[

a bc d

]

denotes a matrix, but

a bc d

denotes a number

equal to the determinant of that matrix.

Examples. (1) For n = 1, the determinant |a11| = a11.

(2) For n = 2, we have:

a11 a12a21 a22

= a11a22 − a12a21.

(3) For n = 3, we have 3! = 6 summands∣

a11 a12 a13a21 a22 a23a31 a32 a33

=

a11a22a33−a12a21a33+a12a23a31−a13a22a31+a13a21a32−a11a23a32

corresponding to permutations(123123

)

,(123213

)

,(123231

)

,(123321

)

,(123312

)

,(123132

)

.

Page 32: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

52 Chapter 2. DRAMATIS PERSONAE

The rule of signs for n = 3 is schematically shown on Figure 27.

Figure 27

Parity of Permutations

The general rule of signs depends on properties of permutations.

Let ∆n denote the following polynomial in in variables x1, . . . , xn:

∆n(x1, . . . , xn) =∏

1≤i<j≤n

(xi − xj).

Examples: ∆2 = x1 − x2, ∆3 = (x1 − x2)(x1 − x3)(x2 − x3). Bydefinition, ∆1 = 1. In general, ∆n is the product of all “n-choose-2”linear factors xi − xj written in such a way that i < j.

Let σ be a permutation of {1, . . . , n}. It acts on polynomi-als P in the variables x1, . . . , xn by permutation of the variables:(σP )(x1, . . . , xn) := P (xσ(1), . . . , xσ(n)).

Example. Let σ =(

1 2 33 1 2

)

. Then

σ∆3 = (x3−x1)(x3−x2)(x1−x2) = (−1)2(x1−x3)(x2−x3)(x1−x2).

One says that σ inverses a pair of indices i < j if σ(i) > σ(j).The total number l(σ) of pairs i < j that σ inverses is called thelength of the permutation σ. Thus, in the previous example, σinverses the pairs (1, 2) and (1, 3), and has length l(σ) = 2.

Lemma. σ∆n = ε(σ)∆n, where ε(σ) = (−1)l(σ).

Proof. Indeed, a permutation of {1, . . . , n} also permutes allpairs i 6= j, and hence permutes all the linear factors in ∆n. However,a factor xi − xj is transformed into xσ(i) − xσ(j), which occurs in theproduct ∆n with the same sign whenever σ(i) < σ(j), and with theopposite sign whenever σ(i) > σ(j). Thus, σ∆n differs from ∆n by

the sign (−1)l(σ). �

Page 33: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 53

A permutation σ is called even or odd depending on the signε(σ), i.e. when the lenght is even or odd respectively.

Examples. (1) The identity permutation id (defined by id(i) =i for all i) is even since l(id) = 0.

(2) Consider a transposition τ , i.e. a permutation that swapstwo indices, say i < j, leaving all other indices in their respectiveplaces. Then τ(j) < τ(i), i.e. τ inverses the pair of indices i < j.Besides, for every index k such that i < k < j we have: τ(j) < τ(k) <τ(i), i.e. both pairs i < k and k < j are inverted. Note that all otherpairs of indices are not inverted by τ , and hence l(τ) = 2(j − i) + 1.In particular, every transposition is odd: ε(τ) = −1.

Proposition. Composition of two even or two odd per-mutations is even, and composition of one even and oneodd permutation is odd:

ε(σσ′) = ε(σ)ε(σ′).

Proof. We have:

ε(σσ′)∆n := (σσ′)∆n = σ(σ′∆n) = ε(σ′)σ∆n = ε(σ′)ε(σ)∆n.

Corollary 1. Inverse permutations have the same parity.

Corollary 2. Whenever a permutation is written as theproduct of transpositions, the parity of the number of thetranspositions in the product remains the same and coin-cides with the parity of the permutation: If σ = τ1 . . . τN ,then ε(σ) = (−1)N .

Here are some illustrations of the above properties in connectionwith the definition of determinants.

Examples. (3) The transposition (21) is odd. That is why theterm a12a21 occurs in 2× 2-determinants with the negative sign.

(4) The permutations(123123

)

,(123213

)

,(123231

)

,(123321

)

,(123312

)

,(123132

)

havelengths l = 0, 1, 2, 3, 2, 1 and respectively signs ε = 1,−1, 1,−1, 1,−1(thus explaining Figure 27). Notice that each next permutation hereis obtained from the previous one by an extra flip.

(5) The permutation(12344321

)

inverses all the 6 pairs of indices andhas therefore length l = 6. Thus the elementary product a14a23a32a41occurs with the sign ε = (−1)6 = +1 in the definition of 4 × 4-determinants.

Page 34: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

54 Chapter 2. DRAMATIS PERSONAE

(6) Since inverse permutations have the same parity, the defini-tion of determinants can be rewritten “by columns:”

detA =∑

σ

ε(σ)aσ(1)1 ...aσ(n)n.

Indeed, each summand in this formula is equal to the summand inthe original definition corresponding to the permutation σ−1, andvice versa. Namely, reordering the factors aσ(1)1...aσ(n)n, so thatσ(1), . . . , σ(n) increase monotonically, yields a1σ−1(1)...anσ−1(n).

Properties of determinants

(i) Transposed matrices have equal determinants:

detAt = detA.

This follows from the last Example. Below, we will think of ann× n matrix as an array A = [a1, . . . ,an] of its n columns of size n(vectors from Cn if you wish) and formulate all further properties ofdeterminants in terms of columns. The same properties hold true forrows, since the transposition of A changes columns into rows withoutchanging the determinant.

(ii) Interchanging any two columns changes the sign ofthe determinant:

det[...,aj , ...,ai, ...] = − det[...,ai, ...,aj , ...].

Indeed, the operation replaces each permutation in the definitionof determinants by its composition with the transposition of the in-dices i and j. Thus changes the parity of the permutation, and thusreverses the sign of each summand.

Rephrasing this property, one says that the determinant, consid-ered as a function of n vectors a1, . . . ,an is totally anti-symmetric,i.e. changes the sign under every odd permutation of the vectors, andstays invariant under even. It implies that a matrix with two equalcolumns has zero determinant. It also allows one to formulate furthercolumn properties of determinants referring to the 1st column only,since the properties of all columns are alike.

(iii) Multiplication of a column by a number multipliesthe determinant by this number:

det[λa1,a2, ...,an] = λdet[a1,a2, ...,an].

Page 35: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 55

Indeed, this operation simply multiplies each of the n! elementaryproducts by the factor of λ.

This property shows that a matrix with a zero column has zerodeterminant.

(iv) The determinant function is additive with respect toeach column:

det[a′1 + a′′1 ,a2, ...,an] = det[a′1,a2, ...,an] + det[a′′1,a2, ...,an].

Indeed, each elementary product contains exactly one factorpicked from the 1-st column and thus splits into the sum of two ele-mentary products a′

σ(1)1aσ(2)2...aσ(n)n and a′′σ(1)1aσ(2)2...aσ(n)n. Sum-

ming up over all permutations yields the sum of two determinantson the right hand side of the formula.

The properties (iv) and (iii) together mean that the determinantfunction is linear with respect to each column separately. Togetherwith the property (ii), they show that adding a multiple of onecolumn to another one does not change the determinant ofthe matrix. Indeed,

|a1 + λa2,a2, ...| = |a1,a2, ...| + λ |a2,a2, ...| = |a1,a2, ...| ,

since the second summand has two equal columns.

The determinant function shears all the above properties with theidentically zero function. The following property shows that thesefunctions do not coincide.

(v) det I = 1.

Indeed, since all off-diagonal entries of the identity matrix arezeroes, the only elementary product in the definition of detA thatsurvives is a11...ann = 1.

The same argument shows that the determinant of any diagonalmatrix equals the product of the diagonal entries. It is not hard togeneralize the argument in order to see that the determinant of anyupper or lower triangular matrix is equal to the product of the diago-nal entries. One can also deduce this from the following factorizationproperty valid for block triangular matrices.

Consider an n×n-matrix

[

A BC D

]

subdivided into four blocks

A,B,C,D of sizes m×m, m× l, l×m and l× l respectively (where

Page 36: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

56 Chapter 2. DRAMATIS PERSONAE

of course m+ l = n). We will call such a matrix block triangularif C or B is the zero matrix 0. We claim that

det

[

A B0 D

]

= detA detD.

Indeed, consider a permutation σ of {1, ..., n} which sends at leastone of the indices {1, ...,m} to the other part of the set,{m+1, ...,m+ l}. Then σ must send at least one of {m+1, ...,m+ l}back to {1, ...,m}. This means that every elementary product in ourn×n-determinant which contains a factor from B must also containa factor from C, and hence vanish, if C = 0. Thus only the permu-tations σ which permute {1, ...,m} separately from {m+1, ...,m+ l}contribute to the determinant in question. Elementary productscorresponding to such permutations factor into elementary prod-ucts from detA and detD and eventually add up to the productdetAdetD.

Of course, the same holds true if B = 0 instead of C = 0.

We will use the factorization formula in the 1st proof of the fol-lowing fundamental property of determinants.

Multiplicativity

Theorem. The determinant is multiplicative with respect tomatrix products: for arbitrary n× n-matrices A and B,

det(AB) = (detA)(detB).

We give two proofs: one ad hoc, the other more conceptual.

Proof I. Consider the auxiliary 2n×2n matrix

[

A 0−I B

]

with

the determinant equal to the product (detA)(detB) according to thefactorization formula. We begin to change the matrix by adding tothe last n columns linear combinations of the first n columns withsuch coefficients that the submatrix B is eventually replaced by zerosubmatrix. Thus, in order to kill the entry bkj we must add thebkj-multiple of the k-th column to the n + j-th column. Accord-ing to the properties of determinants (see (iv)) these operations donot change the determinant but transform the matrix to the form[

A C−I 0

]

. We ask the reader to check that the entry cij of the

submatrix C in the upper right corner equals ai1b1j + ...+ ainbnj so

Page 37: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 57

that C = AB is the matrix product! Now, interchanging the i-thand n+ i-th columns, i = 1, ..., n, we change the determinant by the

factor of (−1)n and transform the matrix to the form

[

C A0 −I

]

.

The factorization formula applies again and yields detC det(−I). Weconclude that detC = detAdetB since det(−I) = (−1)n compen-sates for the previous factor (−1)n.�

Proof II. We will first show that the properties (i – v) com-pletely characterize det[v1, . . . ,vn] as a function of n columns vi ofsize n.

Indeed, consider a function f , which to n columns v1, . . . ,vn,associates a number f(v1, . . . ,vn). Suppose that f is linear withrespect to each column. Let ei denotes the ith column of the identitymatrix. Since v1 =

∑ni=1 vi1ei, we have:

f(v1,v2, . . . ,vn) =

n∑

i=1

vi1f(ei,v2, . . . ,vn).

Using linearity with respect to the 2nd column v2 =∑n

j=1 vj2ej , wesimilarly obtain:

f(v1,v2, . . . vn) =

n∑

i=1

n∑

j=1

vi1vj2f(ei, ej,v3, . . . ,vn).

Proceeding the same way with all columns, we get:

f(v1, . . . ,vn) =∑

i1,...,in

vi11 · · · vinnf(ei1 , . . . , ein).

Thus, f is determined by its values f(ei1 , . . . , ein) on strings of nbasis vectors.

Let us assume now that f is totally anti-symmetric. Then, if anytwo of the indices i1, . . . , in coincide, we have: f(ei1 , . . . , ein) = 0.

All other coefficients correspond to permutations σ =(

1 . . . ni1 . . . in

)

of the indices (1, . . . , n), and hence satisfy:

f(ei1 , . . . , ein) = ε(σ)f(e1, . . . , en).

Therefore, we find:

f(v1, . . . ,vn) =∑

σ

vσ(1)1 . . . vσ(n)nε(σ)f(e1, . . . , en),

= f(e1, . . . , en) det[v1, . . . ,vn].

Page 38: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

58 Chapter 2. DRAMATIS PERSONAE

Thus, we have established:

Proposition 1. Every totally anti-symmetric function ofn coordinate vectors of size n which is linear in each of themis proportional to the determinant function.

Next, given an n× n matrix C, put

f(v1, . . . ,vn) := det[Cv1, . . . , Cvn].

Obviously, the function f is a totally anti-symmetric in all vi (sincedet is). Multiplication by C is linear:

C(λu+ µv) = λCu+ µCv for all u,v and λ, µ.

. Therefore, f is linear with respect to each vi (as composition oftwo linear operations). By the previous result, f is proportional todet. Since Cei are columns of C, we conclude that the coefficientof proportionality f(e1, . . . , en) = detC. Thus, we have found thefollowing interpretation of detC.

Proposition 2. detC is the factor by which the determi-nant function of n vectors vi is multiplied when the vectorsare replaced with Cvi.

Now our theorem follows from the fact that when C = AB, thesubstitution v 7→ Cv is the composition v 7→ Av 7→ ABv of consec-utive substitutions defined by A and B. Under the action of A, thefunction det is multiplied by the factor detA, then under the actionof B by another factor detB. But the resulting factor (detA)(detB)must be equal to detC. �

Corollary. If A is invertible, then detA is invertible.

Indeed, (detA)(detA−1) = det I = 1, and hence detA−1 is recip-rocal to detA. The converse statement: that matrices with invertibledeterminants are invertible, is also true due to the explicit formulafor the inverse matrix, described in the next section.

Page 39: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 59

Remark. Of course, a real or complex number detA is invertiblewhenever detA 6= 0. Yet over the integers Z this is not the case:the only invertible integers are ±1. The above formulation, andseveral similar formulations that follow, which refer to invertibilityof determinants, are preferable as they are more general.

The Cofactor Theorem

In the determinant formula for an n × n-matrix A each elementaryproduct ±a1σ(1)... begins with one of the entries a11, ..., a1n of thefirst row. The sum of all terms containing a11 in the 1-st place isthe product of a11 with the determinant of the (n − 1) × (n − 1)-matrix obtained from A by crossing out the 1-st row and the 1-stcolumn. Similarly, the sum of all terms containing a12 in the 1-stplace looks like the product of a12 with the determinant obtained bycrossing out the 1-st row and the 2-nd column of A. In fact it differsby the factor of −1 from this product, since switching the columns1 and 2 changes signs of all terms in the determinant formula andinterchanges the roles of a11 and a12. Proceeding in this way witha13, ..., a1n we arrive at the cofactor expansion formula for detAwhich can be stated as follows.

Figure 29

ij

ij

1

2

3

4

5

1 52 4311

a a

aa

1n

n1 nn

Figure 28

a

The determinant of the (n− 1)× (n− 1)-matrix obtained from Aby crossing out the row i and column j is called the (ij)-minor of A(Figure 28). Denote it by Mij . The (ij)-cofactor Aij of the matrixA is the number that differs from the minor Mij by a factor ±1:

Aij = (−1)i+jMij .

The chess-board of the signs (−1)i+j is shown on Figure 29. Withthese notations, the cofactor expansion formula reads:

detA = a11A11 + a12A12 + ...+ a1nA1n.

Page 40: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

60 Chapter 2. DRAMATIS PERSONAE

Example.∣

a11 a12 a13a21 a22 a23a31 a32 a33

= a11

a22 a23a32 a33

−a12

a21 a23a31 a33

+a13

a21 a22a31 a32

.

Using the properties (i) and (ii) of determinants we can adjustthe cofactor expansion to the i-th row or j-th column:

detA = ai1Ai1 + ...+ ainAin = a1jA1j + ...+ anjAnj , i, j = 1, ..., n.

These formulas reduce evaluation of n × n-determinants to that of(n − 1) × (n − 1)-determinants and can be useful in recursive com-putations.

Furthermore, we claim that applying the cofactor formula to theentries of the i-th row but picking the cofactors of another row weget the zero sum:

ai1Aj1 + ...+ ainAjn = 0 if i 6= j.

Indeed, construct a new matrix A replacing the j-th row by a copy ofthe i-th row. This forgery does not change the cofactors Aj1, ..., Ajn

(since the j-th row is crossed out anyway) and yields the cofactor

expansion ai1Aj1 + ... + ainAjn for det A. But A has two identical

rows and hence det A = 0. The same arguments applied to thecolumns yield the dual statement:

a1iA1j + ...+ aniAnj = 0 if i 6= j.

All the above formulas can be summarized in a single matrix identity.Introduce the n× n-matrix adj(A), called adjoint to A, by placingthe cofactor Aij on the intersection of j-th row and i-th column. Inother words, each aij is replaced with the corresponding cofactor Aij ,and then the resulting matrix is transposed:

adj

a11 . . . a1n. . . aij . . .an1 . . . ann

=

A11 . . . An1

. . . Aji . . .A1n . . . Ann

.

Theorem. A adj(A) = (detA) I = adj(A) A.

Corollary. If detA is invertible then A is invertible, and

A−1 =1

detAadj(A).

Example. If ad− bc 6= 0, then

[

a bc d

]−1

= 1ad−bc

[

d −b−c a

]

.

Page 41: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 61

Cramer’s Rule

This is an application of the Cofactor Theorem to systems of linearequations. Consider a system

a11x1+ · · ·+ a1nxn = b1· · ·

an1x1+ · · ·+ annxn = bn

of n linear equations with n unknowns (x1, . . . , xn). It can be writtenin the matrix form

Ax = b,

where A is the n×n-matrix of the coefficients aij , b = [b1, . . . , bn]t is

the column of the right hand sides, and x is the column of unknowns.In the following Corollary, ai denote columns of A.

Corollary. If detA is invertible then the system of lin-ear equations Ax = b has a unique solution given by theformulas:

x1 =det[b,a2, ...,an]

det[a1, ...,an], . . . , xn =

det[a1, ...,an−1,b]

det[a1, ...,an].

Indeed, when detA 6= 0, the matrix A is invertible. Multi-plying the matrix equation Ax = b by A−1 on the left, we find:x = A−1b. Thus the solution is unique, and xi = (detA)−1(A1ib1 +... + Anibn) according to the cofactor formula for the inverse ma-trix. But the sum b1A1i + ... + bnAni is the cofactor expansion fordet[a1, ...,ai−1,b,ai+1, ...,an] with respect to the i-th column.

Example. Suppose that a11a22 6= a12a21. Then the system

a11x1 + a12x2 = b1a21x2 + a22x2 = b2

has a unique solution

x1 =

b1 a12b2 a22

a11 a12a21 a22

, x2 =

a11 b1a21 b2

a11 a12a21 a22

.

Page 42: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

62 Chapter 2. DRAMATIS PERSONAE

Three Cool Formulas

We collect here some useful generalizations of previous results.

A. We don’t know of any reasonable generalization of determi-nants to the situation when matrix entries do not commute. However

the following generalization of the formula det

[

a bc d

]

= ad− bc is

instrumental in some non-commutative applications.9

In the block matrix

[

A BC D

]

, assume that D−1 exists.

Then det

[

A BC D

]

= det(A−BD−1C) detD.

Proof:

[

A BC D

] [

I 0−D−1C I

]

=

[

A−BD−1C B0 D

]

.

B. Lagrange’s formula10 below generalizes cofactor expansions.

By a multi-index I of length |I| = k we mean an increasingsequence i1 < · · · < ik of k indices from the set {1, . . . , n}. Givenand n × n-matrix A and two multi-indices I, J of the same lengthk, we define the (IJ)-minor of A as the determinant of the k × k-matrix formed by the entries aiαjβ of A located at the intersectionsof the rows i1, . . . , ik with columns j1, . . . , jk (see Figure 30). Also,denote by I the multi-index complementary to I, i.e. formed bythose n− k indices from {1, . . . , n} which are not contained in I.

For each multi-index I = (i1, . . . , ik), the following cofac-tor expansion with respect to rows i1, . . . , ik holds true:

detA =∑

J :|J |=k

(−1)i1+···+ik+j1+···+jkMIJMI J ,

where the sum is taken over all multi-indices J = (j1, . . . , jk)of length k.

Similarly, one can similarly write Lagrange’s cofactor expansionformula with respect to given k columns.

Example. Let a1,a2,a3,a4 and b1,b2,b3,b4 be 8 vectors on the

plane. Then

a1 a2 a3 a4b1 b2 b3 b4

= |a1 a2||b3 b4| − |a1 a3||b2 b4|+ |a1 a4||b2 b3| − |a2 a3||b1 b4|+ |a2 a4||b1 b3| − |a3 a4||b1 b2|.

9Notably in the definition of Berezinian in super-mathematics [7].10After Joseph-Louis Lagrange (1736–1813).

Page 43: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 63

In the proof of Lagrange’s formula, it suffices to assume that itis written with respect to the first k rows, i.e. that I = (1, . . . , k). In-deed, interchanging them with the rows i1 < · · · < ik takes(i1 − 1) + (i2 − 2) + · · ·+ (ik − k) transpositions, which is accountedfor by the sign (−1)i1+···+ik in the formula.

Next, multiplying out MIJMI J , we find k!(n − k)! elementaryproducts of the form:

±a1,jα1· · · ak,jαk

ak+1,jβ1· · · an,jβn−k

,

where α =(

1 . . . kα1 . . . αk

)

and β =(

1 . . . n− kβ1 . . . βn−k

)

are permu-

tations, and jαµ ∈ J , jβν∈ J . It is clear that the total sum over

multi-indices I contains each elementary product from detA, anddoes it exactly once. Thus, to finish the proof, we need to comparethe signs.

Figure 30

i

i

i

j j2 k

j1

1

2

k

The sign ± in the above formula is equal to ε(α)ε(β), the prod-uct of the signs of the permutations α and β. The sign of thiselementary product in the definition of detA is equal to the sign

of the permutation(

1 . . . k k + 1 . . . njα1

. . . jαkjβ1

. . . jβn−k

)

on the set

J ∪ J = {1, . . . , n}. Reordering separately the first k and last n− kindices in the increasing order changes the sign of the permutationby ε(α)ε(β). Therefore the signs of all summands of detA whichoccur in MIJMI J are coherent. It remains to find the total sign withwhich MIJMI J occurs in detA, by computing the sign of the permu-

tation σ :=(

1 . . . k k + 1 . . . nj1 . . . jk j1 . . . jn−k

)

, where j1 < · · · jk and

j1 < · · · < jn−k.

Starting with the identity permutation (1, 2 . . . , j1, . . . , j2, . . . , n),it takes j1 − 1 transpositions of nearby indices to move j1 to the 1stplace. Then it takes j2− 2 such transpositions to move j2 to the 2nd

Page 44: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

64 Chapter 2. DRAMATIS PERSONAE

place. Continuing this way, we find that

ε(σ) = (−1)(j1−1)+···+(jk−k) = (−1)1+···+k+j1+···+jk .

This agrees with Lagrange’s formula, since I = {1, . . . , k}. �.

C. Let A and B be k × n and n × k matrices (think of k < n).For each multi-index I = (i1, . . . , ik), denote by AI and BI the k×k-matrices formed by respectively: columns of A and rows of B withthe indices i1, . . . , ik.

The determinant of the k × k-matrix AB is given by thefollowing Binet–Cauchy formula:11

detAB =∑

I

(detAI)(detBI).

Note that when k = n, this turns into the multiplicative propertyof determinants: det(AB) = (detA)(detB). Our second proof of itcan be generalized to establish the formula of Binet–Cauchy. Namely,let a1, . . . ,an denote columns of A. Then the jth column of C = ABis the linear combination: cj = a1b1j + · · · + anbnj . Using linearityin each cj , we find:

det[c1, . . . , ck] =∑

1≤i1,...,ik≤k

det[ai1 , . . . ,aik ]bi11 · · · bikk.

If any two of the indices iα coincide, det[ai1 , . . . ,aik ] = 0. Thus the

sum is effectively taken over all permutations

(

1 . . . ki1 . . . ik

)

on the

set12 {i1, . . . , ik}. Reordering the columns ai1 , . . . ,aik in the increas-ing order of the indices (an paying the “fees” ±1 according to paritiesof permutations) we obtain the sum over all multi-indices of lengthk:

i′1<···<i′

k

det[ai′1, . . . ,ai′

k]∑

σ

ε(σ)bi11 · · · bikk.

The sum on the right is taken over permutations σ′ =(

i′1 . . . i′ki1 . . . ik

)

.

It is equal to detBI , where I = (i′1, . . . , i′k). �

Corollary 1. If k > n, detAB = 0.

11After Jacques Binet (1786–1856) and Augustin Louis Cauchy (1789–1857).12Remember that in a set, elements are unordered!

Page 45: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 65

This is because no multi-indices of length k > n can be formedfrom {1, . . . , n}. In the oppositely extreme case k = 1, Binet–Cauchy’s formula turns into the expression utv =

uivi for thedot product of coordinate vectors. A “Pythagorean” interpretationof the following identity will come to light in the next chapter, inconnection with volumes of parallelepipeds.

Corollary 2. detAAt =∑

I(detAI)2.

EXERCISES

121. Prove that the following determinant is equal to 0:

0 0 0 a b0 0 0 c d0 0 0 e fp q r s tv w x y z

. �

122. Compute determinants:

cos x − sinxsinx cos x

,

coshx sinhxsinhx coshx

,

cos x sin ysinx cos y

. �

123. Compute determinants:

0 1 11 0 11 1 0

,

0 1 11 2 31 3 6

,

1 i 1 + i−i 1 01− i 0 1

. �

124. List all the 24 permutations of {1, 2, 3, 4}, find the length and the signof each of them. �

125. Find the length of the following permutation:(

1 2 . . . k k + 1 k + 2 . . . 2k1 3 . . . 2k − 1 2 4 . . . 2k

)

. �

126. Find the maximal possible length of permutations of {1, ..., n}. �

127. Find the length of a permutation

(

1 . . . ni1 . . . in

)

given the length l

of the permutation

(

1 . . . nin . . . i1

)

. �

128. Prove that inverse permutations have the same length. �

129. Compare parities of permutations of the letters a,g,h,i,l,m,o,r,t in thewords logarithm and algorithm. �

130. Prove that the identity permutations are the only ones of length 0.

Page 46: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

66 Chapter 2. DRAMATIS PERSONAE

131. Find all permutations of length 1. �

132.⋆ Show that every permutation σ can be written as the product of l(σ)transpositions of nearby indices. �

133.⋆ Represent the permutation

(

1 2 3 4 54 5 1 3 2

)

as composition of a

minimal number of transpositions. �

134. Do products a13a24a53a41a35 and a21a13a34a55a42 occur in the defin-ing formula for determinants of size 5? �

135. Find the signs of the elementary products a23a31a42a56a14a65 anda32a43a14a51a66a25 in the definition of determinants of size 6 by computingthe numbers of inverted pairs of indices. �

136. Compute the determinants

13247 1334728469 28569

,

246 427 3271014 543 443−342 721 621

. �

137. The numbers 195, 247, and 403 are divisible by 13. Prove that the

following determinant is also divisible by 13:

1 9 52 4 74 0 3

. �

138. Professor Dumbel writes his office and home phone numbers as a 7×1-matrix O and 1× 7-matrix H respectively. Help him compute det(OH). �

139. How does a determinant change if all its n columns are rewritten inthe opposite order? �

140.⋆ Solve the equation

1 x x2 ... xn

1 a1 a21 ... an11 a2 a22 ... an2

...1 an a2n ... ann

= 0, where all a1, ..., an

are given distinct numbers. �

141. Prove that an anti-symmetric matrix of size n has zero determinantif n is odd. �

142. How do similarity transformations of a given matrix affect its deter-minant? �

143. Prove that the adjoint matrix of an upper (lower) triangular matrixis upper (lower) triangular.

144. Which triangular matrices are invertible?

145. Compute the determinants: (∗ is a wild card):

(a)

∗ ∗ ∗ an∗ ∗ . . . 0∗ a2 0 . . .a1 0 . . . 0

, (b)

∗ ∗ a b∗ ∗ c de f 0 0g h 0 0

. �

Page 47: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

3. Determinants 67

146. Compute determinants using cofactor expansions:

(a)

1 2 2 10 1 0 22 0 1 10 2 0 1

, (b)

2 −1 0 0−1 2 −1 00 −1 2 −10 0 −1 2

. �

147. Compute inverses of matrices using the Cofactor Theorem:

(a)

[

1 2 33 1 22 3 1

]

, (b)

[

1 1 10 1 10 0 1

]

. �

148. Solve the systems of linear equations Ax = b where A is one of thematrices of the previous exercise, and b = [1, 0, 1]t. �

149. Compute

1 −1 0 00 1 −1 00 0 1 −10 0 0 1

−1

.

150. Express det(adj(A)) of the adjoint matrix via detA. �

151. Which integer matrices have integer inverses? �

152. Solve systems of equations using Cramer’s rule:

(a)2x1 − x2 − x3 = 4

3x1 + 4x2 − 2x3 = 113x1 − 2x2 + 4x3 = 11

, (b)x1 + 2x2 + 4x3 = 315x1 + x2 + 2x3 = 293x1 − x2 + x3 = 10

. �

153.⋆ Compute determinants:

(a)

0 x1 x2 . . . xn

x1 1 0 . . . 0x2 0 1 . . . 0. . . . .xn 0 . . . 0 1

, (b)

a 0 0 0 0 b0 a 0 0 b 00 0 a b 0 00 0 c d 0 00 c 0 0 d 0c 0 0 0 0 d

� �.

154.⋆ Let Pij , 1 ≤ i < j ≤ 4, denote the 2 × 2-minor of a 2 × 4-matrixformed by the columns i and j. Prove the following Plucker identity13

P12P34 − P13P24 + P14P23 = 0. �

13After Julius Plucker (1801–1868).

Page 48: Dramatis Personaela14.wdfiles.com/local--files/read-the-textbook/la_14_ch...24 Chapter2. DRAMATISPERSONAE The poinwise operations with matrices as functions turn into elemen-twise

68 Chapter 2. DRAMATIS PERSONAE

155. The cross product of two vectors x,y ∈ R3 is defined by

x× y :=

(∣

x2 x3

y2 y3

,

x3 x1

y3 y1

,

x1 x2

y1 y2

)

.

Prove that the length |x× y| =√

|x|2|y|2 − 〈x,y〉2. �

156.⋆ Prove that an +1

an−1 +1

· · ·+ 1

a1 +1

a0

=∆n

∆n−1

,

where ∆n =

a0 1 0 . . . 0−1 a1 1 . . . 0. . . . .0 . . . −1 an−1 10 . . . 0 −1 an

. �

157.⋆ Compute:

λ −1 0 . . . 00 λ −1 . . . 0. . . . .0 . . . 0 λ −1an an−1 . . . a2 λ+ a1

. �

158.⋆ Compute:

1 1 1 . . . 11

(

2

1

) (

3

1

)

. . .(

n

1

)

1(

3

2

) (

4

2

)

. . .(

n+1

2

)

. . . . .1

(

nn−1

) (

n+1

n−1

)

. . .(

2n−2

n−1

)

. � �

159.⋆ Prove Vandermonde’s identity14

1 x1 x21 . . . xn−1

1

1 x2 x22 . . . xn−1

2

. . . . .1 xn x2

n . . . xn−1n

=∏

1≤i<j≤n

(xj − xi). �

160.⋆ Compute:

1 2 3 . . . n1 23 33 . . . n3

. . . . .1 22n−1 32n−1 . . . n2n−1

. � �

14After Alexandre-Theophile Vandermonde (1735–1796).


Recommended