41. Vector spaces

41. Vector spaces

This lecture is intended as a foretaste of things to come. We will develop extra layers of abstraction, and this will allow us to apply the ideas of linear algebra in new contexts.

They key operations we've used on vectors are additon and rescaling: v + w , Ξ» ⁒ v . These operations are all we need to state the definitions of "linear map" and "linear subspace". But we can add and rescale things much more general than column vectors in 𝐑 n .

Definition:

A (real) vector space is a set V with operations add : V Γ— V β†’ V ⁒ Β writtenΒ  ⁒ add ⁒ ( v , w ) = v + w rescale : 𝐑 Γ— V β†’ V ⁒ Β writtenΒ  ⁒ rescale ⁒ ( Ξ» , v ) = Ξ» ⁒ v The elements of V don't have to be elements of 𝐑 n , and add and rescale don't have to be addition and rescaling of vectors in 𝐑 n , they could be any other objects and operations which behave in a similar manner, by which I mean that the following conditions hold (for all u , v , w ∈ V and Ξ» , ΞΌ ∈ 𝐑 ):

  • There is a distinguished element 0 ∈ V .

  • u + ( v + w ) = ( u + v ) + w

  • v + w = w + v

  • v = 0 + v = v + 0

  • v + ( - v ) = 0

  • 1 ⁒ v = v

  • Ξ» ⁒ ( ΞΌ ⁒ v ) = ( Ξ» ⁒ ΞΌ ) ⁒ v

  • Ξ» + ΞΌ ) v = Ξ» v + ΞΌ v

  • Ξ» ⁒ ( v + w ) = Ξ» ⁒ v + Ξ» ⁒ w .

So a vector space is a set with operations that satisfy these conditions. In particular, 𝐑 n equipped with vector addition and rescaling is an example of a vector space. Indeed, it is a theorem that any finite-dimensional vector space is equivalent to 𝐑 n for some n . But if you allow yourself to consider infinite-dimensional vector spaces there are more interesting examples.

Example:

Let V be the set of all functions from 𝐑 to 𝐑 . Given functions f , g , define f + g to be the function whose value at x ∈ 𝐑 is f ⁒ ( x ) + g ⁒ ( x ) , and given a scalar Ξ» ∈ 𝐑 we define Ξ» ⁒ f to be the function whose value at x ∈ 𝐑 is Ξ» ⁒ f ⁒ ( x ) . This gives V the structure of a vector space. It's not the same as 𝐑 n for any n : it is actually an infinite-dimensional vector space.

Example:

Inside our previous example we can find some linear subspaces. For example, the set of continuous functions π’ž 0 ⁒ ( 𝐑 ) is a subspace. Inside that we have another subspace: the space of once continuously-differentiable functions π’ž 1 ⁒ ( 𝐑 ) . Inside that we have the subspace of twice continuously-differentiable functions π’ž 2 ⁒ ( 𝐑 ) . And so on. Inside all of these we have the infinitely-differentiable functions π’ž ∞ ⁒ ( 𝐑 ) , and, inside that, that space of analytic functions π’ž Ο‰ ⁒ ( 𝐑 ) (infinitely-differentiable functions whose Taylor series converges on a neighbourhood of the origin). This gives us an infinite nested sequence of subspaces: π’ž Ο‰ ⁒ ( 𝐑 ) βŠ‚ π’ž ∞ ⁒ ( 𝐑 ) βŠ‚ β‹― βŠ‚ π’ž 2 ⁒ ( 𝐑 ) βŠ‚ π’ž 1 ⁒ ( 𝐑 ) βŠ‚ π’ž 0 ⁒ ( 𝐑 ) βŠ‚ V Inside the space of analytic functions, we have the space of polynomials (which we already met).

Example:

Differentiation defines a linear map d d ⁒ x : π’ž 1 ⁒ ( 𝐑 ) β†’ π’ž 0 ⁒ ( 𝐑 ) , f ↦ d ⁒ f d ⁒ x . In other words, you start with a once continuously-differentiable function and differentiate it to get a continuous function). To show that it's linear, all we need to do is check that d d ⁒ x ⁒ ( f + g ) = d ⁒ f d ⁒ x + d ⁒ g d ⁒ x , d d ⁒ x ⁒ ( Ξ» ⁒ f ) = Ξ» ⁒ d ⁒ f d ⁒ x , Β forΒ  ⁒ Ξ» ∈ 𝐑 .

What is the kernel of d d ⁒ x ? It consists of functions whose derivative is zero, in other words constant functions. That is, ker ⁑ ( d / d ⁒ x ) βŠ‚ π’ž 1 ⁒ ( 𝐑 ) is the 1-dimensional subspace of constant functions.

The eigenfunctions of d / d ⁒ x (with eigenvalue λ ) will be a function f such that d ⁒ f / d ⁒ x = λ ⁒ f . This is a differential equation for f ; its solution is f ⁒ ( x ) = C ⁒ e λ ⁒ x . This is why the exponential functions is so important: it's an eigenfunction of differentiation. Similarly the eigenfunctions of d 2 / d ⁒ x 2 are the solutions of d 2 ⁒ f / d ⁒ x 2 = λ ⁒ f , that is A ⁒ cos ⁑ ( x ⁒ - λ ) + B ⁒ sin ⁑ ( x ⁒ - λ ) . This is a 2-dimensional eigenspace.

Algebraic numbers

Example:

Here is an example from number theory. We have come across the idea that your vectors can have complex coefficients or real coefficients, but we can work much more generally by requiring our coefficients to live in some coefficient field K . In this example, we will take K = 𝐐 (the rational numbers), but you could imagine all sorts of things (the integers modulo 5, the 17-adic numbers, and goodness knows what else). The only difference this makes in the definition of a vector space is that rescaling can only be done by elements of K , that is the rescaling map is rescale : K Γ— V β†’ V .

A number x ∈ 𝐂 is algebraic if there exists a polynomial P ⁒ ( z ) = a n ⁒ z n + β‹― + a 0 with a n , … , a 0 ∈ 𝐐 for which P ⁒ ( x ) = 0 .

  • For example, 2 is an algebraic number because it's a root of z 2 = 2 .

  • i is algebraic because it's a root of z 2 = - 1 .

  • Ο€ and e are not algebraic (i.e. they're transcendental)

The set of all algebraic numbers is called 𝐐 ¯ .

Lemma:

𝐐 ¯ is a 𝐐 -vector space.

We need to show that if x , y are algebraic numbers and λ is a rational number then x + y and λ ⁒ x are algebraic numbers.

To see that Ξ» ⁒ x ∈ 𝐐 Β― , note that there is a polynomial P ⁒ ( z ) = a n ⁒ z n + β‹― + a 0 with a k ∈ 𝐐 and P ⁒ ( x ) = 0 . Now R ⁒ ( z ) = a n Ξ» n ⁒ z n + β‹― + a 1 Ξ» ⁒ z + a 0 satisfies R ⁒ ( Ξ» ⁒ z ) = P ⁒ ( z ) , so R ⁒ ( Ξ» ⁒ x ) = P ⁒ ( x ) = 0 , showing that Ξ» ⁒ x ∈ 𝐐 Β― . Note that we really need Ξ» ∈ 𝐐 (or else the coefficients a k / Ξ» k are not rational), so this is only a 𝐐 -vector space (not a 𝐑 -vector space).

To show that x + y is algebraic is much harder: we can't prove it here. In the words of Pierre Samuel (Algebraic Theory of Numbers), "The reader will have to exert himself to show that 5 3 + 7 3 is an algebraic integer, and will be convinced that the steps which lead to a proof that this number is algebraic may not be easily generalised." The nicest proof uses the theory of modules over rings.

In fact, slightly more is true: the product of two algebraic numbers is also an algebraic number, so 𝐐 Β― is a field. One of the most complicated objects in mathematics is the group G a l ( 𝐐 Β― : 𝐐 ) , the Galois group of 𝐐 . This is the set of invertible 𝐐 -linear maps f : 𝐐 Β― β†’ 𝐐 Β― which preserve the product ( f ⁒ ( a ⁒ b ) = f ⁒ ( a ) ⁒ f ⁒ ( b ) ).

The elements of G a l ( 𝐐 ¯ : 𝐐 ) are like infinitely large matrices with rational entries (because 𝐐 ¯ is infinite-dimensional over 𝐐 ). One way people study this enormously complicated group is using Galois representations: associating to each g ∈ G a l ( 𝐐 ¯ : 𝐐 ) a finite-dimensional matrix M ⁒ ( g ) such that M ⁒ ( g 1 ⁒ g 2 ) = M ⁒ ( g 1 ) ⁒ M ⁒ ( g 2 ) . Constructing Galois representations is a very difficult task, but the payoff can be enormous. Galois representations played an important role in Wiles's proof of Fermat's last theorem, and continue to play an important role in modern number theory.