# Isomorphisms – Serlo

Zur Navigation springen Zur Suche springen

## Isomorphic Structures and Isomorphisms

### Isomorphic Structures

We consider the vector space ${\displaystyle K[X]_{\leq 2}}$ of polynomials of degree less than or equal to ${\displaystyle 2}$ and we consider ${\displaystyle \mathbb {R} ^{3}}$. Vectors in these spaces have a one-to-one correspondence, as we have already seen in the introduction article to vector spaces:

${\displaystyle {\begin{array}{cccc}f(x)&=&{\color {Red}7}x^{2}&+{\color {OliveGreen}3}x&+({\color {NavyBlue}-2})&\leftrightarrow &({\color {Red}7},{\color {OliveGreen}3},{\color {NavyBlue}-2})^{T}\\g(x)&=&{\color {Red}1}x^{2}&+{\color {OliveGreen}0}x&+{\color {NavyBlue}6}&\leftrightarrow &({\color {Red}1},{\color {OliveGreen}0},{\color {NavyBlue}6})^{T}\end{array}}}$

We also found that addition and scalar multiplication work the same way in both vector spaces:

${\displaystyle {\begin{array}{ccc}{\color {Red}7}x^{2}&+{\color {OliveGreen}3}x&+({\color {NavyBlue}-2})&\leftrightarrow &({\color {Red}7},{\color {OliveGreen}3},{\color {NavyBlue}-2})^{T}\\&+&&&+\\{\color {Red}1}x^{2}&+{\color {OliveGreen}0}x&+{\color {NavyBlue}6}&\leftrightarrow &({\color {Red}1},{\color {OliveGreen}0},{\color {NavyBlue}6})^{T}\\&=&&&=\\{\color {Red}8}x^{2}&+{\color {OliveGreen}3}x&+{\color {NavyBlue}4}&\leftrightarrow &({\color {Red}8},{\color {OliveGreen}3},{\color {NavyBlue}4})^{T}\end{array}}}$

${\displaystyle {\begin{array}{ccc}&-1&&&-1\\&\cdot &&&\cdot \\{\color {Red}7}x^{2}&+{\color {OliveGreen}3}x&+({\color {NavyBlue}-2})&\leftrightarrow &({\color {Red}7},{\color {OliveGreen}3},{\color {NavyBlue}-2})^{T}\\&=&&&=\\{\color {Red}-7}x^{2}&+({\color {OliveGreen}-3})x&+{\color {NavyBlue}2}&\leftrightarrow &({\color {Red}-7},{\color {OliveGreen}-3},{\color {NavyBlue}2})^{T}\\\end{array}}}$

In general, vector spaces can be thought of as sets with some structure. In our example, we can match the sets 1 to 1. And also the structures (i.e. addition and multiplication) can be matched. So both vector spaces "essentially carry the same information", although they formally comprise different objects. In such a case, we will call the two vector spaces isomorphic (to each other). The bijection which identifies the two vector spaces is then called an isomorphism.

We now derive what the mathematical definition is of "two vector spaces ${\displaystyle V}$ and ${\displaystyle W}$ are isomorphic":

The identification of the sets is given by a bijective mapping ${\displaystyle f:V\to W}$. Preserving the structure means that addition and scalar multiplication are preserved when mapping back and forth with ${\displaystyle f}$ and ${\displaystyle f^{-1}}$. But "preserving addition and scalar multiplication" for a mapping between vector spaces is nothing else than "being linear". So we want ${\displaystyle f}$ and ${\displaystyle f^{-1}}$ to be linear.

Definition (Isomorphic)

The vector spaces ${\displaystyle V}$ and ${\displaystyle W}$ are isomorphic if there is a bijective map ${\displaystyle f\colon V\to W}$ between them such that ${\displaystyle f}$ and ${\displaystyle f^{-1}}$ are linear. We then write ${\displaystyle V\cong W}$.

Let us now return to our example from above. In this case, the identification map we are looking for from the Definition would look like this:

{\displaystyle {\begin{aligned}f\colon \mathbb {R} ^{3}&\to K[X]_{\leq 2}\\[0.3em](a,b,c)&\mapsto ax^{2}+bx+c\end{aligned}}}

### Isomorphism

We also want to give a name to the map ${\displaystyle f}$ introduced above:

Definition (Isomorphism)

An isomorphism between vector spaces ${\displaystyle V}$ and ${\displaystyle W}$ is a bijective map ${\displaystyle f\colon V\to W}$ such that ${\displaystyle f}$ and ${\displaystyle f^{-1}}$ are linear.

### Alternative Derivation

Now let's look at the term "vector space" from a different point of view. We can also think of a vector space as a basis together with corresponding linear combinations of the basis. So we can call vector spaces "equal" if we can identify the bases 1 to 1 and the corresponding linear combinations are generated in the same way. In other words, we are looking for a mapping that preserves both bases and linear combinations. What property must the mapping have in order to generate the same linear combinations? The answer is almost in the name: The mapping must be linear.

Let us now turn to the question of what property a linear map needs in order to map bases to bases. A basis is nothing else than a linearly independent generator. Thus, the map must preserve generators and linear independence. A linear map that preserves a generator is called an epimorphism - that is, a surjective linear map. A linear map that preserves linear independence is called a monomorphism and is thus an injective linear map. So the function we are looking for is an epimorphism and a monomorphism at the same time. As a monomorphism it must be injective. As an epimorphism, on the other hand, the mapping must be surjective. So overall we get a bijective linear map. This we again call an isomorphism. This gives us the alternative definition:

Definition (Alternative definition of isomorphism and isomorphism)

Two vector spaces ${\displaystyle V}$ and ${\displaystyle W}$ are isomorphic if there is a bijective linear map ${\displaystyle f\colon V\to W}$ between them.

A map ${\displaystyle f\colon V\to W}$ is called an isomorphism if it is a bijective linear map.

## Inverse Mappings of Linear Bijections are Linear

We have derived two descriptions for isomorphisms. Thus we have also two different definitions. The first one seems to require more than the second one: In the first definition, an isomorphism ${\displaystyle f}$ must additionally satisfy that ${\displaystyle f^{-1}}$ is linear. Does this give us two different mathematical objects, or does linearity of ${\displaystyle f}$ already imply linearity of ${\displaystyle f^{-1}}$? According to our intuition, both definitions should define the same objects. So ${\displaystyle f}$ being linear should then imply ${\displaystyle f^{-1}}$ being linear. And indeed, this is the case:

Theorem (The inverse map of a bijective linear map is again linear.)

Let ${\displaystyle f\colon V\to W}$ be a bijective linear map. Then the inverse mapping ${\displaystyle g\colon W\to V}$ is also linear.

How to get to the proof? (The inverse map of a bijective linear map is again linear.)

We want to show that ${\displaystyle g}$ is linear. For this, both ${\displaystyle g(w+w')=g(w)+g(w')}$ and ${\displaystyle g(\lambda \cdot w)=\lambda \cdot g(w)}$ must hold for all vectors ${\displaystyle w,w'\in W}$ and scalars ${\displaystyle \lambda \in K}$.

We have given that ${\displaystyle f}$ is linear and bijective with inverse map ${\displaystyle g}$. How can we use this to show the linearity of ${\displaystyle g}$? Since ${\displaystyle g}$ is the inverse map of ${\displaystyle f}$, we have:

{\displaystyle {\begin{aligned}&f\circ g=id_{W}{\text{, so }}f(g(w))=w{\text{ for all }}w\in W\end{aligned}}}

Together with the linearity of ${\displaystyle f}$, this gives us:

{\displaystyle {\begin{aligned}g(w+w')&=g(f(g(w))+f(g(w')))\\[0.3em]&{\color {OliveGreen}\left\downarrow \ {\text{ linearity of }}f\right.}\\[0.3em]&=g(f(g(w)+g(w')))=g(w)+g(w').\end{aligned}}}

In the same way we can proceed for homogeneity ${\displaystyle g(\lambda \cdot w)=\lambda \cdot g(w)}$.

Proof (The inverse map of a bijective linear map is again linear.)

For the inverse ${\displaystyle g}$ of ${\displaystyle f}$ , it holds that:

{\displaystyle {\begin{aligned}&f\circ g=id_{W};\\[0.3em]&g\circ f=id_{V}.\end{aligned}}}

So for every vector ${\displaystyle v\in V}$ and every vector ${\displaystyle w\in W}$

{\displaystyle {\begin{aligned}&(f\circ g)(w)=f(g(w))=w;\\[0.3em]&(g\circ f)(v)=g(f(v))=v.\end{aligned}}}

Proof step: ${\displaystyle g}$ is additive.

Let ${\displaystyle w}$ and ${\displaystyle w'\in W}$ be two vectors. Then we have:

{\displaystyle {\begin{aligned}&g(w+w')\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {g} {\text{ is the inverse of }}\operatorname {f} \right.}\\[0.3em]=\,&g(f(g(w))+f(g(w')))\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {f} {\text{ is linear }}\right.}\\[0.3em]=\,&g(f(g(w)+g(w'))\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {g} {\text{ is the inverse of }}\operatorname {f} \right.}\\[0.3em]=\,&g(w)+g(w').\end{aligned}}}

Thus the inverse function is additive.

Proof step: ${\displaystyle g}$ Is homogeneous.

Let ${\displaystyle w\in W}$ be a vector and ${\displaystyle \lambda \in K}$ is a scalar. Then we have that

{\displaystyle {\begin{aligned}&g(\lambda \cdot w)\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {g} {\text{ is the inverse of }}\operatorname {f} \right.}\\[0.3em]=\,&g(\lambda \cdot f(g(w)))\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {f} {\text{ is linear }}\right.}\\[0.3em]=\,&g(f(\lambda \cdot g(w)))\\[0.3em]&{\color {OliveGreen}\left\downarrow \ \operatorname {g} {\text{ is the inverse of }}\operatorname {f} \right.}\\[0.3em]=\,&\lambda \cdot g(w).\end{aligned}}}

Thus the inverse function is homogeneous.

Thus have shown that the inverse ${\displaystyle g}$ is also linear.

## Classifying Isomorphic Structures

### Bijections of Bases Generate an Isomorphism

In the alternative derivation , we used the intuition that an isomorphism is a linear map that "preserves bases". This means that bases are sent to bases and linear combinations are preserved. So, describing it a bit more formally, we considered the following:

We already know the following: If ${\displaystyle f\colon V\to W}$ is a linear map between two vector spaces and ${\displaystyle f}$ is an isomorphism, then ${\displaystyle f}$ maps bases of ${\displaystyle V}$ to bases of ${\displaystyle W}$.

But we don't know yet whether a linear map that sends a basis to a basis, is already an isomorphism. This statement indeed turns out to be true.

Theorem

Let ${\displaystyle K}$ be a field, ${\displaystyle V,W}$ two ${\displaystyle K}$-vector spaces, ${\displaystyle B_{V}}$ a basis of ${\displaystyle V}$ and ${\displaystyle f:V\to W}$ a linear map.

Then ${\displaystyle f}$ is an isomorphism if and only if ${\displaystyle B_{V}}$ is mapped by ${\displaystyle f}$ to a basis of ${\displaystyle W}$.

Proof

Proof step: ${\displaystyle \implies }$

Let ${\displaystyle f}$ be an isomorphism. Then ${\displaystyle f}$ is by definition both a monomorphism and an epimorphism.

We want to show that ${\displaystyle f}$ preserves bases. That is, the image of ${\displaystyle B_{V}}$ under ${\displaystyle f}$ is a linearly independent generator of ${\displaystyle W}$.

Proof step: ${\displaystyle f(B_{V})}$ is linearly independent

We know from the article on monomorphisms that those preserve linear independence. The set ${\displaystyle B_{V}}$ is a basis and thus linearly independent. So its image under ${\displaystyle f}$ is also linearly independent.

Proof step: ${\displaystyle f(B_{V})}$ is a generator of ${\displaystyle W}$

We know from the article on epimorphisms that those preserve generators . The set ${\displaystyle B_{V}}$ is a basis and hence a generator of ${\displaystyle V}$. So its image under ${\displaystyle f}$ is a generator of ${\displaystyle W}$.

Proof step: ${\displaystyle \Longleftarrow }$

${\displaystyle f}$ maps ${\displaystyle B_{V}}$ to a basis ${\displaystyle B_{W}}$ of ${\displaystyle W}$ .

Proof step: Injectivity

Since ${\displaystyle f}$ maps the linearly independent set ${\displaystyle B_{V}}$ to the linearly independent set ${\displaystyle B_{W}}$, ${\displaystyle f}$ preserves linear independence. From the article on monomorphisms we know that ${\displaystyle f}$ must thus be injective.

Proof step: Surjectivity

${\displaystyle f}$ maps the basis ${\displaystyle B_{V}}$, (which is in particular a generator), to the basis ${\displaystyle B_{W}}$ (which is also a generator). From the article on epimorphisms we know that ${\displaystyle f}$ must thus be surjective.

${\displaystyle f}$ is linear by premise. Together with injectivity and surjectivity it follows that ${\displaystyle f}$ is an isomorphism.

Theorem

Let ${\displaystyle V}$ and ${\displaystyle W}$ be two ${\displaystyle K}$-vector spaces with bases${\displaystyle B_{V}}$ and ${\displaystyle B_{W}}$. Let further ${\displaystyle h\colon B_{V}\to B_{W}}$ be a bijective mapping. Then there is exactly one isomorphism ${\displaystyle f\colon V\to W}$ with ${\displaystyle f|_{B_{V}}=h}$.

Proof

From the article about linear continuation we know that we can find a unique linear map ${\displaystyle f}$ with ${\displaystyle f(b)=h(b)\in B_{W}}$ for all ${\displaystyle b\in B_{V}}$. Thus, as required by the premise, ${\displaystyle f|_{B_{V}}=h}$.

We still have to show that the mapping ${\displaystyle f}$ is an isomorphism. By the previous theorem, we must show that ${\displaystyle f}$ maps a basis of ${\displaystyle V}$ to a basis of ${\displaystyle W}$. Now we have constructed ${\displaystyle f}$ exactly such that ${\displaystyle f|_{B_{V}}=h}$. That is, ${\displaystyle f}$ maps the basis ${\displaystyle B_{V}}$ to the basis ${\displaystyle B_{W}}$ since ${\displaystyle h}$ is bijective. So ${\displaystyle f}$ is an isomorphism.

If we have given a bijection between bases, then there is a nice description of the inverse of ${\displaystyle f}$: We know that ${\displaystyle f^{-1}}$ is characterized by the conditions ${\displaystyle f^{-1}\circ f=\operatorname {id} _{V}}$ and ${\displaystyle f\circ f^{-1}=\operatorname {id} _{W}}$. Further, the principle of linear continuation tells us that we need to know ${\displaystyle f^{-1}}$ only on a basis of ${\displaystyle W}$ to describe it completely. Now we have already chosen the basis ${\displaystyle B_{W}}$ of W. That is, we are interested in ${\displaystyle f^{-1}(b_{W})}$ for ${\displaystyle b_{W}\in B_{W}}$. Because ${\displaystyle h}$ is bijective, there is exactly one ${\displaystyle b_{V}\in B_{V}}$ with ${\displaystyle h(b_{V})=b_{W}}$. Therefore, we get ${\displaystyle f^{-1}(b_{W})=f^{-1}(f(b_{V}))=b_{V}}$ from the above conditions. Now how can we describe this element ${\displaystyle b_{V}}$ more precisely? ${\displaystyle b_{V}}$ is the unique preimage of ${\displaystyle b_{W}}$ under ${\displaystyle h}$. So ${\displaystyle b_{V}=h^{-1}(b_{W})}$. In other words, ${\displaystyle f^{-1}}$ is the linear map induced by ${\displaystyle h^{-1}}$ from ${\displaystyle W}$ to ${\displaystyle V}$.

### Classification of Finite Dimensional Vector Spaces

When are two finite-dimensional vector spaces isomorphic? If ${\displaystyle V}$ and ${\displaystyle W}$ are finite-dimensional vector spaces, then we have bases ${\displaystyle \{b_{1},\ldots ,b_{n}\}}$ of ${\displaystyle V}$ and ${\displaystyle \{c_{1},\ldots ,c_{m}\}}$ of ${\displaystyle W}$. From the previous theorem we know that an isomorphism is uniquely characterized by the bijection of the bases. When do we find a bijection between these two sets? Exactly when they have the same size, so ${\displaystyle n=m}$. Or in other words, if ${\displaystyle V}$ and ${\displaystyle W}$ have the same dimension:

Theorem (Finite dimensional vector spaces with the same dimension are isomorphic)

Let ${\displaystyle V,W}$ be finite-dimensional vector spaces. Then: ${\displaystyle \operatorname {dim} (V)=\operatorname {dim} (W)\Longleftrightarrow V\cong W}$

Proof (Finite dimensional vector spaces with the same dimension are isomorphic)

Proof step: ${\displaystyle \implies }$

Let ${\displaystyle \operatorname {dim} (V)=\operatorname {dim} (W)}$.

Two vector spaces are called isomorphic if there exists an isomorphism between them. We know that an isomorphism exists between vector spaces if we can find a bijective mapping between the bases of them. Since ${\displaystyle \operatorname {dim} (V)=\operatorname {dim} (W)}$, we find a bijective mapping between bases. Thus, there exists an isomorphism between ${\displaystyle V}$ and ${\displaystyle W}$.

Thus, ${\displaystyle V}$ and ${\displaystyle W}$ are isomorphic.

Proof step: ${\displaystyle \Longleftarrow }$

Let ${\displaystyle V\cong W}$.

Let ${\displaystyle f}$ be an isomorphism between ${\displaystyle V}$ and ${\displaystyle W}$. We know that an isomorphism maps bases to bases. That is, ${\displaystyle f(B_{V})}$ is a basis of ${\displaystyle W}$. In particular, since the mapping is an isomorphism, it is bijective. Thus ${\displaystyle |B_{V}|=|f(B_{V})|}$.

This implies ${\displaystyle \operatorname {dim} (V)=\operatorname {dim} (W)}$.

We have shown that all ${\displaystyle K}$-vector spaces of dimension ${\displaystyle n}$ are isomorphic. In particular, all such vector spaces are isomorphic to the vector space ${\displaystyle K^{n}}$. Because the ${\displaystyle K^{n}}$ is a well-describable model for a vector space, let us examine in more detail the isomorphism constructed in the last theorem.

Let ${\displaystyle V}$ be an ${\displaystyle n}$-dimensional ${\displaystyle K}$-vector space. We now follow the proof of the last theorem to understand the construction of the isomorphism. We use that bases of ${\displaystyle V}$ and of ${\displaystyle K^{n}}$ have the same size. For the isomorphism, we construct a bijection between a basis of ${\displaystyle V}$ and a basis of ${\displaystyle K^{n}}$. The space ${\displaystyle K^{n}}$ has as kind of "standard basis", given by the canonical basis ${\displaystyle \{e_{1},\dots ,e_{n}\}}$.

Following the proof of the last theorem, we see that we must choose a basis of ${\displaystyle V}$ and a basis of ${\displaystyle K^{n}}$. For ${\displaystyle K^{n}}$ we choose the standard basis ${\displaystyle E}$ and for ${\displaystyle V}$ we choose some basis ${\displaystyle B}$ of ${\displaystyle V}$. Next, we need a bijection ${\displaystyle h\colon E\to B}$ between the standard basis and the basis ${\displaystyle B}$. That is, we need to associate exactly one ${\displaystyle b\in B}$ with each ${\displaystyle e_{i}}$. We can thus name the images of ${\displaystyle e_{i}}$ as ${\displaystyle b_{i}=h(e_{i})}$. Because ${\displaystyle h}$ is bijective, we get ${\displaystyle B=\{b_{1},\dots ,b_{n}\}}$. In essence, we have used this to number the elements of B. Mathematically, numbering the elements of ${\displaystyle B}$ is the same as giving a bijection from ${\displaystyle E}$ to ${\displaystyle B}$, since we can simply map ${\displaystyle e_{i}}$ to the ${\displaystyle i}$-th element of ${\displaystyle B}$.

The principle of linear continuation now provides us with an isomorphism ${\displaystyle f\colon K^{n}\to V}$. By linear continuation, this isomorphism sends the vector ${\displaystyle (x_{1},\dots ,x_{n})^{T}\in K^{n}}$ to the element ${\displaystyle f((x_{1},\dots ,x_{n})^{T})=x_{1}\cdot b_{1}+\dots +x_{n}\cdot b_{n}}$.

Now what about the map that sends ${\displaystyle B}$ to ${\displaystyle E}$, i.e., the inverse map ${\displaystyle f^{-1}}$ of ${\displaystyle f}$?

We have already computed above what the mapping ${\displaystyle f^{-1}}$ looks like in this case. ${\displaystyle f^{-1}}$ is just the mapping induced by ${\displaystyle h^{-1}}$ via the principle of linear continuation. That is, for basis vectors, we know that ${\displaystyle f}$ maps ${\displaystyle b_{i}\in B}$ to ${\displaystyle h^{-1}(b_{i})=e_{i}}$. And where does it map a general vector ${\displaystyle v\in V}$? HEre, we use the principle of linear continuation: We write ${\displaystyle v}$ as a linear combination of our basis ${\displaystyle v=\lambda _{1}b_{1}+\dots +\lambda _{n}b_{n}}$. By linearity, the mapping ${\displaystyle f^{-1}}$ now sends ${\displaystyle v}$ to ${\displaystyle f^{-1}(v)=(\lambda _{1},\dots ,\lambda _{n})^{T}}$. In particular, the ${\displaystyle \lambda _{i}}$ describe where ${\displaystyle v}$ is located with respect to the basis vectors ${\displaystyle b_{i}}$. This is just like GPS coordinates, which tells you your position with respect to certain anchor points (there prime meridian and equator). Therefore, we can say that ${\displaystyle f^{-1}}$ sends each vector to its coordinates with respect to the basis ${\displaystyle B}$.

Definition (Coordinate mapping)

Let ${\displaystyle V}$ be a ${\displaystyle n}$-dimensional ${\displaystyle K}$-vector space and ${\displaystyle b_{1},\dots ,b_{n}}$ a basis of ${\displaystyle V}$. We define the isomorphism ${\displaystyle k_{b_{1},\dots ,b_{n}}\colon V\to K^{n}}$ as the continuation of the following bijection between the base ${\displaystyle b_{1},\dots ,b_{n}}$ and the standard basis of ${\displaystyle K^{n}}$:

${\displaystyle \{b_{1},\dots ,b_{n}\}\to \{e_{1},\dots ,e_{n}\};b_{i}\mapsto e_{i}}$

We call ${\displaystyle k_{b_{1},\dots ,b_{n}}}$ the coordinate mapping with respect to ${\displaystyle b_{1},\dots ,b_{n}}$.

We now want to investigate how many choices the construction of the coordinate map depends on.

Example (Coordinate mapping between the vector space of real quadratic polynomials and ${\displaystyle \mathbb {R} ^{3}}$)

We consider the two ${\displaystyle \mathbb {R} }$-vector spaces ${\displaystyle W:=\mathbb {R} ^{3}}$ and that of real polynomials of degree ${\displaystyle \leq 2:=P_{\leq 2}}$. The coordinate mapping then looks like this:

${\displaystyle k_{1,x,x^{2}}:P_{2}\to W}$, ${\displaystyle a_{0}+a_{1}x+a_{2}x^{2}\mapsto (a_{0},a_{1},a_{2})}$.

The coordinate mapping depends on the choice of the basis. If you have different bases, you get different mappings.

Example (Different bases create different coordinate mappings)

We consider the following two bases of ${\displaystyle \mathbb {R} ^{2}}$: ${\displaystyle B=\{(1,0)^{T},(0,1)^{T}\}}$ and ${\displaystyle C=\{(1,1)^{T},(1,-1)^{T}\}}$.

For ${\displaystyle x=(x_{1},x_{2})\in \mathbb {R} ^{2}}$ we have

${\displaystyle x=x_{1}\cdot {\begin{pmatrix}1\\0\end{pmatrix}}+x_{2}\cdot {\begin{pmatrix}0\\1\end{pmatrix}}}$

So the coordinate mapping with respect to ${\displaystyle B}$ looks like this

${\displaystyle k_{(1,0)^{T},(0,1)^{T}}\colon \mathbb {R} ^{2}\to \mathbb {R} ^{2},\quad {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}\mapsto {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}}$

For the base ${\displaystyle C}$ we have

${\displaystyle x={\frac {x_{1}+x_{2}}{2}}\cdot {\begin{pmatrix}1\\1\end{pmatrix}}+{\frac {x_{1}-x_{2}}{2}}\cdot {\begin{pmatrix}1\\-1\end{pmatrix}}}$

Thus, the coordinate mapping with respect to ${\displaystyle C}$ is.

${\displaystyle k_{(1,1)^{T},(1,-1)^{T}}\colon {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}\mapsto {\begin{pmatrix}{\frac {x_{1}+x_{2}}{2}}\\{\frac {x_{1}-x_{2}}{2}}\end{pmatrix}}.}$

These two mappings are not the same. For example

${\displaystyle k_{(1,0)^{T},(0,1)^{T}}\left({\begin{pmatrix}1\\3\end{pmatrix}}\right)={\begin{pmatrix}1\\3\end{pmatrix}}\neq {\begin{pmatrix}2\\-1\end{pmatrix}}=k_{(1,1)^{T},(1,-1)^{T}}\left({\begin{pmatrix}1\\3\end{pmatrix}}\right)}$

Even if we only change the numbering of the elements of a base, we already get different coordinate mappings.

Example (Different numbering of the basis result in different coordinate images)

We consider the standard basis ${\displaystyle B=\{(1,0)^{T},(0,1)^{T}\}}$ of ${\displaystyle \mathbb {R} ^{2}}$. We want to find out what the coordinate mappings ${\displaystyle k_{(1,0)^{T},(0,1)^{T}}}$ and ${\displaystyle k_{(0,1)^{T},(1,0)^{T}}}$ look like. For ${\displaystyle k_{(1,0)^{T},(0,1)^{T}}}$ we already know this:

${\displaystyle k_{(1,0)^{T},(0,1)^{T}}\colon \mathbb {R} ^{2}\to \mathbb {R} ^{2},\quad {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}\mapsto {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}}$

For ${\displaystyle x=(x_{1},x_{2})^{T}}$ we have

${\displaystyle x=x_{2}\cdot {\begin{pmatrix}0\\1\end{pmatrix}}+x_{1}\cdot {\begin{pmatrix}1\\0\end{pmatrix}}.}$

The construction of the coordinate mapping thus provides us with the following description

${\displaystyle k_{(0,1)^{T},(1,0)^{T}}\colon \mathbb {R} ^{2}\to \mathbb {R} ^{2},\quad {\begin{pmatrix}x_{1}\\x_{2}\end{pmatrix}}\mapsto {\begin{pmatrix}x_{2}\\x_{1}\end{pmatrix}}}$

These two mappings are different. For example

${\displaystyle k_{(1,0)^{T},(0,1)^{T}}\left({\begin{pmatrix}2\\1\end{pmatrix}}\right)={\begin{pmatrix}2\\1\end{pmatrix}}\neq {\begin{pmatrix}1\\2\end{pmatrix}}=k_{(0,1)^{T},(1,0)^{T}}\left({\begin{pmatrix}2\\1\end{pmatrix}}\right).}$

In order to speak of the coordinate mapping, we must also specify the order of the basis elements. A basis where we also specify the order of the basis elements is called an ordered basis.

Definition (Ordered basis)

Let ${\displaystyle K}$ be a field and ${\displaystyle V}$ a finite-dimensional ${\displaystyle K}$-vector space. Let ${\displaystyle B=(b_{1},\dots ,b_{n})\in V^{n}}$. Then we call ${\displaystyle B}$ an ordered basis of ${\displaystyle V}$ if ${\displaystyle \{b_{1},\dots ,b_{n}\}}$ is a basis of ${\displaystyle V}$.

With this notion we can simplify the notation of the coordinate mapping. If ${\displaystyle B=(b_{1},\dots ,b_{n})}$ is an ordered basis, we also denote the coordinate mapping ${\displaystyle k_{b_{1},\dots ,b_{n}}}$ as ${\displaystyle k_{B}}$.

We have now talked about a class of isomorphisms from ${\displaystyle V}$ to ${\displaystyle K^{n}}$. Are there any other isomorphisms from ${\displaystyle V}$ to ${\displaystyle K^{n}}$? That is, are there isomorphisms that are not coordinate mappings? In fact, every isomorphism from ${\displaystyle V}$ to ${\displaystyle K^{n}}$ is a coordinate mapping with respect to a proper basis.

Theorem (All isomorphisms ${\displaystyle V\to K^{n}}$ are coordinate mappings)

Let ${\displaystyle f\colon V\to K^{n}}$ be an isomorphism. Then there is exactly one ordered basis ${\displaystyle B}$ of ${\displaystyle V}$ such that ${\displaystyle f=k_{B}}$.

How to get to the proof? (All isomorphisms ${\displaystyle V\to K^{n}}$ are coordinate mappings)

We have constructed the coordinate mapping as an inverse mapping. For this we bijectively mapped the standard basis of ${\displaystyle K^{n}}$ to a basis of ${\displaystyle V}$. To reconstruct this basis, we need to consider the preimages of the standard basis under ${\displaystyle f}$. That is, we need ${\displaystyle B=f^{-1}(E)}$, which requires choosing an ordering of ${\displaystyle B}$. For instance, we may set ${\displaystyle b_{i}=f^{-1}(e_{i})}$. Now, we know that ${\displaystyle B}$ is a basis because ${\displaystyle f^{-1}}$ is an isomorphism. Further, we have just above applied the principle of linear continuation backwards, which told us that all of ${\displaystyle f^{-1}}$ is induced by only the bijection ${\displaystyle e_{i}\mapsto b_{i}}$. Further above, we have also seen that ${\displaystyle f}$ is already induced by the bijection ${\displaystyle b_{i}\mapsto e_{i}}$. But this gives exactly the coordinate mapping with respect to ${\displaystyle B}$.

Proof (All isomorphisms ${\displaystyle V\to K^{n}}$ are coordinate mappings)

We define ${\displaystyle b_{i}:=f^{-1}(e_{i})}$ for ${\displaystyle 1\leq i\leq n}$. Then ${\displaystyle \{b_{1},\dots ,b_{n}\}}$ is the image of the standard basis under the mapping ${\displaystyle f^{-1}}$. Since ${\displaystyle f^{-1}}$ is an isomorphism, it maps bases to bases. Thus ${\displaystyle b_{1},\ldots ,b_{n}}$ is a basis of ${\displaystyle V}$.

Define the ordered basis ${\displaystyle B:=(b_{1},\dots ,b_{n})}$. We now show ${\displaystyle f=k_{B}}$. For this it is sufficient to prove equality on the basis ${\displaystyle B}$, since ${\displaystyle f}$ and ${\displaystyle k_{B}}$ are linear. For any ${\displaystyle b_{i}}$ it holds that

${\displaystyle f(b_{i})=f(f^{-1}(e_{i}))=e_{i}=k_{B}(b_{i}).}$

So indeed, ${\displaystyle f=k_{B}}$.

## Examples of vector space isomorphisms

Example (Real polynomials of ${\displaystyle n}$-th degree and ${\displaystyle \mathbb {R} ^{n+1}}$)

For ${\displaystyle n=2}$ , we can establish an isomorphism between the space of polynomials of at most second degree ${\displaystyle \mathbb {R} [X]_{\leq 2}}$ and the space ${\displaystyle \mathbb {R} ^{3}}$.

We define the mapping ${\displaystyle f:\mathbb {R} [X]_{\leq 2}\to \mathbb {R} ^{3}}$ vis ${\displaystyle f(p)=(p(-1),\,p(0),\,p(1))^{T}}$.

Claim: ${\displaystyle f}$ is an isomorphism.

For this, we need to prove three things:

1. ${\displaystyle f}$ is a linear map ${\displaystyle f:\mathbb {R} [X]_{\leq 2}\to \mathbb {R} ^{3}}$
2. ${\displaystyle f}$ is injektive
3. ${\displaystyle f}$ is surjektive

Proof step: Linearity of ${\displaystyle f}$

Since ${\displaystyle f(p)}$ is defined for every polynomial ${\displaystyle p\in \mathbb {R} [X]_{\leq 2}}$ and has values in ${\displaystyle \mathbb {R} ^{3}}$, ${\displaystyle f}$ is well-defined as a mapping.

So we still have to prove that for ${\displaystyle p_{1},p_{2}\in \mathbb {R} [X]_{\leq 2}}$ and ${\displaystyle \lambda \in \mathbb {R} }$ it always holds that ${\displaystyle f(p_{1}+p_{2})=f(p_{1})+f(p_{2})}$ and ${\displaystyle f(\lambda p)=\lambda f(p)}$.

This is completely analogous to this calculation.

Proof step: Injectivity of ${\displaystyle f}$

Let ${\displaystyle p\in \mathbb {R} [X]_{\leq 2}}$ and ${\displaystyle f(p)=(0,0,0)^{T}}$.

This means that the polynomial of the highest second degree ${\displaystyle p}$ has three zeros:${\displaystyle p(-1)=p(0)=p(1)=0}$. It follows (e.g., with polynomial division) that we can write ${\displaystyle p}$ as ${\displaystyle p(x)=(x+1)\cdot (x-0)\cdot (x-1)\cdot q(x)}$, where ${\displaystyle q}$ is again a polynomial (or a constant, i.e., a zero-degree polynomial). But because the degree of ${\displaystyle p}$ is at most two, ${\displaystyle q}$ must be constant and equal to ${\displaystyle 0}$, and thus ${\displaystyle p}$ is then the zero polynomial, i.e. the zero vector of the vector space ${\displaystyle \mathbb {R} [X]_{\leq 2}}$.

Now, since the kernel of ${\displaystyle f}$ consists only of the zero vector, ${\displaystyle f}$ is injective.

Proof step: Surjectivity of ${\displaystyle f}$

In proving this assertion, we use polynomial interpolation in the Lagrangian form.

For this purpose we define three polynomials ${\displaystyle p_{1},p_{2},p_{3}\in \mathbb {R} [X]_{\leq 2}}$ via

{\displaystyle {\begin{aligned}p_{1}(x)&={\dfrac {(x-0)(x-1)}{(-1-0)(-1-1)}}\\p_{2}(x)&={\dfrac {(x-(-1))(x-1)}{(0-(-1))(0-1)}}\\p_{3}(x)&={\dfrac {(x-(-1))(x-0)}{(1-(-1))(1-0)}}\end{aligned}}}

${\displaystyle p_{1}}$ has zeros at ${\displaystyle x=0}$ and ${\displaystyle x=1}$, and the denominator is the numerator at the position ${\displaystyle x=-1}$. Hence ${\displaystyle p_{1}(-1)=1}$, since the numerator and denominator then contain the same number.

Quite analogously, ${\displaystyle p_{2}(-1)=p_{2}(1)=0}$ and ${\displaystyle p_{2}(0)=1}$ as well as ${\displaystyle p_{3}(-1)=p_{3}(0)=}$ and ${\displaystyle p_{3}(1)=1}$.

Now, if we have any vector ${\displaystyle x=(x_{1},x_{2},x_{3})^{T}\in \mathbb {R} ^{3}}$, then we define the polynomial ${\displaystyle p\in \mathbb {R} [X]_{\leq 2}}$ by

${\displaystyle p(x)=x_{1}\cdot p_{1}(x)+x_{2}\cdot p_{2}(x)+x_{3}\cdot p_{3}(x)}$

Then, ${\displaystyle p(-1)=x_{1}\cdot 1+x_{2}\cdot 0+x_{3}\cdot 0=x_{1}}$ ans analogously ${\displaystyle p(0)=x_{1}\cdot 0+x_{2}\cdot 1+x_{3}\cdot 0=x_{2}}$ as well as ${\displaystyle p(1)=x_{1}\cdot 0+x_{2}\cdot 0+x_{3}\cdot 1=x_{3}}$.

Thus we have shown that ${\displaystyle f}$ is surjective.

We also see that we can use this procedure for arbitrary degrees of polynomials and arbitrary points, as long as the number of points is equal to the maximum degree of the polynomials plus 1. We can also replace ${\displaystyle \mathbb {R} }$ everywhere by ${\displaystyle \mathbb {Q} }$ or ${\displaystyle \mathbb {C} }$ without the need to change anything in the proof.

Example (Convergent sequences modulo zero sequences)

Let ${\displaystyle K\in \{\mathbb {R} ,\mathbb {C} \}}$. We already know the vector spaces ${\displaystyle c,c_{0}\subseteq \omega }$ of convergent sequences and of zero sequences. We also know that ${\displaystyle c_{0}\subseteq c}$ is a subspace. Therefore, we can form the factor space ${\displaystyle V:=c/c_{0}}$.

In the following, we will show that ${\displaystyle V\cong K}$.

We define a mapping

{\displaystyle {\begin{aligned}\Phi \colon K&\to V\\\lambda &\mapsto [(\lambda )_{i\in \mathbb {N} }].\end{aligned}}}

So the image of ${\displaystyle \lambda \in K}$ under this map is the coset of the sequence which takes the constant value ${\displaystyle \lambda }$. This is convergent with limit ${\displaystyle \lambda }$. We have to show that ${\displaystyle \Phi }$ is linear and bijective.

Proof step: Linearity of ${\displaystyle \Phi }$

We need to show additivity and homogeneity of ${\displaystyle \Phi }$ in order to get linearity.

Proof step: Additivity of ${\displaystyle \Phi }$

Let ${\displaystyle \lambda ,\mu \in K}$. Then

{\displaystyle {\begin{aligned}\Phi (\lambda )+\Phi (\mu )&=[(\lambda )_{i\in \mathbb {N} }]+[(\mu )_{i\in \mathbb {N} }]\\&=[(\lambda )_{i\in \mathbb {N} }+(\mu )_{i\in \mathbb {N} }]\\&=[(\lambda +\mu )_{i\in \mathbb {N} }]\\&=\Phi (\lambda +\mu ).\end{aligned}}}

Proof step: Homogeneity of ${\displaystyle \Phi }$

Let ${\displaystyle \lambda ,\kappa \in K}$. We consider ${\displaystyle \lambda }$ as a vector of the ${\displaystyle K}$-vector space ${\displaystyle K}$ and ${\displaystyle \kappa }$ as a scalar. Then

{\displaystyle {\begin{aligned}\Phi (\kappa \cdot \lambda )&=[(\kappa \cdot \lambda )_{i\in \mathbb {N} }]\\&=[\kappa \cdot (\lambda )_{i\in \mathbb {N} }]\\&=\kappa \cdot [(\lambda )_{i\in \mathbb {N} }]\\&=\kappa \cdot \Phi (\lambda ).\end{aligned}}}

Proof step: Injectivity of ${\displaystyle \Phi }$

We need to show that ${\displaystyle \ker(\Phi )=0}$ . So let ${\displaystyle \lambda \in \ker(\Phi )}$. That means, ${\displaystyle 0=\Phi (\lambda )=[(\lambda )_{i\in \mathbb {N} }]}$. Thus, we have ${\displaystyle (\lambda )_{i\in \mathbb {N} }\in c_{0}}$.

Therefore, ${\displaystyle \lambda =\lim _{i\to \infty }\lambda =0}$ which establishes the assertion.

Proof step: Surjectivity of ${\displaystyle \Phi }$

Let ${\displaystyle x=[(x_{i})_{i\in \mathbb {N} }]\in V}$ and let ${\displaystyle \lambda =\lim _{i\to \infty }x_{i}}$. We set ${\displaystyle y=[(\lambda )_{i\in \mathbb {N} }]=\Phi (\lambda )}$. Then ${\displaystyle \lim _{i\to \infty }|x_{i}-y_{i}|=\lim _{i\to \infty }|x_{i}-\lambda |=0}$.

Therefore ${\displaystyle (x_{i})_{i\in \mathbb {N} }-(\lambda )_{i\in \mathbb {N} }\in c_{0}}$, which implies ${\displaystyle x=y=\Phi (\lambda )}$. This establishes surjectivity.

Example (The isomorphism theorem)

One of the most important examples is the isomorphism between the image space of a linear map ${\displaystyle L:V\to W}$ and the quotient space ${\displaystyle V/\ker L}$.

All this is described here.

## Exercises

Exercise (complex ${\displaystyle \mathbb {R} }$-vector spaces)

Let ${\displaystyle V}$ be a finite-dimensional ${\displaystyle \mathbb {C} }$-vector space. Show that ${\displaystyle V\cong \mathbb {R} ^{2\operatorname {dim} _{\mathbb {C} }(V)}}$ (interpreted as ${\displaystyle \mathbb {R} }$-vector spaces).

Solution (complex ${\displaystyle \mathbb {R} }$-vector spaces)

Set ${\displaystyle n:=\operatorname {dim} _{\mathbb {C} }(V)}$. We choose a ${\displaystyle \mathbb {C} }$ basis ${\displaystyle {\mathcal {B}}=\{b_{1},\dots ,b_{n}\}}$ of ${\displaystyle V}$. Define ${\displaystyle c_{j}:=i\cdot b_{j}}$ for all ${\displaystyle 1\leq j\leq n}$.

We have to show that ${\displaystyle \{b_{1},\dots ,b_{n},c_{1},\dots ,c_{n}\}}$ is an ${\displaystyle \mathbb {R} }$-basis of ${\displaystyle V}$. Then, ${\displaystyle \operatorname {dim} _{\mathbb {R} }(V)=2n=\operatorname {dim} _{\mathbb {R} }(\mathbb {R} ^{2n})}$. According to a theorem above, we have ${\displaystyle V\cong \mathbb {R} ^{2n}}$ as ${\displaystyle \mathbb {R} }$-vector spaces.

We now show ${\displaystyle \mathbb {R} }$-linear independence.

Proof step: ${\displaystyle \{b_{1},\dots ,b_{n},c_{1},\dots ,c_{n}\}}$ is ${\displaystyle \mathbb {R} }$-linearly independent

Let ${\displaystyle \beta _{1},\dots ,\beta _{n},\gamma _{1},\dots ,\gamma _{n}\in \mathbb {R} }$ and assume that ${\displaystyle \sum _{j=1}^{n}\beta _{j}\cdot b_{j}+\sum _{j=1}^{n}\gamma _{j}\cdot c_{j}=0}$. We substitute the definition for ${\displaystyle c_{j}}$, conclude the sums and obtain ${\displaystyle \sum _{j=1}^{n}(\beta _{j}+i\cdot \gamma _{j})\cdot b_{j}=0}$. By ${\displaystyle \mathbb {C} }$-linear independence of ${\displaystyle b_{j}}$ we obtain ${\displaystyle \beta _{j}+i\cdot \gamma _{j}=0}$ for all ${\displaystyle j\in \{1,\dots ,n\}}$. Thus, ${\displaystyle \beta _{j}=\gamma _{j}=0}$ for all ${\displaystyle j\in \{1,\dots ,n\}}$. This establishes the ${\displaystyle \mathbb {R} }$-linear independence.

Now only one step is missing:

Proof step: ${\displaystyle \{b_{1},\dots ,b_{n},c_{1},\dots ,c_{n}\}}$ is a generator with respect to ${\displaystyle \mathbb {R} }$

Let ${\displaystyle v\in V}$ be arbitrary.

Since ${\displaystyle {\mathcal {B}}}$ is a ${\displaystyle \mathbb {C} }$-basis of ${\displaystyle V}$ , we can find some ${\displaystyle \lambda _{1},\dots ,\lambda _{n}\in \mathbb {C} }$, such that ${\displaystyle v=\sum _{j=1}^{n}\lambda _{j}\cdot b_{j}}$ . We write ${\displaystyle \lambda _{j}=\beta _{j}+\gamma _{j}i}$ with ${\displaystyle \beta _{j},\gamma _{j}\in \mathbb {R} }$ for all ${\displaystyle j}$. Then we obtain

{\displaystyle {\begin{aligned}v&=\sum _{j=1}^{n}\lambda _{j}\cdot b_{j}\\&=\sum _{j=1}^{n}(\beta _{j}+\gamma _{j}i)\cdot b_{j}\\&=\sum _{j=1}^{n}(\beta _{j}\cdot b_{j}+\gamma _{j}\cdot (i\cdot b_{j}))\\&=\sum _{j=1}^{n}(\beta _{j}\cdot b_{j}+\gamma _{j}\cdot c_{j})\\&=\sum _{j=1}^{n}\beta _{j}\cdot b_{j}+\sum _{j=1}^{n}\gamma _{j}\cdot c_{j}.\end{aligned}}}

So ${\displaystyle v}$ is inside the ${\displaystyle \mathbb {R} }$-span of ${\displaystyle \{b_{1},\dots ,b_{n},c_{1},\dots ,c_{n}\}}$. This establishes the assertion.

Exercise (Isomorphic coordinate spaces)

Let ${\displaystyle K}$ be a field and consider ${\displaystyle n,m\in \mathbb {N} _{0}}$. Prove that ${\displaystyle K^{n}\cong K^{m}}$ holds if and only if ${\displaystyle m=n}$.

Solution (Isomorphic coordinate spaces)

We know that ${\displaystyle \operatorname {dim} (K^{k})=k}$ for all ${\displaystyle k\in \mathbb {N} _{0}}$. We use the theorem above, which states that finite-dimensional vector spaces are isomorphic exactly if their dimensions coincide. So ${\displaystyle K^{n}\cong K^{m}}$ holds if and only if ${\displaystyle n=\operatorname {dim} (K^{n})=\operatorname {dim} (K^{m})=m}$ .

Exercise (Isomorphism criteria for endomorphisms)

Let ${\displaystyle K}$ be a field, ${\displaystyle V}$ a finite-dimensional ${\displaystyle K}$-vector space and ${\displaystyle f:V\to V}$ a ${\displaystyle K}$-linear map. Prove that the following three statements are equivalent:

(i) ${\displaystyle f}$ is an isomorphism.

(ii) ${\displaystyle f}$ is injective.

(iii) ${\displaystyle f}$ is surjective.

(Note: For this task, it may be helpful to know the terms kernel and image of a linear map. Using the dimension theorem, this exercise becomes much easier. However, we give a solution here, which works without the dimension theorem).

Solution (Isomorphism criteria for endomorphisms)

(i)${\displaystyle \implies }$(ii) and (iii): According to the definition of an isomorphism, ${\displaystyle f}$ is bijective, i.e. injective and surjective. Therefore (ii) and (iii) hold.

(ii)${\displaystyle \implies }$(i): Let ${\displaystyle f}$ be an injective mapping. We need to show that ${\displaystyle f}$ is also surjective. The image ${\displaystyle \mathrm {im} (f):=\{f(v)~|~v\in V\}}$ of ${\displaystyle f}$ is a subspace of ${\displaystyle V}$. This can be verified by calculation. We now define a mapping ${\displaystyle f'}$ that does the same thing as ${\displaystyle f}$, except that it will be surjective by definition. This mapping is defined as follows:

{\displaystyle {\begin{aligned}f':V&\to \mathrm {im} (f)\\v&\mapsto f(v)\end{aligned}}}

The surjectivity comes from the fact that every element ${\displaystyle w\in \mathrm {im} (f)}$ can be written as ${\displaystyle w=f(v')}$, for a suitable ${\displaystyle v'\in V}$. Moreover, the mapping ${\displaystyle f'}$ is injective and linear. This is because ${\displaystyle f}$ already has these two properties. So ${\displaystyle V}$ and ${\displaystyle \mathrm {im} (f)}$ are isomorphic. Therefore, ${\displaystyle V}$ and ${\displaystyle \mathrm {im} (f)}$ have the same finite dimension. Since ${\displaystyle \mathrm {im} (f)}$ is a subspace of ${\displaystyle V}$, ${\displaystyle \mathrm {im} (f)=V}$ holds. This can be seen by choosing a basis in ${\displaystyle \mathrm {im} (f)}$, for instance the basis given by the vectors ${\displaystyle v_{1},\dots ,v_{n}\in \mathrm {im} (f)}$. These ${\displaystyle v_{1},\dots ,v_{n}}$ are also linearly independent in ${\displaystyle V}$, since ${\displaystyle \mathrm {im} (f)\subseteq V}$. And since ${\displaystyle V}$ and ${\displaystyle \mathrm {im} (f)}$ have the same dimension, the ${\displaystyle v_{1},\dots ,v_{n}}$ are also a basis in ${\displaystyle V}$. So the two vector spaces ${\displaystyle V}$ and ${\displaystyle \mathrm {im} (f)}$ must now be the same, because all elements from them are ${\displaystyle K}$-linear combinations formed with the ${\displaystyle v_{1},\dots ,v_{n}}$. Thus we have shown that ${\displaystyle f}$ is surjective.

(iii)${\displaystyle \implies }$(i): Now suppose ${\displaystyle f}$ is surjective. We need to show that ${\displaystyle f}$ is also injective. Let ${\displaystyle \mathrm {ker} (f):=\{v\in V~|~f(v)=0\}}$ be the kernel of the mapping ${\displaystyle f}$. You may convince yourself by calculation, that this kernel is a subspace of ${\displaystyle V}$. Let ${\displaystyle v_{1},\dots ,v_{k}}$ be a basis of ${\displaystyle \mathrm {ker} (f)}$. We can complete this (small) basis to a (large) basis of ${\displaystyle V}$, by including the additional vectors ${\displaystyle v_{k+1},\dots ,v_{n}}$. We will now show that ${\displaystyle f(v_{k+1}),\dots ,f(v_{n})}$ are linearly independent. So let coefficients ${\displaystyle \lambda _{k+1},\dots ,\lambda _{n}\in K}$ be given such that

{\displaystyle {\begin{aligned}\lambda _{k+1}f(v_{k+1})+\dots +\lambda _{n}f(v_{n})=0.\end{aligned}}}

By linearity of ${\displaystyle f}$ we conclude: ${\displaystyle f(\lambda _{k+1}v_{k+1}+\dots \lambda _{n}v_{n})=0}$. This means that the linear combination

{\displaystyle {\begin{aligned}\lambda _{k+1}v_{k+1}+\dots +\lambda _{n}v_{n}\end{aligned}}}

is in the kernel of ${\displaystyle f}$. But we already know a basis of ${\displaystyle \mathrm {ker} (f)}$. Therefore there are coefficients ${\displaystyle \lambda _{1},\dots ,\lambda _{k}\in K}$, such that

{\displaystyle {\begin{aligned}\lambda _{k+1}v_{k+1}+\dots +\lambda _{n}v_{n}=\lambda _{1}v_{1}+\dots +\lambda _{k}v_{k}.\end{aligned}}}

Because of the linear independence of ${\displaystyle v_{1},\dots ,v_{n}}$ it now follows that ${\displaystyle \lambda _{1},\dots ,\lambda _{n}=0}$. Therefore, the ${\displaystyle f(v_{k+1}),\dots ,f(v_{n})}$ are linearly independent. Next, we will show that these vectors also form a basis of ${\displaystyle V}$. To do this, we show that each vector in ${\displaystyle V}$ can be written as a linear combination of the ${\displaystyle f(v_{k+1}),\dots ,f(v_{n})}$. Let ${\displaystyle w\in V}$. Because of the surjectivity of ${\displaystyle f}$, there is a ${\displaystyle v\in V}$, with ${\displaystyle w=f(v)}$. Since the ${\displaystyle v_{1},\dots ,v_{n}}$ form a basis of ${\displaystyle V}$, there are coefficients ${\displaystyle \lambda _{1},\dots ,\lambda _{n}\in K}$ such that

{\displaystyle {\begin{aligned}v=\lambda _{1}v_{1}+\dots +\lambda _{n}v_{n}\end{aligned}}}

If we now apply ${\displaystyle f}$ to this equation, we get:

{\displaystyle {\begin{aligned}w=f(v)=\lambda _{1}\underbrace {f(v_{1})} _{=0}+\dots +\lambda _{k}\underbrace {f(v_{k})} _{=0}+\lambda _{k+1}f(v_{k+1})+\dots +\lambda _{n}f(v_{n}).\end{aligned}}}

Here we used the linearity of ${\displaystyle f}$. Since the first ${\displaystyle k}$ elements of our basis are in the kernel, their images are ${\displaystyle 0}$. So we get the desired representation of ${\displaystyle w}$:

{\displaystyle {\begin{aligned}w=f(v)=\lambda _{k+1}f(v_{k+1})+\dots +\lambda _{n}f(v_{n}).\end{aligned}}}

Thus we have shown that ${\displaystyle f(v_{k+1}),\dots ,f(v_{n})}$ forms a linearly independent generator of ${\displaystyle V}$. So these vectors form a basis of ${\displaystyle V}$. Now if ${\displaystyle k}$ were not ${\displaystyle 0}$, two finite bases in ${\displaystyle V}$ would not contain equally many elements. This cannot be the case. Therefore, ${\displaystyle k=0}$, so ${\displaystyle \mathrm {ker} (f)}$ is the trivial vector space and ${\displaystyle f}$ is indeed injective.

Exercise (Function spaces)

Let ${\displaystyle X}$ be a finite set with ${\displaystyle n\in \mathbb {N} }$ elements and let ${\displaystyle K}$ be a field. We have seen that the set of functions from ${\displaystyle X}$ to ${\displaystyle K}$ forms a ${\displaystyle K}$-vector space, denoted by ${\displaystyle \operatorname {Fun} (X,K)}$. Show that ${\displaystyle \operatorname {Fun} (X,K)\cong K^{n}}$.

Solution (Function spaces)

We already know according to a theorem above that two finite dimensional vector spaces are isomorphic exactly if they have the same dimension. So we just need to show that ${\displaystyle \operatorname {dim} (\operatorname {fun} (X,K))=n=\operatorname {dim} (K^{n})}$ holds.

To show this, we first need a basis of ${\displaystyle \operatorname {Fun} (X,K)}$. For this, let ${\displaystyle x_{1},\dots ,x_{n}}$ be the elements of the set ${\displaystyle X}$. We define ${\displaystyle f_{1},\dots ,f_{n}\in \operatorname {Fun} (X,K)}$ by

${\displaystyle f_{j}(x_{i}):=\delta _{i,j}={\begin{cases}1,{\text{ for }}i=j\\0,{\text{ for }}i\neq j.\end{cases}}}$

We now show that the functions ${\displaystyle f_{1},\dots ,f_{n}}$ indeed form a basis of ${\displaystyle \operatorname {Fun} (X,K)}$.

Proof step: ${\displaystyle f_{1},\dots ,f_{n}}$ are linearly independent

Let ${\displaystyle \alpha _{1},\dots ,\alpha _{n}\in K}$ with ${\displaystyle \sum _{k=1}^{n}\alpha _{k}\cdot f_{k}=0}$ being the zero function. If we apply this function to any ${\displaystyle x_{j}}$ with ${\displaystyle j\in \{1,\dots ,n\}}$, then we obtain: ${\displaystyle \sum _{k=1}^{n}\alpha _{k}\cdot f_{k}(x_{j})=0}$. By definition of ${\displaystyle f_{1},\dots ,f_{n}}$ it follows that

${\displaystyle 0=\sum _{k=1}^{n}\alpha _{k}f_{k}(x_{j})=\alpha _{j}\cdot f_{j}(x_{j})=\alpha _{j}\cdot 1=\alpha _{j}}$.

Since ${\displaystyle j}$ was arbitrary and ${\displaystyle \sum _{k=1}^{n}\alpha _{k}\cdot f_{k}(x_{j})=0}$ must hold for all ${\displaystyle x_{j}\in X}$, it follows that ${\displaystyle \alpha _{1}=\dots =\alpha _{n}=0}$. So we have shown that ${\displaystyle f_{1},\dots ,f_{n}}$ are linearly independent.

Proof step: ${\displaystyle f_{1},\dots ,f_{n}}$ generate ${\displaystyle \operatorname {Fun} (X,K)}$

Let ${\displaystyle g\in \operatorname {Fun} (X,K)}$ be arbitrary. We now want to write ${\displaystyle g}$ as a linear combination of ${\displaystyle f_{1},\dots ,f_{n}}$. For this we show ${\displaystyle g=\sum _{j=1}^{n}g(x_{j})\cdot f_{j}}$, i.e., ${\displaystyle g}$ is a linear combination of ${\displaystyle f_{1},\dots ,f_{n}}$ with coefficients ${\displaystyle g(x_{1}),\dots ,g(x_{n})\in K}$. We now verify that ${\displaystyle g(x_{i})=\sum _{j=1}^{n}g(x_{j})\cdot f_{j}(x_{i})}$ for all ${\displaystyle i\in \{1,\dots ,n\}}$. Let ${\displaystyle i\in \{1,\dots ,n\}}$ be arbitrary. By definition of ${\displaystyle f_{1},\dots ,f_{n}}$ we obtain:

${\displaystyle \sum _{j=1}^{n}g(x_{j})\cdot f_{j}(x_{i})=g(x_{i})\cdot f_{i}(x_{i})=g(x_{i})\cdot 1=g(x_{i})}$.

Since equality holds for all ${\displaystyle i}$, the functions agree at every point and are therefore identical. So we have shown that ${\displaystyle f_{1},\dots ,f_{n}}$ generate ${\displaystyle \operatorname {Fun} (X,K)}$.

Thus we have proved that ${\displaystyle f_{1},\dots ,f_{n}}$ is a basis of ${\displaystyle \operatorname {Fun} (X,K)}$. Since we have ${\displaystyle n}$ basis elements of ${\displaystyle \operatorname {Fun} (X,K)}$, it follows that ${\displaystyle \operatorname {dim} (\operatorname {Fun} (X,K))=n=\operatorname {dim} (K^{n})}$.

Hint

If ${\displaystyle X}$ is an infinite set, then ${\displaystyle \operatorname {Fun} (X,K)}$ is infinite-dimensional. In the special case ${\displaystyle X=\mathbb {N} }$, ${\displaystyle \operatorname {Fun} (X,K)}$ is isomorphic to the sequence space ${\displaystyle \omega }$.