Linear maps are special maps between vector spaces that are compatible with the vector space structure. They are one of the most important concepts of linear algebra and have numerous applications in science and technology.
We have learned about the structure of vector spaces and studied various properties of them. Now we want to consider not only isolated vector spaces, but also maps between them. Some of these maps fit well with the underlying vector space structure and are therefore called linear maps or vector space homomorphisms. They are a generalization of linear functions through the origin in one dimension, whose graphs are lines (hence the name).
It is a typical approach in algebra to study maps that preserve the structure of an algebraic object, such as a vector space. For many algebraic objects such as groups, rings or fields, one often studies the corresponding structure-preserving maps between the respective algebraic structures - group homomorphisms, ring homomorphisms and field homomorphisms. For vector spaces, the structure-preserving maps are the linear maps (= vector space homomorphisms).
So let and be two vector spaces. When is a map structure-preserving or well compatible with the underlying vector space structures in and ? For this, let's repeat what the vector space structure is all about: They basically allow for two operations:
Addition of vectors: two vectors can be added, in a similar way to how numbers are added.
Scalar multiplication: vectors with a scaling factor (which is an element of the field) can be scaled. That means: compressed, stretched or mirrored.
Let's start with of addition of vectors: when is a function compatible with the additions and on the respective vector spaces and ? The most natural definition is the following:
A map is compatible with the addition if a sum is preserved by the map. Meaning, if is a sum within the vector space , then the images of , and , which are situated in vector space , also form a corresponding sum:
Thus, a map compatible with addition satisfies for all the implication:
This implication can be summarized in one equation by substituting the premise into the second equation. It thus suffices to require for all that:
This equation describes the first characteristic property of the linear map, namely "being compatible with vector addition". We can visualize it well for maps . A map is compatible with addition if and only if the triangle given by the vectors , and is preserved under applying the map. That means, also the three vectors , and hive to form a triangle:
If is not compatible with addition, there are vectors and with . The triangle generated by , and is then not preserved, because the triangle side of the initial triangle is not mapped to the triangle side in the target space:
Compatibility with scalar multiplication[Bearbeiten]
Analogously, we can naturally define that a map is compatible with scalar multiplication if and only if it is preserved by the map. So it should hold for all and for all scalars that
Note that is a scalar and not a vector and thus is not changed by the map under consideration. In other words, it can be "pulled out of the bracket". This move is only allowed if both vector spaces have the same underlying field. Both the domain of definition and the range of values must be vector spaces over the same .
Linear maps thus preserve scalings. From one may conclude . For the case where , straight lines of the form are mapped to the straight line . The above implication can be summarized in an equation. For all and , we require that:
For maps this means that a scaled vector is mapped to the correspondingly scaled version of the image vector:
If a map is not compatible with scalar multiplication, there is a vector and a scaling factor such that :
A linear map is a special map between vector spaces that is compatible with the structure of the underlying vector spaces. In particular, this means that a linear map has the following two characteristic properties:
compatibility with addition:.
compatibility with scalar multiplication:
The compatibility with addition is called additivity and the compatibility with scalar multiplication is called homogeneity.
Now let be a map between these vector spaces. We call a linear map from to if the following two properties are satisfied:
additivity: For all we have that
homogeneity: For all and we have that
If it's clear from the context, in the future we'll also just write "" instead of and . Similarly, "" is often used instead of and are used. Sometimes the dot for scalar multiplication is completely omitted.
In the literature, the term vector space homomorphism or homomorphism for short is also used as a synonym for the term linear map. The ancient Greek word homós stands for equal, morphé stands for shape. Literally translated, a vector space homomorphism is a map between vector spaces, which leaves the "shape" of the vector spaces invariant.
The characteristic equations of the linear map are and . What do these two properties intuitively mean? According to the additivity property, it doesn't matter whether you first add and and then map them, or whether you first map both vectors and then add them. Both ways lead to the same result:
What does the homogeneity property mean? Regardless of whether you first scale by and then map it or first map the vector and then scale it by , the result is the same:
The characteristic properties of linear maps mean that the orders of function mapping and vector space operations do not matter.
Charakterization: linear combinations are mapped to linear combinations[Bearbeiten]
Besides the defining property that linear maps get along well with the underlying vector space structure, linear maps can also be characterized by the following property:
This is an important property because linear combinations are used to define important structures on vector spaces such as the linear independence or having generators. Also the definition of the basis relies on the notion of linear combination. The connection to linear combinations can be seen by looking at the two characteristic equations of linear maps:
We can apply the two formulas above step-by-step to a linear combination like for vectors and from . This allows us to "get the linear combination out of the bracket":
The linear combination is mapped by to and thus keeps its structure. The situation is similar for other linear combinations. For by the property sums "can be pulled out of the bracket" and by the property scalar multiplications "can be pulled out of the bracket". We thus obtain the following alternative characterization of the linear map: linear combinations are mapped to linear combinations.
Our first example is a stretch by the factor in -direction in the plane . Here, every vector is mapped to . The following figure shows this map for . The -coordinate remains the same and the -coordinate is doubled:
Now let's see if this map is compatible with addition. So let's take two vectors and , sum them and then stretch them in -direction. The result is the same as if we first stretch both vectors in -direction and then add them:
This can also be shown mathematically. Our map is the function . We can now check the property :
Now let's check the compatibility with scalar multiplication. The following figure shows that it doesn't matter if the vector is first scaled by a factor of and then stretched in -direction or first stretched in -direction and then scaled by :
This can also be shown formally: For and we have that
In the following, we consider a rotation of the plane by the angle (measured counter-clockwise) with the origin as center of rotation. Thus, it is a map that assigns to every vector the vector rotated by the angle :
Rotating a vector by the angle
Let us now convince ourselves that is indeed a linear map. To do this, we need to show that:
is additive: for all , we have .
is homogeneous: For all and we have .
First, we check additivity, that is, the equation . If we add two vectors and then rotate their sum by the angle , the same vector should come out, as if we first rotate the vectors by the angle and then add the rotated vectors and . This can be visualized by the following two videos:
Rotate first, then add
Add first, then rotate
Now we come to homogeneity: . If we first stretch a vector by a factor and then rotate the result by the angle , we should get the same vector as if we first rotate the by an angle and then scale the result by the factor . This is again visualized by two videos:
Rotate first, then scale
Scale first, then rotate
Thus, rotations in are indeed linear maps.
Linear maps between vector spaces of different dimension[Bearbeiten]
An example of a linear map between two vector spaces with different dimensions is the following projection of the space onto the plane :
We now check whether the vector addition is preserved. That means, for vectors we need that
Linear maps are used in almost all technological fields. Here is just a very tiny collection of some examples:
In order to make predictions or control machines, complicated functions are often approximated by linear ones (regression). Mainly because linear maps are easy to handle.
The best known case where linear maps make our lives easier are computer graphics. Any scaling of a photo or graphic is a linear map. Even different screen resolutions ended up being linear maps.
Search engines use page ranks of a website to sort their search results. our "Serlo-page", also gets a ranking this way. To determine the page rank, a so-called Markov chain is used, which is a somewhat more sophisticated linear map.
A linear map, also called vector space homomorphism, preserves the structure of the vector space. This is shown in the following properties of a linear mapping :
The zero vector is mapped to the zero vector: .
Inverses are mapped to inverses: .
Linear combinations are mapped to linear combinations.
Compositions of linear maps are again linear
Images of subspaces are subspaces
The image of a span is the span of the individual image vectors: ( is supposed to be an arbitrary set)
Relation to linear functions and affine maps[Bearbeiten]
Linear functions in one dimension take the form with . They are only linear maps in some cases, namely for . As an example, for and :
Maps are in fact linear, if and only if , i.e., the map takes the form with . The functions of the form are called affine-linear maps or simply affine maps: They are the sum of a linear map and a constant translational term . Every linear map is affine-linear, but not the other way round!
However, affine maps still map straight lines to straight lines and preserve parallel lines and ratios of distances.
We can always decompose an affine map into a linear map and a translation . We have that also . Because the translations are easy to describe, the linear part is usually more interesting. In the theory we therefore only look at the linear part.
We believe that anyone can understand higher level maths - with the right explanation! That's why we - a group of passionate students and teachers of mathematics - are translating our legendary co-created textbook for real analysis (“Analysis 1”) from German into English. If you know German and want to help us to make higher level maths accessible to everyone, then we look forward to hearing from you at firstname.lastname@example.org.
Feedback? Do you want to join?
If you have questions concerning the content, or didn't understand something, the feel free to contact us! We would love to answer your questions! Also we are thankful for critics and/or comments!
If you share our vision to explain university math in an comprehensible way, then contact us under:
This article is licensed under the free license CC-BY-SA 3.0. With that you can use it, modify it or share it freely, as long as you name „Serlo“ as source and put you changes under the same CC-BY-SA 3.0 oder an compatible license. On the page „Kopier uns!“ we explain you what you have to pay attention to, when using our texts, picture or videos.