LaTex2Web logo

LaTeX2Web, a web authoring and publishing system

If you see this, something is wrong

Collapse and expand sections

To get acquainted with the document, the best thing to do is to select the "Collapse all sections" item from the "View" menu. This will leave visible only the titles of the top-level sections.

Clicking on a section title toggles the visibility of the section content. If you have collapsed all of the sections, this will let you discover the document progressively, from the top-level sections to the lower-level ones.

Cross-references and related material

Generally speaking, anything that is blue is clickable.

Clicking on a reference link (like an equation number, for instance) will display the reference as close as possible, without breaking the layout. Clicking on the displayed content or on the reference link hides the content. This is recursive: if the content includes a reference, clicking on it will have the same effect. These "links" are not necessarily numbers, as it is possible in LaTeX2Web to use full text for a reference.

Clicking on a bibliographical reference (i.e., a number within brackets) will display the reference.

Speech bubbles indicate a footnote. Click on the bubble to reveal the footnote (there is no page in a web document, so footnotes are placed inside the text flow). Acronyms work the same way as footnotes, except that you have the acronym instead of the speech bubble.

Discussions

By default, discussions are open in a document. Click on the discussion button below to reveal the discussion thread. However, you must be registered to participate in the discussion.

If a thread has been initialized, you can reply to it. Any modification to any comment, or a reply to it, in the discussion is signified by email to the owner of the document and to the author of the comment.

Table of contents

First published on Saturday, Jul 6, 2024 and last modified on Thursday, Apr 10, 2025 by François Chaplais.

The Linear Mappings and their Matrices

Fabienne Chaplais Mathedu SAS

1 Introduction

We will discover here the linear mappings in the euclidean plane, that preserve the addtion and scalar multiplication of vectors.

A very important fact with the linear mappings is that they behave similarly as their matrices in the canonical basis, making them quite easy to handle.

2 The Linear Mappings, Definition and Examples

2.1 Definition of the Linear Mappings in \( \mathbb{P}\)

Definition 1

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a mapping in the euclidean plane. Then is said to be a linear mapping in if and only if:

  • it preserves the addition of vectors, that is \( \forall\;(\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^2,\; f(\overrightarrow{u}+\overrightarrow{v}) =f(\overrightarrow{u})+f(\overrightarrow{v})\)

  • and it preserves the multiplication of a vector by a scalar, that is \( \forall\;(\lambda,\overrightarrow{u})\in\mathbb{R}\times\mathbb{P},\; f(\lambda\overrightarrow{u})=\lambda f(\overrightarrow{u})\)

2.2 Properties of the Linear Mappings in \( \mathbb{P}\)

Theorem 1

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane.

Then the following assertions hold.

  1. It keeps the null vector unchanged: \( f(\overrightarrow{0})=\overrightarrow{0}\)

  2. The image of the opposite of a vector is the opposite of the image of the vector: \( \forall\;\overrightarrow{u}\in\mathbb{P},\;f(-\overrightarrow{u})=-f(\overrightarrow{u})\)

  3. It preserves the subtraction of vectors: \( \forall\;(\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^2,\;f(\overrightarrow{u}-\overrightarrow{v}) =f(\overrightarrow{u})-f(\overrightarrow{v})\)

  4. It preserves the linear combinations of vectors: \( \forall\;(\lambda,\mu,\overrightarrow{u},\overrightarrow{v})\in\mathbb{R}\times\mathbb{R}\times\mathbb{P}\times\mathbb{P},\; \; f(\lambda\overrightarrow{u}+\mu\overrightarrow{v})=\lambda f(\overrightarrow{u})+\mu f(\overrightarrow{v})\)

  5. And more generally, if \( \overrightarrow{v}\in\mathbb{P}\) is a vector such that \( \overrightarrow{v}=\lambda_1\overrightarrow{v}_1+\lambda_2\overrightarrow{v}_2+…+\lambda_n\overrightarrow{v}_n\) , with \( (\overrightarrow{v}_1,\overrightarrow{v}_2,…,\overrightarrow{v}_n)\in\mathbb{P}^n\) vectors and \( (\lambda_1,\lambda_2,…,\lambda_n)\in\mathbb{R}^n\) scalars, then \( f(\overrightarrow{v})=\lambda_1f(\overrightarrow{v}_1)+\lambda_2f(\overrightarrow{v}_2)+…+\lambda_nf(\overrightarrow{v}_n)\) .

Corollary 1

The translation \( t_{u_0}\) of vector \( \overrightarrow{u}_0\) , defined for any vector \( \overrightarrow{u}\in\mathbb{P}\) as \( t_{u_0}(\overrightarrow{u})=\overrightarrow{u}+\overrightarrow{u}_0\) is not a linear mapping unless \( \overrightarrow{u}_0=\overrightarrow{0}\) .

This is because, if \( \overrightarrow{u}_0\ne \overrightarrow{0}\) , \( t_{u_0}\) doesn’t preserve the null vector.

Proof (of the theorem 1)

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane.

  1. Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector.

    Then, as \( f\) is a linear mapping:

    \( f(\overrightarrow{0})=f(0\times\overrightarrow{u})=0\times f(\overrightarrow{u})=\overrightarrow{0}\)

  2. Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector.

    Then, as \( f\) is a linear mapping:

    \( f(-\overrightarrow{u})=f((-1)\times\overrightarrow{u})=(-1)\times f(\overrightarrow{u})=-f(\overrightarrow{u})\)

  3. Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^2\) are vectors.

    Then, as \( f\) is a linear mapping:

    \( f(\overrightarrow{u}-\overrightarrow{v}) =f(\overrightarrow{u}+(-1)\times\overrightarrow{v}) =f(\overrightarrow{u})+f((-1)\times\overrightarrow{v})\)

    \( =f(\overrightarrow{u})+(-1)\times f(\overrightarrow{v}) =f(\overrightarrow{u})-f(\overrightarrow{v})\)

  4. Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^2\) are vectors and that \( (\lambda,\mu)\in\mathbb{R}^{2}\) are scalars.

    Then, as \( f\) is a linear mapping:

    \( f(\lambda\overrightarrow{u}+\mu\overrightarrow{v}) f(\lambda\overrightarrow{u})+f(\mu\overrightarrow{v}) =\lambda f(\overrightarrow{u})+\mu f(\overrightarrow{v})\)

  5. Let’s prove the assertion by recursion on the number \( n\ge 1\) of terms of the linear combination.

    Initialisation \( n=1\) .

    The assertion is equivalent to the preservation of the multiplication of a vector by a scalar.

    Recursion

    Recursion Hypothesis

    For some \( n\ge 1\) , for any vectors \( (\overrightarrow{v}_1,\overrightarrow{v}_2,…,\overrightarrow{v}_n)\in\mathbb{P}^n\) , and for any scalars \( (\lambda_1,\lambda_2,…,\lambda_n)\in\mathbb{R}^n\) , if \( \overrightarrow{v}=\lambda_1\overrightarrow{v}_1+\lambda_2\overrightarrow{v}_2+…+\lambda_n\overrightarrow{v}_n\) , then \( f(\overrightarrow{v})=\lambda_1f(\overrightarrow{v}_1)+\lambda_2f(\overrightarrow{v}_2)+…+\lambda_nf(\overrightarrow{v}_n)\) .

    step \( n+1\)

    Assume that \( (\overrightarrow{v}_1,\overrightarrow{v}_2,…,\overrightarrow{v}_n,\overrightarrow{v}_{n+1})\in\mathbb{P}^{n+1}\) are vectors and \( (\lambda_1,\lambda_2,…,\lambda_n,\lambda_{n+1})\in\mathbb{R}^{n+1}\) are scalars, and denote \( \overrightarrow{v}=\lambda_1\overrightarrow{v}_1+\lambda_2\overrightarrow{v}_2+…+\lambda_n\overrightarrow{v}_n+\lambda_{n+1}\overrightarrow{v}_{n+1}\) .

    Then, because of the previous assertion:

    \( f(\overrightarrow{v})=f(\lambda_1\overrightarrow{v}_1+\lambda_2\overrightarrow{v}_2+…+\lambda_n\overrightarrow{v}_n) +\lambda_{n+1}f(\overrightarrow{v}_{n+1})\) .

    And because of the recursion hypothesis:

    \( f(\overrightarrow{v})=\lambda_1f(\overrightarrow{v}_1)+\lambda_2f(\overrightarrow{v}_2)+…+\lambda_nf(\overrightarrow{v}_n)) +\lambda_{n+1}f(\overrightarrow{v}_{n+1})\) .

    That proves the assertion for \( n+1\) , and that ends the proof by recursion.

2.3 Examples of Linear Mappings in \( \mathbb{P}\)

2.3.1 The null mapping \( o\)

The null mapping is the mapping \( o:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) that maps any vector of the euclidean plane to the null vector.

It is a linear mapping, because:

  • As \( \overrightarrow{0}+\overrightarrow{0}=\overrightarrow{0}\) , it preserves the addition of vectors.

  • And as \( \forall\;\lambda\in\mathbb{R},\;\lambda\overrightarrow{0}=\overrightarrow{0}\) , it preserves the multiplication of a vector by a scalar.

2.3.2 The identity mapping \( \text{Id}_{\mathbb{P}}\)

The identity mapping is the mapping \( \text{Id}_{\mathbb{P}}:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) that maps any vector of the euclidean plane to itself.

It preserves trivially the addition of vectors and the multiplication of a vector by a scalar.

It is thus a linear mapping.

2.3.3 The homotheries

The homothety of factor \( \lambda\in\mathbb{R}\) is a linear mapping, because:

  • As \( \forall\;(\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^2,\; \lambda(\overrightarrow{u}+\overrightarrow{v})=\lambda\overrightarrow{u}+\lambda\overrightarrow{v}\) , it preserves the addition of vectors.

  • And as \( \forall\;(\mu,\overrightarrow{u})\in\mathbb{R}\times\mathbb{P},\; \lambda(\mu\overrightarrow{u})=\mu(\lambda \overrightarrow{u})\) , it preserves the multiplication of a vector by a scalar.

2.3.4 The rotations

The rotation of angle \( \theta\in\mathbb{R}\) is a linear mapping, because it corresponds to the multiplication by \( e^{i\theta}\) in \( \mathbb{C}\) and, as will be proved in the last section of the course:

  • the multiplication is distributive on the addition in \( \mathbb{C}\) .

  • and the multiplication in \( \mathbb{C}\) is associative and commutative.

3 The Matrix of a Linear Mapping

Applying a linear mapping to a vector is multiplying the corresponding column vector by the matrix of the linear mapping.

3.1 Examples of Matrices of Linear Mappings in \( \mathbb{P}\)

3.1.1 The null mapping \( o\)

The null mapping \( o\) has as matrix the null matrix \( O_{22}=\begin{bmatrix}0&0\\0&0 \end{bmatrix}\) because, for any column vector \( X=\begin{bmatrix}x_1\\x_2 \end{bmatrix}\) , with \( (x_1,x_2)\in\mathbb{R}^2\) real numbers, \( O_{22}X=\begin{bmatrix}0&0\\0&0 \end{bmatrix}\begin{bmatrix}x_1\\x_2 \end{bmatrix} =\begin{bmatrix}0\times x_1+0\times x_2\\0\times x_1+0\times x_2 \end{bmatrix} =\begin{bmatrix}0\\0\end{bmatrix}=O_{21}\) , the null column vector.

3.1.2 The identity mapping \( \text{Id}_{\mathbb{P}}\)

The identity mapping \( \text{Id}_{\mathbb{P}}\) has as matrix the identity matrix \( I=\begin{bmatrix}1&0\\0&1 \end{bmatrix}\) because, for any column vector \( X\) with two real elements, \( IX=X\) , as has been proved in the lecture 10.

3.1.3 The homotheries

The homothety \( h_{\lambda}\) of factor \( \lambda\in\mathbb{R}\) has as matrix the scalar matrix \( H_{\lambda}=\begin{bmatrix}\lambda&0\\0&\lambda \end{bmatrix}\) because, for any column vector \( X=\begin{bmatrix}x_1\\x_2 \end{bmatrix}\) , with \( (x_1,x_2)\in\mathbb{R}^2\) real numbers, \( H_{\lambda}X=\begin{bmatrix}\lambda&0\\0&\lambda \end{bmatrix}\begin{bmatrix}x_1\\x_2 \end{bmatrix} =\begin{bmatrix}\lambda x_1+0\times x_2\\0\times x_1+\lambda x_2\end{bmatrix} =\begin{bmatrix}\lambda x_1\\\lambda x_2\end{bmatrix}=\lambda X\) .

3.1.4 The rotations

The rotation \( \rho_{\theta}\) of angle \( \theta\in\mathbb{R}\) has as matrix the skew-symmetric matrix \( R_{\theta}=\begin{bmatrix}\cos(\theta)&-\sin(\theta)\\ \sin(\theta)&\cos(\theta) \end{bmatrix}\) because, for any vector \( \overrightarrow{u}\in\mathbb{P}\) with column vector of coordinates \( X\) , \( R_{\theta}X=\rho_{\theta}(\overrightarrow{u})\) .

Indeed, if we consider:

  • the column vector \( X=\begin{bmatrix}x_1\\x_2\end{bmatrix}\) corresponding to the vector \( \overrightarrow{u}\) and to the complex number \( z_v=y_1+iy_2\) ,

  • and the column vector \( Y=R_{\theta}X=\begin{bmatrix}y_1\\y_2\end{bmatrix}\) corresponding to the vector \( \overrightarrow{v}\) and to the complex number \( z_v=y_1+iy_2\) ,

then \( z_v=(\cos(\theta)x_1-\sin(\theta)x_2) +i(\sin(\theta)x_1+\cos(\theta)x_2)=e^{i\theta}z_u\) .

3.2 Definition of the Matrix of a Linear Mapping

Definition 2

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane, and consider the matrix \( A=\begin{bmatrix} a_{11}&a_{12}\\a_{21}&a_{22} \end{bmatrix}\) such that:

  • the first column \( C_1=\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of the image \( f(\overrightarrow{i})\) of the first vector \( \overrightarrow{i}\) of the canonical basis,

  • and the second column \( C_2=\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of the image \( f(\overrightarrow{j})\) of the second vector \( \overrightarrow{j}\) of the canonical basis.

Then the matrix \( A\) is called the matrix of the linear mapping \( f\) in the canonical basis.

Theorem 2

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector, and consider the vector \( \overrightarrow{v}=f(\overrightarrow{u})\) .

Consider the matrix \( A\) of \( f\) in the canonical basis \( (\overrightarrow{i},\overrightarrow{j})\) .

Consider the following column vectors:

  • \( X=\begin{bmatrix}x_1\\x_2 \end{bmatrix}\) the column vector of the cartesian coordinates \( (x_{1},x_{2})\) of the vector \( \overrightarrow{u}\) ,

  • and \( Y=\begin{bmatrix}y_1\\y_2 \end{bmatrix}\) the column vector of the cartesian coordinates \( (y_{1},y_{2})\) of the vector \( \overrightarrow{v}=f(\overrightarrow{u})\) .

Then \( Y=AX\) .

That’s why the matrix \( A\) is called the matrix of the linear mapping \( f\) in the canonical basis.

Proof (of the theorem 2)

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector, and consider the vector \( \overrightarrow{v}=f(\overrightarrow{u})\) .

Consider the matrix \( A\) of \( f\) in the canonical basis \( (\overrightarrow{i},\overrightarrow{j})\) .

Consider the following column vectors:

  • \( X=\begin{bmatrix}x_1\\x_2 \end{bmatrix}\) the column vector of the cartesian coordinates \( (x_{1},x_{2})\) of the vector \( \overrightarrow{u}\) ,

  • and \( Y=\begin{bmatrix}y_1\\y_2 \end{bmatrix}\) the column vector of the cartesian coordinates \( (y_{1},y_{2})\) of the vector \( \overrightarrow{v}=f(\overrightarrow{u})\) .

Then \( \overrightarrow{u}=x_1\overrightarrow{i}+x_2\overrightarrow{j}\) , so that \( f(\overrightarrow{u})=f(x_1\overrightarrow{i}+x_2\overrightarrow{j}) =x_1f(\overrightarrow{i})+x_2f(\overrightarrow{j})\) .

Consequently, with the column vectors of the cartesian coordinates of the different vectors, \( Y=x_1C_1+x_2C_2 =\begin{bmatrix} a_{11}x_1+a_{12}x_2\\a_{21}x_1+a_{22}x_2 \end{bmatrix}=AX\) .

4 Addition, Subtraction and Scalar Multiplication of Linear Mappings

In parallel, we shall see the same operations on matrices, element by element.

4.1 Add Linear Matppings and Matrices

4.1.1 Add two Linear Mappings in \( \mathbb{P}\)

Definition 3

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Then we define the sum \( s=f+g\) of \( f\) and \( g\) as the mapping in \( \mathbb{P}\) :

\[ \begin{matrix}s:&\mathbb{P}&\rightarrow&\mathbb{P}\\&\overrightarrow{u}&\mapsto&s(\overrightarrow{u})=f(\overrightarrow{u})+g(\overrightarrow{u})\end{matrix} \]

Theorem 3

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Consider the sum \( s=f+g\) of \( f\) and \( g\) .

Then \( s\) is a linear mapping.

Proof

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Consider the sum \( s=f+g\) of \( f\) and \( g\) .

Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) are vectors and that \( \lambda\in\mathbb{R}\) is a scalar.

Then, because \( f\) and \( g\) are linear mappings, the following calculations may be performed.

  • As the addition of vectors is commutative and associative, we have:

    \( s(\overrightarrow{u}+\overrightarrow{v}) =f(\overrightarrow{u}+\overrightarrow{v})+g(\overrightarrow{u}+\overrightarrow{v}) =(f(\overrightarrow{u})+f(\overrightarrow{v}))+(g(\overrightarrow{u})+g(\overrightarrow{v})) =(f(\overrightarrow{u})+g(\overrightarrow{u}))+(f(\overrightarrow{v})+g(\overrightarrow{v})) =s(\overrightarrow{u})+s(\overrightarrow{v})\)

  • As the multiplication by a scalar is distributive on the addition of vectors, we have:

    \( s(\lambda\overrightarrow{u}) =f(\lambda\overrightarrow{u})+g(\lambda\overrightarrow{u}) =\lambda f(\overrightarrow{u})+\lambda g(\overrightarrow{u}) =\lambda(f(\overrightarrow{u})+g(\overrightarrow{u})) =\lambda s(\overrightarrow{u})\)

As this is true for any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) and any scalar \( \lambda\in\mathbb{R}\) , \( s\) is a linear mapping.

4.1.2 Add two \( 2\times 2\) Matrices

Definition 4

Assume that \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) and \( B=\begin{bmatrix}b_{11}&b_{12}\\b_{21}&b_{22}\end{bmatrix}\) are matrices with real elements.

The we define the sum element by element \( S=A+B\) as the matrix \( S=\begin{bmatrix}a_{11}+b_{11}&a_{12}+b_{12}\\a_{21}+b_{21}&a_{22}+b_{22}\end{bmatrix}\) .

4.1.3 Link the Addition of Linear Mappings and the Addition of Matrices

Theorem 4

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B\) in the canonical basis.

Then the sum \( s=f+g\) of \( f\) and \( g\) has the matrix \( S=A+B\) in the canonical basis.

Proof

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B=\begin{bmatrix}b_{11}&b_{12}\\b_{21}&b_{22}\end{bmatrix}\) in the canonical basis.

Then the first column of the matrix \( S=A+B\) is equal to:

\( C_1=\begin{bmatrix} a_{11}+b_{11}\\a_{21}+b_{21} \end{bmatrix} =\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}+\begin{bmatrix} b_{11}\\b_{21} \end{bmatrix}\) .

It is the column vector of the coordinates of \( f(\overrightarrow{i})+g(\overrightarrow{i})=s(\overrightarrow{i})\) .

And the second column of the matrix \( S=A+B\) is equal to:

\( C_2=\begin{bmatrix} a_{12}+b_{12}\\a_{22}+b_{22} \end{bmatrix} =\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}+\begin{bmatrix} b_{12}\\b_{22} \end{bmatrix}\) .

It is the column vector of the coordinates of \( f(\overrightarrow{j})+g(\overrightarrow{j})=s(\overrightarrow{j})\) .

Consequently, \( S\) is the matrix of \( s\) in the canonical basis.

4.1.4 Properties of the Addition of Linear Mappings and \( 2\times 2\) Matrices with Real Elements

Theorem 5

Assume that \( f\) , \( g\) and \( h\) are linear mappings in the euclidean plane.

Then the following properties are verified.

  1. Commutativity: \( f+g=g+f\) .

  2. Associativity: \( (f+g)+h=f+(g+h)\) .

  3. The null mapping \( o\) is neutral for the addition of linear mappings: \( f+o=o+f=f\) .

  4. If \( -f\) is the mapping opposite to \( f\) , such that \( \forall\;\overrightarrow{u}\in\mathbb{P},\;(-f)(\overrightarrow{u})=-f(\overrightarrow{u})\) , then \( -f\) is a linear mapping and \( f+(-f)=(-f)+f=o\) , the null mapping.

The last item justifies the denomination of “opposite of \( f\) ” for \( -f\) .

Corollary 2

Assume that \( A\) , \( B\) and \( C\) are \( 2\times 2\) matrices with real elements.

Then the following properties are verified.

  1. Commutativity: \( A+B=B+A\) .

  2. Associativity: \( (A+B)+C=A+(B+C)\) .

  3. The null \( 2\times 2\) matrix \( O_{22}\) is neutral for the addition of l\( 2\times 2\) matrices: \( f+o=o+f=f\) .

  4. If the opposite of \( A\) , \( -A=\begin{bmatrix}-a_{11}&-a_{12}\\{-a_{21}}&-a_{22}\end{bmatrix}\) , is the matrix opposite to \( A\) , then \( A+(-A)=(-A)+A=O_{22}\) , the null \( 2\times 2\) matrix.

The last item justifies the denomination of “opposite of \( A\) ” for \( -A\) .

Lemma 1

Assume that \( A\) is a \( 2\times 2\) matrix with real elements.

Then there exists a unique linear mapping \( f\) in the euclidean plane having \( A\) as its matrix in the canonical basis.

Lemma 2

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Then \( -f\) is a linear mapping that has the matrix \( -A\) in the canonical basis

Proof (of the lemma 1)

Assume that \( A=\begin{bmatrix} a_{11}&a_{12}\\a_{21}&a_{22} \end{bmatrix}\) is a matrix with real elements.

  1. If there exists a linear mapping \( f\) in the euclidean plane having \( A\) as its matrix in the canonical basis, then it is unique.

    Indeed, the first column \( C_1=\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of the image \( f(\overrightarrow{i})\) of the first vector \( \overrightarrow{i}\) of the canonical basis, and the second column \( C_2=\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of the image \( f(\overrightarrow{j})\) of the second vector \( \overrightarrow{j}\) of the canonical basis.

    So \( f(\overrightarrow{i})\) and \( f(\overrightarrow{j})\) are known.

    Moreover, for any vector \( \overrightarrow{u}\in\mathbb{P}\) with cartesian coordinates \( (x_{1},x_{2})\) , \( f(\overrightarrow{u})=x_{1}f(\overrightarrow{i})+x_{2}f(\overrightarrow{j})\) is also known.

    So that the knowledge of \( A\) gives the knowledge of \( f\) , that is thus unique.

  2. Let’s now build \( f\) .

    Consider the vector \( \overrightarrow{I}\) with column vector of coordinates the first column \( C_1=\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}\) of \( A\) , and the vector \( \overrightarrow{J}\) with column vector of coordinates the second column \( C_2=\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}\) of \( A\) .

    Now, define the mapping \( f\) in the euclidean plane such as, for any vector \( \overrightarrow{u}\in\mathbb{P}\) with cartesian coordinates \( (x_{1},x_{2})\) , \( f(\overrightarrow{u})=x_{1}\overrightarrow{I}+x_{2}\overrightarrow{J}\) .

    Then \( f\) is a linear mapping, because of the following statements.

    • For any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) , with cartesian coordinates \( (x_{1},x_{2})\) and \( (y_{1},y_{2})\) we have:

      \( f(\overrightarrow{u}+\overrightarrow{v})=(x_{1}+y_{1})\overrightarrow{I}+(x_{2}+y_{2})\overrightarrow{J} =(x_{1}\overrightarrow{I}+y_{1}\overrightarrow{I})+(x_{2}\overrightarrow{J}+y_{2}\overrightarrow{J})\)

      \( =(x_{1}\overrightarrow{I}+x_{2}\overrightarrow{J})+(y_{1}\overrightarrow{I}+y_{2}\overrightarrow{J}) =f(\overrightarrow{u})+f(\overrightarrow{v})\)

    • And for any vector \( \overrightarrow{u}\in\mathbb{P}\) with cartesian coordinates \( (x_{1},x_{2})\) , and for any scalar \( \lambda\in\mathbb{R}\) , we have:

      \( f(\lambda\overrightarrow{u})=(\lambda x_{1})\overrightarrow{I}+(\lambda x_{2})\overrightarrow{J} =\lambda (x_{1}\overrightarrow{I})+\lambda (x_{2}\overrightarrow{J})\)

      \( =\lambda (x_{1}\overrightarrow{I}+x_{2}\overrightarrow{J}) =\lambda f(\overrightarrow{u})\)

Proof (of the lemma 2)

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A=\begin{bmatrix} a_{11}&a_{12}\\a_{21}&a_{22} \end{bmatrix}\) in the canonical basis.

Then \( -f\) is a linear mapping, because of the following statements.

  • For any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) , with cartesian coordinates \( (x_{1},x_{2})\) and \( (y_{1},y_{2})\) we have:

    \( (-f)(\overrightarrow{u}+\overrightarrow{v})=-f(\overrightarrow{u}+\overrightarrow{v}) =-(f(\overrightarrow{u})+f(\overrightarrow{v}))=(-f(\overrightarrow{u}))+(-f(\overrightarrow{v})) =(-f)(\overrightarrow{u})+(-f)(\overrightarrow{v})\)

  • And for any vector \( \overrightarrow{u}\in\mathbb{P}\) with cartesian coordinates \( (x_{1},x_{2})\) , and for any scalar \( \lambda\in\mathbb{R}\) , we have:

    \( (-f)(\lambda\overrightarrow{u})=-f(\lambda\overrightarrow{u}) =-(\lambda f(\overrightarrow{u}))=\lambda (-f(\overrightarrow{u}))=\lambda (-f)(\overrightarrow{u})\)

Moreover we have:

  • the first column \( C_1=\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of \( f(\overrightarrow{i})\) , so that the first column \( -C_1=\begin{bmatrix} -a_{11}\\{-a_{21}} \end{bmatrix}\) of \( -A\) is the column vector of the cartesian coordinates of \( -f(\overrightarrow{i})=(-f)(\overrightarrow{i})\) ,

  • and the second column \( C_2=\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}\) of \( A\) is the column vector of the cartesian coordinates of \( f(\overrightarrow{j})\) , so that the second column \( -C_2=\begin{bmatrix} -a_{12}\\{-a_{22}} \end{bmatrix}\) of \( -A\) is the column vector of the cartesian coordinates of \( -f(\overrightarrow{j})=(-f)(\overrightarrow{j})\) .

Consequently, \( -A\) is the matrix of \( -f\) in the canonical basis.

Proof (of the corollary 2)

Assume that \( A\) , \( B\) and \( C\) are \( 2\times 2\) matrices with real elements.

Consider the linear mappings \( f\) , \( g\) and \( h\) having \( A\) , \( B\) and \( C\) as matrices in the canonical basis (cf. the lemma 1).

Then, because of the theorems 4 and 5, we have:

  1. \( A+B\) is the matrix of \( f+g\) in the canonical basis, and \( B+A\) is the matrix of \( g+f=f+g\) in the canonical basis. Consequently, \( A+B=B+A\) .

  2. \( (A+B)+C\) is the matrix of \( (f+g)+h\) in the canonical basis, and \( A+(B+C)\) is the matrix of \( f+(g+h)=(f+g)+h\) in the canonical basis. Consequently, \( (A+B)+C=A+(B+C)\) .

  3. \( A+O_{22}\) is the matrix of \( f+o=f\) in the canonical basis, and \( O_{22}+A\) is the matrix of \( o+f=f\) in the canonical basis.

    Consequently, \( A+O_{22}=O_{22}+A=A\) .

  4. Because of the lemma 2, \( -A\) is the matrix of \( -f\) in the canonical basis.

    Consequently, \( A+(-A)\) is the matrix of \( f+(-f)=o\) in the canonical basis, and \( (-A)+A\) is the matrix of \( (-f)+f=o\) in the canonical basis.

    Consequently, \( A+(-A)=(-A)+A=O_{22}\) .

Proof (of the theorem 5)

Assume that \( f\) , \( g\) and \( h\) are linear mappings in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because of the commutativity of the addition of vectors, we have:

    \( (f+g)(\overrightarrow{u})=f(\overrightarrow{u})+g(\overrightarrow{u}) =g(\overrightarrow{u})+f(\overrightarrow{u}) =(g+f)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f+g=g+f\) .

  2. Because of the associativity of the addition of vectors, we have:

    \( ((f+g)+h)(\overrightarrow{u})=(f+g)(\overrightarrow{u})+h(\overrightarrow{u})\)

    \( =(f(\overrightarrow{u})+g(\overrightarrow{u}))+h(\overrightarrow{u}) =f(\overrightarrow{u})+(g(\overrightarrow{u})+h(\overrightarrow{u}))\)

    \( =f(\overrightarrow{u})+(g+h)(\overrightarrow{u}) =(f+(g+h))(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( (f+g)+h=f+(g+h)\) .

  3. As the null vextor \( \overrightarrow{0}\) is neutral for the addition of vectors, we have:

    \( (f+o)(\overrightarrow{u})=f(\overrightarrow{u})+\overrightarrow{0}=f(\overrightarrow{u})\) .

    And also:

    \( (o+f)(\overrightarrow{u})=\overrightarrow{0}+f(\overrightarrow{u})=f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f+o=o+f=f\) .

  4. \( -f\) is a linear mapping because of the lemma 2.

    Moreover, because the opposite \( -\overrightarrow{v}\) of a vector \( \overrightarrow{v}\) is such that

    \( \overrightarrow{v}+(-\overrightarrow{v})=(-\overrightarrow{v})+\overrightarrow{v}=\overrightarrow{0}\) , we have:

    \( (f+(-f))(\overrightarrow{u})=f(\overrightarrow{u})+(-f)(\overrightarrow{u}) =f(\overrightarrow{u})+(-f(\overrightarrow{u}))=\overrightarrow{0}\) ,

    and:

    \( ((-f)+f)(\overrightarrow{u})=(-f)(\overrightarrow{u})+f(\overrightarrow{u}) =(-f(\overrightarrow{u}))+f(\overrightarrow{u})=\overrightarrow{0}\)

As the properties of the corollary 2 are verified for any \( 2\times 2\) matrices with real elements \( A\) , \( B\) and \( C\) , then \( (\mathbb{M}_{22},+)\) is a commutative group, where \( \mathbb{M}_{22}\) is the set of all \( 2\times 2\) matrices with real elements.

4.2 Subtract Linear Mappings and Matrices

4.2.1 Subtract two Linear Mappings in \( \mathbb{P}\)

Definition 5

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Then we define the difference \( d=f-g\) of \( f\) and \( g\) as the mapping in \( \mathbb{P}\) :

\[ \begin{matrix}d:&\mathbb{P}&\rightarrow&\mathbb{P}\\&\overrightarrow{u}&\mapsto&d(\overrightarrow{u})=f(\overrightarrow{u})-g(\overrightarrow{u})\end{matrix} \]

Theorem 6

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Consider the difference \( d=f-g\) of \( f\) and \( g\) .

Then \( d\) is a linear mapping.

Proof

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) and \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) are linear mappings in the euclidean plane.

Consider the difference \( d=f-g\) of \( f\) and \( g\) .

Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) are vectors and that \( \lambda\in\mathbb{R}\) is a scalar.

Then, because \( f\) and \( -g\) are linear mappings, the following calculations may be performed.

  • As the addition of vectors is commutative and associative and as \( \overrightarrow{u}_{1}-\overrightarrow{u}_{2}=\overrightarrow{u}_{1}+(-\overrightarrow{u}_{2})\) , we have:

    \( d(\overrightarrow{u}+\overrightarrow{v}) =f(\overrightarrow{u}+\overrightarrow{v})-g(\overrightarrow{u}+\overrightarrow{v})\)

    \( =f(\overrightarrow{u}+\overrightarrow{v})+(-g(\overrightarrow{u}+\overrightarrow{v}))\)

    \( =(f(\overrightarrow{u})+f(\overrightarrow{v}))+((-g)(\overrightarrow{u})+g(\overrightarrow{v}))\)

    \( =(f(\overrightarrow{u})+f(\overrightarrow{v}))+((-g)(\overrightarrow{u})+(-g)(\overrightarrow{v}))\)

    \( =(f(\overrightarrow{u})+f(\overrightarrow{v}))+((-g(\overrightarrow{u}))+(-g(\overrightarrow{v})))\)

    \( =(f(\overrightarrow{u})+(-g(\overrightarrow{u})))+(f(\overrightarrow{v})+(-g(\overrightarrow{v})))\)

    \( =(f(\overrightarrow{u})-g(\overrightarrow{u}))+(f(\overrightarrow{v})-g(\overrightarrow{v})) =d(\overrightarrow{u})+d(\overrightarrow{v})\)

  • As the multiplication by a scalar is distributive on the addition of vectors, we have:

    \( d(\lambda\overrightarrow{u}) =f(\lambda\overrightarrow{u})-g(\lambda\overrightarrow{u}) =f(\lambda\overrightarrow{u})+(-g)(\lambda\overrightarrow{u}) =\lambda f(\overrightarrow{u})+\lambda (-g)(\overrightarrow{u}))\)

    \( =\lambda(f(\overrightarrow{u})+(-g)(\overrightarrow{u})) =\lambda(f(\overrightarrow{u})+(-g(\overrightarrow{u}))) =\lambda(f(\overrightarrow{u})-g(\overrightarrow{u})) =\lambda d(\overrightarrow{u})\)

As this is true for any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) and any scalar \( \lambda\in\mathbb{R}\) , \( d\) is a linear mapping.

4.2.2 Subtract two \( 2\times 2\) Matrices

Definition 6

Assume that \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) and \( B=\begin{bmatrix}b_{11}&b_{12}\\b_{21}&b_{22}\end{bmatrix}\) are matrices with real elements.

Then we define the difference element by element \( D=A-B\) as the matrix \( S=\begin{bmatrix}a_{11}-b_{11}&a_{12}-b_{12}\\a_{21}-b_{21}&a_{22}-b_{22}\end{bmatrix}\) .

4.2.3 Link the Subtraction of Linear Mappings and the Addition of Matrices

Theorem 7

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B\) in the canonical basis.

Then the difference \( d=f-g\) of \( f\) and \( g\) has the matrix \( D=A-B\) in the canonical basis.

Proof

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B=\begin{bmatrix}b_{11}&b_{12}\\b_{21}&b_{22}\end{bmatrix}\) in the canonical basis.

Then the first column of the matrix \( D=A-B\) is equal to:

\( C_1=\begin{bmatrix} a_{11}-b_{11}\\a_{21}-b_{21} \end{bmatrix} =\begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}-\begin{bmatrix} b_{11}\\b_{21} \end{bmatrix}\) .

It is the column vector of the coordinates of \( f(\overrightarrow{i})-g(\overrightarrow{i})=d(\overrightarrow{i})\) .

And the second column of the matrix \( D=A-B\) is equal to:

\( C_2=\begin{bmatrix} a_{12}-b_{12}\\a_{22}-b_{22} \end{bmatrix} =\begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}-\begin{bmatrix} b_{12}\\b_{22} \end{bmatrix}\) .

It is the column vector of the coordinates of \( f(\overrightarrow{j})-g(\overrightarrow{j})=d(\overrightarrow{j})\) .

Consequently, \( D\) is the matrix of \( d\) in the canonical basis.

4.2.4 Properties of the Subtraction of Linear Mappings and \( 2\times 2\) Matrices with Real Elements

Theorem 8

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the null linear mapping \( o\) .

Then the following assertions hold:

  1. \( f-o=f\) ,

  2. \( o-f=-f\) ,

  3. and \( f-f=o\) .

Corollary 3

Assume that \( A\) is a \( 2\times 2\) matrix with real elements, and consider the null \( 2\times 2\) matrix \( O_{22}\) .

Then the following assertions hold:

  1. \( A-O_{22}=A\) ,

  2. \( O_{22}-A=-A\) ,

  3. and \( A-A=O_{22}\) .

Proof (of the corollary 3)

Assume that \( A\) is a \( 2\times 2\) matrix with real elements.

Consider the linear mapping \( f\) having \( A\) as matrix in the canonical basis (cf. the lemma 1).

Note that the linear mapping having \( O_{22}\) as matrix in the canonical basis is the null mapping \( o\)

Then, because of the theorems 7 and 8, and the lemma 2, we have:

  1. \( A-O_{22}\) is the matrix of \( f-o=f\) in the canonical basis, and \( f\) has the matrix \( A\) in the canonical basis. Consequently, \( A-O_{22}=A\) .

  2. \( O_{22}-A\) is the matrix of \( o-f=-f\) in the canonical basis, and \( -f\) has the matrix \( -A\) in the canonical basis. Consequently, \( O_{22}-A=-A\) .

  3. \( A-A\) is the matrix of \( f-f=o\) in the canonical basis, and \( o\) has the matrix \( O_{22}\) in the canonical basis. Consequently, \( A-A=O_{22}\) .

Proof (of the theorem 8)

Assume that \( f\) is a linear mapping in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because the null vector is neutral for the addition of vectors and is equal to its opposite, we have:

    \( (f-o)(\overrightarrow{u})=f(\overrightarrow{u})-\overrightarrow{0} =f(\overrightarrow{u})+(-\overrightarrow{0}) =f(\overrightarrow{u})-\overrightarrow{0}=f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f-o=f\) .

  2. Because the null vector is neutral for the addition of vectors, we have:

    \( (o-f)(\overrightarrow{u})=\overrightarrow{0}-f(\overrightarrow{u}) =\overrightarrow{0}+(-f(\overrightarrow{u})) =\overrightarrow{0}+(-f)(\overrightarrow{u})=(-f)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( o-f=-f\) .

  3. Because a vector minus itsels is equal to the null vector, we have:

    \( (f-f)(\overrightarrow{u})=f(\overrightarrow{u})-f(\overrightarrow{u}) =\overrightarrow{0}=o(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f-f=o\) .

4.2.5 Properties of the Addition and Subtraction of Linear Mappings and Matrices with Real Elements

Theorem 9

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane.

Then the following properties are verified.

  1. Subtracting a linear mapping is adding its opposite: \( f-g=f+(-g)\) .

  2. Addition and subtraction of linear mappings are mutually reciprocal:

    1. \( (f+g)-g=f\)

    2. and \( (f-g)+g=f\) .

Corollary 4

Assume that \( A\) and \( B\) are matrices with real elements.

Then the following properties are verified.

  1. Subtracting a matrix is adding its opposite: \( A-B=A+(-B)\) .

  2. Addition and subtraction of matrices are mutually reciprocal:

    1. \( (A+B)-B=A\) ,

    2. and \( (A-B)+B=A\) .

Proof (of the corollary 4)

Assume that \( A\) and \( B\) are \( 2\times 2\) matrices with real elements.

Consider the linear mappings \( f\) and \( g\) having respectively \( A\) and \( B\) as matrices in the canonical basis (cf. the lemma 1).

Then, because of the theorems 4, 7 and 9, we have:

  1. \( A-B\) is the matrix of \( f-g\) in the canonical basis, and \( A+(-B)\) is the matrix of \( f+(-g)=f-g\) in the canonical basis. Consequently, \( A-B=A+(-B)\) .

  2. Because of the previous item, the corollary 3 and the associativity of the addition of the \( 2\times 2\) matrices with real elements, we have:

    1. \( (A+B)-B=(A+B)+(-B)=A+(B+(-B))=A+O_{22}=A\) ,

    2. and \( (A-B)+B=(A+(-B))+B=A+((-B)+B)=A+O_{22}=A\)

Proof (of the theorem 9)

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because subtracting a vector is adding its opposite,

    \( (f-g)(\overrightarrow{u})=f(\overrightarrow{u})-g(\overrightarrow{u}) =f(\overrightarrow{u})+(-g(\overrightarrow{u})) =f(\overrightarrow{u})+(-g)(\overrightarrow{u}) =(f+(-g))(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f-g=f+(-g)\) .

  2. Because of the previous item, the theorem 8 and the associativity of the addition of the linear mappings, we have:

    1. \( (f+g)-g=(f+g)+(-g)=f+(g+(-g))=f+o=f\)

    2. \( (f-g)+g=(f+(-g))+g)=f+((-g)+g)=f+o=f\)

4.3 Multiply Linear Matppings and Matrices by a Scalar

4.3.1 Multiply a Linear Mapping in \( \mathbb{P}\) by a Scalar

Definition 7

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane and that \( \lambda\in\mathbb{R}\) is a real number.

Then we define the product \( p=\lambda f\) of \( f\) by \( \lambda\) as the mapping in \( \mathbb{P}\) :

\[ \begin{matrix}p:&\mathbb{P}&\rightarrow&\mathbb{P}\\&\overrightarrow{u}&\mapsto&p(\overrightarrow{u})=\lambda f(\overrightarrow{u})\end{matrix} \]

Theorem 10

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane and that \( \lambda\in\mathbb{R}\) is a real number.

Then the product \( p=\lambda f\) of \( f\) by \( \lambda\) is a linear mapping.

Proof

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is a linear mapping in the euclidean plane and that \( \lambda\in\mathbb{R}\) is a real number.

Consider the product \( p=\lambda f\) of \( f\) by \( \lambda\) .

Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) are vectors and that \( \mu\in\mathbb{R}\) is a scalar.

Then, because \( f\) is a linear mappings, the following calculations may be performed.

  • As the multiplication of a vector by a scalar is distributive on the addition of vectors, we have:

    \( p(\overrightarrow{u}+\overrightarrow{v})=\lambda f(\overrightarrow{u}+\overrightarrow{v}) =\lambda(f(\overrightarrow{u})+f(\overrightarrow{v}))\)

    \( =\lambda f(\overrightarrow{u})+\lambda f(\overrightarrow{v}) =p(\overrightarrow{u})+p(\overrightarrow{v})\)

  • Because of the associativity law of the multiplication of a vector by a scalar, we have:

    \( p(\mu\overrightarrow{u})=\lambda f(\mu\overrightarrow{u}) =\lambda (\mu f(\overrightarrow{u})) =(\lambda\mu) f(\overrightarrow{u})=(\mu\lambda) f(\overrightarrow{u})\)

    \( =\mu(\lambda f(\overrightarrow{u}))=\mu p(\overrightarrow{u}))\)

As this is true for any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) and any scalar \( \mu\in\mathbb{R}\) , \( p\) is a linear mapping.

4.3.2 Multiply a \( 2\times 2\) Matrix by a Scalar

Definition 8

Assume that \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) is a matrix with real elements and that \( \lambda\in\mathbb{R}\) is a real number.

Then we define the product element by element \( P=\lambda A\) of \( A\) by \( \lambda\) as the matrix \( P=\begin{bmatrix}\lambda a_{11}&\lambda a_{12}\\\lambda a_{21}&\lambda a_{22}\end{bmatrix}\)

4.3.3 Link the Multiplication by a Scalar of a Linear Mapping and of a Matrix

Theorem 11

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Assume that \( \lambda\in\mathbb{R}\) is a real number.

Then the product \( p=\lambda f\) of \( f\) by \( \lambda\) has the matrix \( P=\lambda A\) in the canonical basis.

Proof

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A=\begin{bmatrix}a_{11}&a_{12}\\a_{21}&a_{22}\end{bmatrix}\) in the canonical basis.

Assume that \( \lambda\in\mathbb{R}\) is a real number.

Then the first column of the matrix \( P=\lambda A\) is equal to:

\( C_1=\begin{bmatrix} \lambda a_{11}\\\lambda a_{21} \end{bmatrix} =\lambda \begin{bmatrix} a_{11}\\a_{21} \end{bmatrix}\) .

It is the column vector of the coordinates of \( \lambda f(\overrightarrow{i})=p(\overrightarrow{i})\) .

And the second column of the matrix \( P=\lambda A\) is equal to:

\( C_2=\begin{bmatrix} \lambda a_{12}\\\lambda a_{22} \end{bmatrix} =\lambda \begin{bmatrix} a_{12}\\a_{22} \end{bmatrix}\) .

It is the column vector of the coordinates of \( \lambda f(\overrightarrow{j})=p(\overrightarrow{j})\) .

Consequently, \( P\) is the matrix of \( p\) in the canonical basis.

4.3.4 Properties of the Multiplication by Scalars of Linear Mappings and \( 2\times 2\) Matrices with Real Elements

Theorem 12

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the null linear mapping \( o\) .

Then the following assertions hold:

  1. \( 0\times f=o\) ,

  2. and \( 1\times f=f\) .

Corollary 5

Assume that \( A\) is a \( 2\times 2\) matrix with real elements, and consider the null \( 2\times 2\) matrix \( O_{22}\) .

Then the following assertions hold:

  1. \( 0\times A=O_{22}\) ,

  2. and \( 1\times A=A\) .

Proof (of the corollary 5)

Assume that \( A\) is a \( 2\times 2\) matrix with real elements.

Consider the linear mapping \( f\) having \( A\) as matrix in the canonical basis (cf. the lemma 1).

Note that the linear mapping having \( O_{22}\) as matrix in the canonical basis is the null mapping \( o\)

Then, because of the theorems 11 and 12, we have:

  1. \( 0\times A\) is the matrix of \( 0\times f=o\) in the canonical basis, and \( o\) has the matrix \( O_{22}\) in the canonical basis. Consequently, \( 0\times A=O_{22}\) .

  2. \( 1\times A\) is the matrix of \( 1\times f=f\) in the canonical basis, and \( f\) has the matrix \( A\) in the canonical basis. Consequently, \( 1\times A=A\) .

Proof (of the theorem 12)

Assume that \( f\) is a linear mapping in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because \( 0\) multipliplied by any vector is equal to the null vector, we have:

    \( (0\times f)(\overrightarrow{u})=0\times (f(\overrightarrow{u})) \overrightarrow{0}=o(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( 0\times f=o\) .

  2. Because \( 1\) multipliplied by any vector is equal to that vector, we have:

    \( (1\times f)(\overrightarrow{u})=1\times (f(\overrightarrow{u})) =f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( 1\times f=f\) .

4.3.5 Properties of the Addition and Multiplication by Scalars of Linear Mappings and Matrices with Real Elements

Theorem 13

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Then the following properties are verified.

  1. The first distributivity law says that: \( (\lambda+\mu) f=\lambda f+\mu f\) .

  2. The second distributivity law says that: \( \lambda(f+ g)=\lambda f + \lambda g\) .

  3. And the associativity law says that: \( \lambda(\mu f)=(\lambda\mu) f\)

Corollary 6

Assume that \( A\) and \( B\) are matrices with real elements, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Then the following properties are verified.

  1. The first distributivity law says that: \( (\lambda+\mu) A=\lambda A+\mu A\) .

  2. The second distributivity law says that: \( \lambda(A+B)=\lambda A + \lambda B\) .

  3. And the associativity law says that: \( \lambda(\mu A)=(\lambda\mu) A\) .

Proof (of the corollary 4)

Assume that \( A\) and \( B\) are \( 2\times 2\) matrices with real elements, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Consider the linear mappings \( f\) and \( g\) having respectively \( A\) and \( B\) as matrices in the canonical basis (cf. the lemma 1).

Then, because of the theorems 4, 7 and 11, we have:

  1. \( (\lambda+\mu) A\) is the matrix of \( (\lambda+\mu) f\) in the canonical basis, and \( \lambda A+\mu A\) is the matrix of \( \lambda f+\mu f=(\lambda+\mu) f\) in the canonical basis. Consequently, \( (\lambda+\mu) A=\lambda A+\mu A\) .

  2. \( \lambda(A+B)\) is the matrix of \( \lambda(f+g)\) in the canonical basis, and \( \lambda A + \lambda B=\lambda(f+g)\) is the matrix of \( \lambda f + \lambda g\) in the canonical basis. Consequently, \( \lambda(A+B)=\lambda A + \lambda B\) .

  3. \( \lambda(\mu A)\) is the matrix of \( \lambda(\mu f)\) in the canonical basis, and \( (\lambda \mu) A\) is the matrix of \( (\lambda \mu) f=\lambda(\mu f)\) in the canonical basis. Consequently, \( \lambda(\mu A)=(\lambda\mu) A\) .

Proof (of the theorem 9)

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because of the first distributivity law in the euclidean plane \( \mathbb{P}\) ,

    \( ((\lambda+\mu) f)(\overrightarrow{u})=(\lambda+\mu)f(\overrightarrow{u}) =\lambda f(\overrightarrow{u})+\mu f(\overrightarrow{u}) =(\lambda f)(\overrightarrow{u})+(\mu f)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( (\lambda+\mu) f=\lambda f+\mu f\) .

  2. Because of the second distributivity law in the euclidean plane \( \mathbb{P}\) ,

    \( (\lambda(f+ g))(\overrightarrow{u})=\lambda(f+g)(\overrightarrow{u}) =\lambda (f(\overrightarrow{u})+g(\overrightarrow{u})) =\lambda f(\overrightarrow{u})+\lambda g(\overrightarrow{u})) =(\lambda f)(\overrightarrow{u})+(\lambda g)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( \lambda(f+ g)=\lambda f + \lambda g\) .

  3. Because of the associativity law in the vector plane \( \mathbb{P}\) ,

    \( (\lambda(\mu f))(\overrightarrow{u})=\lambda(\mu f)(\overrightarrow{u}) =\lambda (\mu f(\overrightarrow{u}))=(\lambda \mu) f(\overrightarrow{u} =((\lambda \mu) f)(\overrightarrow{u}\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( \lambda(\mu f)=(\lambda\mu) f\) .

As \( (\mathbb{M}_{22},+)\) is a commutative group and the properties of the corollary 6 are verified for any \( 2\times 2\) matrices with real elements \( A\) and \( B\) and any real numbers \( \lambda\) and \( \mu\) , then \( (\mathbb{M}_{22},+,\cdot)\) is a vector space, where ‘\( \cdot\) ’ denotes the multiplication of the \( 2\times 2\) matrices with real elements by scalars.

Moreover, the vector space \( \mathbb{M}_{22}\) is of dimension 4, with the canonical basis \( \left(\begin{bmatrix}1&0\\0&0\end{bmatrix}, \begin{bmatrix}0&1\\0&0\end{bmatrix}, \begin{bmatrix}0&0\\1&0\end{bmatrix}, \begin{bmatrix}0&0\\0&1\end{bmatrix}\right)\).

That means that any \( 2\times 2\) matrix with real element is a unique linear combination of the matrices in the canonical basis.

Indeed, if \( A=\begin{bmatrix}a&b\\c&d\end{bmatrix}\) , then the only way to write \( A\) as a sum of scalar products of the matrices in the canonical basis of \( \mathbb{M}_{22}\) is

\( A=a\begin{bmatrix}1&0\\0&0\end{bmatrix}+ b\begin{bmatrix}0&1\\0&0\end{bmatrix}+ c\begin{bmatrix}0&0\\1&0\end{bmatrix}+ d\begin{bmatrix}0&0\\0&1\end{bmatrix}\) .

5 Composition of Linear Mappings

The composition of linear mappings may be quite easily described with their matrices in the canonical basis, because the matrix of the composed mapping (that is a linear mapping) is the product of the composed linear mappings.

5.1 Composition of Mappings in \( \mathbb{P}\)

Let’s remember the definition of the composition of mappings in \( \mathbb{P}\) .

Definition 9

The composed mapping \( f\circ g\) of two mappings \( f\) and \( g\) in \( \mathbb{P}\) is the mapping in \( \mathbb{P}\) that, to any vector \( \overrightarrow{u}\in\mathbb{P}\) assigns the vector \( (f\circ g)(\overrightarrow{u})=f(g(\overrightarrow{u}))\) .

5.2 Examples of Composition of Linear Mappings

5.2.1 Composition of Two Homotheties

Theorem 14

Assume that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers, and consider the homotheties of factors \( \lambda\) and \( \mu\) :

\[ \begin{matrix}h_{\lambda}:&\mathbb{P}&\rightarrow&\mathbb{P}\\ \\&\overrightarrow{u}&\mapsto&h_{\lambda}(\overrightarrow{u})=\lambda\overrightarrow{u}\end{matrix} \]

and

\[ \begin{matrix}h_{\mu}:&\mathbb{P}&\rightarrow&\mathbb{P}\\ \\&\overrightarrow{u}&\mapsto&h_{\mu}(\overrightarrow{u})=\mu\overrightarrow{u}\end{matrix} \]

Then the composed mapping \( h_{\lambda}\circ h_{\mu}\) is the homothety \( h_{\lambda\mu}\) of factor \( \lambda\mu\) .

Proof

Assume that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers, and consider the homotheties \( h_{\lambda}\) and \( h_{\mu}\) of factors \( \lambda\) and \( \mu\) .

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then because of the associativity law in the vector plane \( \mathbb{P}\) , we have:

\( (h_{\lambda}\circ h_{\mu})(\overrightarrow{u})=h_{\lambda}(h_{\mu}(\overrightarrow{u})) =\lambda(\mu\overrightarrow{u})=(\lambda\mu)\overrightarrow{u} =h_{\lambda\mu}(\overrightarrow{u})\) .

As this is true for any vector \( \overrightarrow{u}\) , \( h_{\lambda}\circ h_{\mu}=h_{\lambda\mu}\) .

5.2.2 Composition of Two Rotations

Theorem 15

Assume that \( (\theta,\mu)\in\mathbb{R}^2\) are real numbers, and consider the rotations \( \rho_{\theta}\) and \( \rho_{\mu}\) of angles \( \theta\) and \( \mu\) .

Then the composed mapping \( \rho_{\theta}\circ\rho_{\mu}\) is the rotation \( \rho_{(\theta+\mu)}\) of angle \( \theta+\mu\) .

Proof

Assume that \( (\theta,\mu)\in\mathbb{R}^2\) are real numbers, and consider the rotations \( \rho_{\theta}\) and \( \rho_{\mu}\) of angles \( \theta\) and \( \mu\) .

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane, and consider the corresponding complex number \( z_{u}\) .

Then the following assertions hold:

  • \( \rho_{\mu}(\overrightarrow{u})\) corresponds to the complex number \( e^{i\mu}z_{u}\) .

  • Because of the associativity of the multiplication in \( \mathbb{C}\) (that will be proved in the last section of the course), and because \( e^{i\theta}e^{i\mu}=e^{i(\theta+\mu)}\) , \( (\rho_{\theta}\circ\rho_{\mu})(\overrightarrow{u})=\rho_{\theta}(\rho_{\mu}(\overrightarrow{u}))\) corresponds to the complex number: \( e^{i\theta}(e^{i\mu}z_{u})=(e^{i\theta}e^{i\mu})z_{u}=e^{i(\theta+\mu)}z_{u}\) , that corresponds to the vector \( \rho_{(\theta+\mu)}(\overrightarrow{u})\) .

As this is true for any vector \( \overrightarrow{u}\) , \( \rho_{\theta}\circ\rho_{\mu}=\rho_{(\theta+\mu)}\) .

5.2.3 Composition of the Null Mapping with Any Linear Mapping to the Left and to the Right

Theorem 16

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the null mapping \( o\) in the euclidean plane. Then the composed mappings \( o\circ f\) and \( f\circ o\) are both equal to the null mapping \( o\) .

Proof

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the null mapping \( o\) in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following assertions hold:

  • By definition of the null mapping \( o\) , \( (o\circ f)(\overrightarrow{u})=o(f(\overrightarrow{u}))=\overrightarrow{0}=o(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( o\circ f=o\) .

  • And by definition of the null mapping \( o\) , together with the fact that \( f(\overrightarrow{0})=\overrightarrow{0}\) , \( (f\circ o)(\overrightarrow{u})=f(o(\overrightarrow{u}))=f(\overrightarrow{0})=\overrightarrow{0}=o(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f\circ o=o\) .

5.2.4 Composition of the Identity Mapping with Any Linear Mapping to the Left and to the Right

Theorem 17

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the identity mapping \( \text{Id}_{\mathbb{P}}\) in the euclidean plane.

Then the composed mappings \( \text{Id}_{\mathbb{P}}\circ f\) and \( f\circ \text{Id}_{\mathbb{P}}\) are both equal to the linear mapping \( f\) .

Proof

Assume that \( f\) is a linear mapping in the euclidean plane, and consider the identity mapping \( \text{Id}_{\mathbb{P}}\) in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following assertions hold:

  • By definition of the identity mapping \( \text{Id}_{\mathbb{P}}\) , \( (\text{Id}_{\mathbb{P}}\circ f)(\overrightarrow{u})=\text{Id}_{\mathbb{P}}(f(\overrightarrow{u})) =f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( \text{Id}_{\mathbb{P}}\circ f=f\) .

  • And by definition of the identity mapping \( \text{Id}_{\mathbb{P}}\) again,

    \( (f\circ \text{Id}_{\mathbb{P}})(\overrightarrow{u})=f(\text{Id}_{\mathbb{P}}(\overrightarrow{u})) =f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f\circ \text{Id}_{\mathbb{P}}=f\) .

5.3 Composition of Linear Mappings and Multiplication of Matrices

Theorem 18

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B\) in the canonical basis.

Then the composed mapping \( f\circ g\) of \( f\) and \( g\) is a linear mapping that has the product matrix\( AB\) as matrix in the canonical basis.

Proof

Assume that the linear mapping \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( A\) in the canonical basis.

Assume that the linear mapping \( g:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) has the matrix \( B\) in the canonical basis.

Consider the composed mapping \( f\circ g\) of \( f\) and \( g\) .

Assume that \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) are vectors and that \( \lambda\in\mathbb{R}\) is a scalar.

Then, because \( f\) and \( g\) are linear mappings, the following calculations may be performed.

  • \( (f\circ g)(\overrightarrow{u}+\overrightarrow{v}) =f(g(\overrightarrow{u}+\overrightarrow{v})) =f(g(\overrightarrow{u})+g(\overrightarrow{v})) =f(g(\overrightarrow{u}))+f(g(\overrightarrow{v})) =(f\circ g)(\overrightarrow{u})+(f\circ g)(\overrightarrow{v})\)

  • \( (f\circ g)(\lambda\overrightarrow{u}) =f(g(\lambda\overrightarrow{u})) =f(\lambda g(\overrightarrow{u})) =\lambda f(g(\overrightarrow{u})) =\lambda (f\circ g)(\overrightarrow{u})\)

As this is true for any vectors \( (\overrightarrow{u},\overrightarrow{v})\in\mathbb{P}^{2}\) and any scalar \( \lambda\in\mathbb{R}\) , \( f\circ g\) is a linear mapping.

Assume that \( X\) , \( Y\) and \( Z\) are the column vectors of the cartesian coordinates of the vectors \( \overrightarrow{u}\) , \( g(\overrightarrow{u})\) and \( (f\circ g)(\overrightarrow{u})=f(g(\overrightarrow{u}))\) .

Then \( Y=BX\) and \( Z=AY=A(BX)=(AB)X\) because of the associativity lemma in the pdf of the section 3 about the solution of \( 2\times 2\) linear systems.

As this is true for any vector \( \overrightarrow{u}\) the product matrix \( AB\) is the matrix of \( f\circ g\) in the canonical basis.

5.4 Properties of the Composition of Linear Mappings and of the Multiplication of Matrices with Real Elements

Theorem 19

Assume that \( f\) , \( g\) and \( h\) are linear mappings in the euclidean plane.

Then the following properties are verified for the composition of linear mappings.

  1. Associativity: \( (f\circ g)\circ h=f\circ (g\circ h)\) .

  2. The identity mapping \( \text{Id}_{\mathbb{P}}\) is neutral: \( \text{Id}_{\mathbb{P}}\circ f=f\circ \text{Id}_{\mathbb{P}}=f\) .

  3. The null mapping \( o\) is absorbant: \( o\circ f=f\circ o=o\) .

Corollary 7

Assume that \( A\) , \( B\) and \( C\) are \( 2\times 2\) matrices with real elements.

Then the following properties are verified for the multiplication of matrices.

  1. Associativity: \( (AB)C=A(BC)\) .

  2. The identity \( 2\times 2\) matrix \( I\) is neutral: \( AI=IA=A\) .

  3. The null \( 2\times 2\) matrix \( O_{22}\) is absorbant: \( AO_{22}=O_{22}A=O_{22}\) .

Beware, the multiplication of matrices and thus the composition of linear mappings is not commutative in general.

For instance, with \( A=\begin{bmatrix}1&1\\0&0\end{bmatrix}\) and \( B=\begin{bmatrix}0&1\\0&1\end{bmatrix}\) , we have:

\( AB=\begin{bmatrix}1&1\\0&0\end{bmatrix}\begin{bmatrix}0&1\\0&1\end{bmatrix} =\begin{bmatrix} 1\times 0+1\times 0&1\times 1+1\times 1\\ 0\times 0+0\times 0&0\times 1+0\times 1 \end{bmatrix} =\begin{bmatrix}0&1\\0&0\end{bmatrix}\) .

And

\( BA=\begin{bmatrix}0&1\\0&1\end{bmatrix}\begin{bmatrix}1&1\\0&0\end{bmatrix} =\begin{bmatrix} 0\times 1+1\times 0&0\times 1+1\times 0\\ 0\times 1+1\times 0&0\times 1+1\times 0 \end{bmatrix} =\begin{bmatrix}0&0\\0&0\end{bmatrix}\ne AB\) .

Proof (of the corollary 7)

Assume that \( A\) , \( B\) and \( C\) are \( 2\times 2\) matrices with real elements.

Consider the linear mappings \( f\) , \( g\) and \( h\) having \( A\) , \( B\) and \( C\) as matrices in the canonical basis (cf. the lemma 1).

Then, because of the theorem 18, we have:

  1. \( (AB)C\) is the matrix of \( (f\circ g)\circ h\) in the canonical basis, and \( A(BC)\) is the matrix of \( f\circ (g\circ h)=f\circ g)\circ h\) in the canonical basis. Consequently, \( (AB)C=A(BC)\) .

  2. \( AI\) is the matrix of \( f\circ \text{Id}_{\mathbb{P}}=f\) in the canonical basis, and \( IA\) is the matrix of \( (\text{Id}_{\mathbb{P}}\circ f=f\) in the canonical basis.

    Consequently, \( AI=IA=A\) .

  3. \( AO_{22}\) is the matrix of \( f\circ o=o\) in the canonical basis, and \( O_{22}A\) is the matrix of \( o\circ f=o\) in the canonical basis.

    Moreover, \( o\) has the matrix \( O_{22}\) i the canonical basis.

    Consequently, \( AO_{22}=O_{22}A=O_{22}\) .

Proof (of the theorem 19)

Assume that \( f\) , \( g\) and \( h\) are linear mappings in the euclidean plane.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. \( ((f\circ g)\circ h)(\overrightarrow{u})=f((g\circ h)(\overrightarrow{u})) =f(g(h(\overrightarrow{u})))\)

    \( =f((g\circ h)(\overrightarrow{u})) =(f\circ (g\circ h))(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( (f\circ g)\circ h=f\circ (g\circ h)\) .

  2. \( (\text{Id}_{\mathbb{P}}\circ f)(\overrightarrow{u})=\text{Id}_{\mathbb{P}}(f(\overrightarrow{u})) =f(\overrightarrow{u})\) .

    And

    \( (f\circ \text{Id}_{\mathbb{P}})(\overrightarrow{u})=f(\text{Id}_{\mathbb{P}}(\overrightarrow{u})) =f(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( \text{Id}_{\mathbb{P}}\circ f=f\circ \text{Id}_{\mathbb{P}}=f\) .

  3. \( (o\circ f)(\overrightarrow{u})=o(f(\overrightarrow{u})) =\overrightarrow{0}=o(\overrightarrow{u})\) .

    And

    \( (f\circ o)(\overrightarrow{u})=f(o(\overrightarrow{u})) =f(\overrightarrow{0})=\overrightarrow{0}=o(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( o\circ f=f\circ o=o\) .

5.5 Properties, Shared with the Addition and Scalar Multiplication, of the Composition of Linear Mappings and of the Multiplication of Matrices with Real Elements

Theorem 20

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Then the following properties are verified.

  1. The composition of linear mappings is distributive to the left on the addtion of linear mappings: \( f\circ (g+ h)=f\circ g+f\circ h\) .

  2. The composition of linear mappings is distributive to the right on the addtion of linear mappings: \( (f+g)\circ h=f\circ h+g\circ h\) .

  3. The composition of linear mappings is compatible with the scalar multiplication of linear mappings: \( (\lambda f)\circ (\mu g)=(\lambda\mu)(f\circ g)\)

Corollary 8

Assume that \( A\) and \( B\) are matrices with real elements, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Then the following properties are verified.

  1. The multiplication of matrices is distributive to the left on the addtion of matrices: \( A(B+C)=AB+AC\) .

  2. The multiplication of matrices is distributive to the right on the addtion of matrices: \( (A+B)C=AC+BC\) .

  3. The multiplication of matrices is compatible with the scalar multiplication of matrices: \( (\lambda A)(\mu B)=(\lambda\mu)(AB)\) .

Proof (of the corollary 8)

Assume that \( A\) and \( B\) are \( 2\times 2\) matrices with real elements, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Consider the linear mappings \( f\) and \( g\) having respectively \( A\) and \( B\) as matrices in the canonical basis (cf. the lemma 1).

Then, because of the theorems 4, 11, 18 and 20, we have:

  1. \( A(B+C)\) is the matrix of \( f\circ (g+ h)\) in the canonical basis, and \( AB+AC\) is the matrix of \( f\circ g+f\circ h=f\circ (g+ h)\) in the canonical basis. Consequently, \( A(B+C)=AB+AC\) .

  2. \( (A+B)C\) is the matrix of \( (f+g)\circ h\) in the canonical basis, and \( AC+BC\) is the matrix of \( f\circ h+g\circ h=f\circ (g+ h)\) in the canonical basis. Consequently, \( (A+B)C=AC+BC\) .

  3. \( (\lambda A)(\mu B)\) is the matrix of \( (\lambda f)\circ (\mu g)\) in the canonical basis, and \( (\lambda \mu) (AB)\) is the matrix of \( (\lambda \mu) (f\circ g)=(\lambda f)\circ (\mu g)\) in the canonical basis. Consequently, \( (\lambda A)(\mu B)=(\lambda\mu)(AB)\) .

Proof (of the theorem 20)

Assume that \( f\) and \( g\) are linear mappings in the euclidean plane, and that \( (\lambda,\mu)\in\mathbb{R}^2\) are real numbers.

Assume that \( \overrightarrow{u}\in\mathbb{P}\) is a vector in the euclidean plane.

Then the following calculations may be performed.

  1. Because \( f\) is a linear mapping, we have:

    \( (f\circ (g+ h))(\overrightarrow{u})=f((g+ h)(\overrightarrow{u})) =f(g(\overrightarrow{u})+ h(\overrightarrow{u})) =f(g(\overrightarrow{u}))+ f(h(\overrightarrow{u})) =(f\circ g)(\overrightarrow{u})+ (f\circ h)(\overrightarrow{u}) =(f\circ g+ f\circ h)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( f\circ (g+ h)=f\circ g+ f\circ h\) .

  2. \( ((f+g)\circ h)(\overrightarrow{u})=(f+g)(h(\overrightarrow{u})) =f(h(\overrightarrow{u}))+g(h(\overrightarrow{u})) =(f\circ h)(\overrightarrow{u})+(g\circ h)(\overrightarrow{u}) =(f\circ h + g\circ h)(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( (f+g)\circ h=f\circ h+g\circ h\) .

  3. Because \( f\) and \( g\) are linear mappings, and because of the associativity law in the vector plane \( \mathbb{P}\) ,

    \( ((\lambda f)\circ (\mu g))(\overrightarrow{u}) =(\lambda f)((\mu g)(\overrightarrow{u})) =\lambda f(\mu g(\overrightarrow{u})) =\lambda(\mu f(g(\overrightarrow{u})))\)

    \( =(\lambda\mu)(f(g(\overrightarrow{u})) =(\lambda\mu)(f\circ g)(\overrightarrow{u}) =((\lambda\mu)(f\circ g))(\overrightarrow{u})\) .

    As this is true for any vector \( \overrightarrow{u}\) , \( (\lambda f)\circ (\mu g)=(\lambda\mu)(f\circ g)\) .

Because that:

  • \( (\mathbb{M}_{22},+)\) is a commutative group,

  • the multiplication of matrices ‘\( \times\) ’ is associative,

  • it has a neutral element (the identity matrix \( I\) ),

  • and it is distributive to the left and to the right on the addition of matrices,

\( (\mathbb{M}_{22},+,\times) \) is a unitary ring.

And because that:

  • \( (\mathbb{M}_{22},+,\cdot)\) is a vector space (of dimension 4),

  • \( (\mathbb{M}_{22},+,\times) \) is a unitary ring,

  • and the multiplication of matrices is compatible with the scalar multiplication,

\( (\mathbb{M}_{22},+,\cdot,\times) \) is a unitary algebra.

6 Inverses of Linear Mappings and of Matrices

6.1 Inverse of a Mapping in \( \mathbb{P}\)

Definition 10

The inverse of a mapping \( f\) in \( \mathbb{P}\) exists if and only if \( f\) is one-to-one from \( \mathbb{P}\) onto \( \mathbb{P}\) , and then it assigns to any vector \( \overrightarrow{v}\in\mathbb{P}\) the unique vector \( \overrightarrow{u}\in\mathbb{P}\) such that \( f(\overrightarrow{u})=\overrightarrow{v}\) .

If the inverse of \( f\) exists it is denoted \( f^{-1}\) and \( f\) is said to be invertible.

Theorem 21

Assume that \( f\) is an invertible mapping in \( \mathbb{P}\) .

Then \( f^{-1}\circ f=f\circ f^{-1}=\text{Id}_{\mathbb{P}}\) .

Proof

Assume that \( f\) is an invertible mapping in \( \mathbb{P}\) .

Assume that \( \overrightarrow{v}\in\mathbb{P}\) is a vector, and consider the vector\( \overrightarrow{u}\in\mathbb{P}\) such that \( \overrightarrow{v}=f(\overrightarrow{u})\) .

Then the following calculations may be performed.

  • \( (f^{-1}\circ f)(\overrightarrow{u}) =f^{-1}(f(\overrightarrow{u}))=f^{-1}(\overrightarrow{v}) =\overrightarrow{u}=\text{Id}_{\mathbb{P}}(\overrightarrow{u})\) .

    As for any vector \( \overrightarrow{u}\in\mathbb{P}\) , we may consider the vector \( \overrightarrow{v}=f(\overrightarrow{u})\) , this is true for any vector \( \overrightarrow{u}\in\mathbb{P}\) .

    Consequently, \( f^{-1}\circ f=\text{Id}_{\mathbb{P}}\)

  • \( (f\circ f^{-1})(\overrightarrow{v}) =f(f^{-1}(\overrightarrow{v}))=f(\overrightarrow{u}) =\overrightarrow{v}=\text{Id}_{\mathbb{P}}(\overrightarrow{v})\) .

    As this is true for any vector \( \overrightarrow{v}\in\mathbb{P}\) , so that \( f\circ f^{-1}=\text{Id}_{\mathbb{P}}\) .

6.2 Examples of invertible and not invertible linear mappings

  • The null mapping \( o\) is not one to one, because every vectors in the euclidean plane have the null vector as image. So, it is not invertible.

  • The identity mapping \( \text{Id}_{\mathbb{P}}\) is invertible with inverse itself, because for any vector \( \overrightarrow{v}\in\mathbb{P}\) , \( \overrightarrow{u}=\overrightarrow{v}\) is the only vector such that \( \overrightarrow{v}=\text{Id}_{\mathbb{P}}(\overrightarrow{u})\) .

  • The homothety \( h_{\lambda}\) of factor \( \lambda\in\mathbb{R}\) is invertible or not, on the following conditions.

    • If \( \lambda\ne 0\) , then \( h_{\lambda}\) is invertible with inverse \( h_{\frac{1}{\lambda}}\) , because, for any vector \( \overrightarrow{v}\in\mathbb{P}\) , \( \overrightarrow{u}=\frac{1}{\lambda}\overrightarrow{v}\) is the only vector such that \( \overrightarrow{v}=\lambda\overrightarrow{u}\) .

    • If \( \lambda=0\) , then \( h_{\lambda}=h_0=o\) , the null mapping, that is not invertible.

  • The rotation \( \rho_{\theta}\) of angle \( \theta\) is invertible with inverse \( \rho_{-\theta}\) , as was already seen in lecture 40 about the rotations as multiplications of complex exponentials.

6.3 Inverse of a \( 2\times 2\) Matrix with Real Elements

Definition 11

The inverse of a \( 2\times 2\) matrix \( A\) with real elements exists if and only if there exists a \( 2\times 2\) matrix \( B\) with real elements such that \( AB=I\) .

If it exists, the inverse of \( A\) is denoted \( A^{-1}\) , and the matrix \( A\) is say to be invertible.

As was proved in the lecture 9 about the matrix proof of the solution of a linear system, the following theorem holds.

Theorem 22

If \( A=\begin{bmatrix} a&b\\ c&d \end{bmatrix}\) , with \( (a,b,c,d)\in\mathbb{R}^{4}\) real numbers, then the inverse of \( A\) exists if and only if its determinant \( \det(A)=ad-bc\) is non zero, and in that case \( A^{-1}=\frac{1}{ad-bc}\begin{bmatrix} d&-b\\ {-c}&a \end{bmatrix}\) .

Corollary 9

If \( A\) is an invertible \( 2\times 2\) matrix with real elements, then \( A^{-1}A=AA^{-1}=I\) , the identity matrix.

Proof (Of the corollary)

Assume that \( (a,b,c,d)\in\mathbb{R}^{4}\) are real numbers such that \( ad-bc\ne 0\) , and consider the invertible matrix \( A=\begin{bmatrix} a&b\\ c&d \end{bmatrix}\) .

Let’s denote \( A^{*}=\begin{bmatrix} d&-b\\ -c&a \end{bmatrix}\) , so that \( A^{-1}=\frac{1}{ad-bc}A^{*}\) .

Then, because of the compatiblility of the matrix multiplication and the scalar multiplication, \( A^{-1}A=\frac{1}{ad-bc}A^{*}A\) and \( AA^{-1}=\frac{1}{ad-bc}AA^{*}\) .

But the following calculations may be performed.

  • \( A^{*}A =\begin{bmatrix} d&-b\\ -c&a \end{bmatrix}\begin{bmatrix} a&b\\ c&d \end{bmatrix} =\begin{bmatrix} da-bc&db-bd\\ -ca+ac&-cb+ad \end{bmatrix}\)

    \( =\begin{bmatrix} ad-bc&0\\ 0&ad-bc \end{bmatrix}=(ad-bc)I\) .

    Because of the compatibility of the scalar multiplication and the matrix multiplication, we may conclude that \( A^{-1}A=I\) .

  • \( AA^{*} =\begin{bmatrix} a&b\\ c&d \end{bmatrix}\begin{bmatrix} d&-b\\ -c&a \end{bmatrix} =\begin{bmatrix} ad-bc&-ab+ba\\ cd-dc&-cb+da \end{bmatrix}\)

    \( =\begin{bmatrix} ad-bc&0\\ 0&ad-bc \end{bmatrix}=(ad-bc)I\) .

    Because of the compatibility of the scalar multiplication and the matrix multiplication, we may conclude that \( AA^{-1}=I\) .

6.4 Inverses of Linear Mappings and of Matrices

Theorem 23

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is an invertible linear mapping having the matrix \( A\) in the canonical basis.

Then the following assertions hold.

  1. The inverse \( f^{-1}\) of \( f\) is a linear mapping.

  2. The matrix \( A\) is invertible.

  3. The matrix of \( f^{-1}\) in the canonical basis is the inverse \( A^{-1}\) of \( A\) .

Proof

Assume that \( f:\;\mathbb{P}\;\rightarrow\;\mathbb{P}\) is an invertible linear mapping having the matrix \( A\) in the canonical basis.

  1. Assume that \( (\overrightarrow{v_{1}},\overrightarrow{v_{2}})\in\mathbb{P}^{2}\) are vectors and that \( \lambda\in\mathbb{R}\) is a scalar.

    Consider the vectors \( \overrightarrow{u_{1}}=f^{-1}(\overrightarrow{v_{1}})\) and \( \overrightarrow{u_{2}}=f^{-1}(\overrightarrow{v_{2}})\) .

    Then \( \overrightarrow{v_{1}}=f(\overrightarrow{u_{1}})\) and \( \overrightarrow{v_{2}}=f(\overrightarrow{u_{2}})\)

    Then, because \( f\) is a linear mapping, the following calculations may be performed.

    • \( f(\overrightarrow{u_{1}}+\overrightarrow{u_{2}}) =f(\overrightarrow{u_{1}})+f(\overrightarrow{u_{2}}) =\overrightarrow{v_{1}}+\overrightarrow{v_{2}}\) .

      Consequently,

      \( f^{-1}(\overrightarrow{v_{1}}+\overrightarrow{v_{2}}) =\overrightarrow{u_{1}}+\overrightarrow{u_{2}} =f^{-1}(\overrightarrow{v_{1}})+f^{-1}(\overrightarrow{v_{2}})\)

    • \( f(\lambda\overrightarrow{u_{1}}) =\lambda f(\overrightarrow{u_{1}}) =\lambda\overrightarrow{v_{1}}\)

      Consequently,

      \( f^{-1}(\lambda\overrightarrow{v_{1}}) =\lambda\overrightarrow{u_{1}} =\lambda f^{-1}(\overrightarrow{v_{1}})\)

    • As this is true for any vectors \( \overrightarrow{u_{1}}\) and \( \overrightarrow{u_{2}}\) and any scalar \( \lambda\) , \( f^{-1}\) is a linear mapping.

  2. As \( f\) is invertible, it is one-to-one from \( \mathbb{P}\) onto \( \mathbb{P}\) .

    Assume that \( \overrightarrow{v}\in\mathbb{P}\) is a vector with column vector of cartesian coordinates \( Y\) .

    Consider the unique vector \( \overrightarrow{u}\in\mathbb{P}\) such that \( f(\overrightarrow{u})=\overrightarrow{v}\) .

    Then the column vector \( X\) of the cartesian coordinates of the vector \( \overrightarrow{u}\) is the unique solution of the matrix equation \( AX=Y\) .

    Consequently, \( A\) is invertible (with a non zero determinant).

  3. With the conventions of the previous item, as \( Y=AX\) , then \( X=A^{-1}Y\) .

    Consequently, as \( \overrightarrow{u}=f^{-1}(\overrightarrow{v})\) , \( A^{-1}\) is the matrix of \( f^{-1}\) in the canonical basis.

7 Kernel, Range and Rank of a Linear Mapping

We enter here into the structure of the different linear mappings, with their kernels, ranges and ranks.

7.1 The Kernel of a Linear Mapping

7.1.1 Definition of the Kernel of a Linear Mapping

Definition 12

We define the kernel of a linear mapping \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) as the set of all vectors \( \overrightarrow{u}\in\mathbb{P}\) such that \( f(\overrightarrow{u})=\overrightarrow{0}\) :

\[ \ker(f)=\{\overrightarrow{u}\in\mathbb{P} : f(\overrightarrow{u})=\overrightarrow{0}) \} \]

7.1.2 Examples of Kernels of Linear Mappings

The null mapping \( o\) As for any vector \( \overrightarrow{u}\in\mathbb{P}\) , \( o(\overrightarrow{u})=\overrightarrow{0}\) , the kernel of the null mapping \( o\) is the whole plane \( \mathbb{P}\) .

Reversely, if the kernel of a linear mapping \( f\) is the whole plane \( \mathbb{P}\) then, for any vector \( \overrightarrow{u}\in\mathbb{P}\) , \( f(\overrightarrow{u})=\overrightarrow{0}\) , so that \( f\) is the null mapping \( o\) .

Consequently, the null mapping \( o\) is the only mapping having \( \mathbb{P}\) as kernel.

The identity mapping \( \text{Id}_{\mathbb{P}}\) As \( \text{Id}_{\mathbb{P}}\) is one to one and \( \text{Id}_{\mathbb{P}}(\overrightarrow{0})=\overrightarrow{0}\) , then \( \overrightarrow{0}\) is the only vector \( \overrightarrow{u}\) such that \( \text{Id}_{\mathbb{P}}(\overrightarrow{u})=\overrightarrow{0}\) .

Consequently, the kernel of the identity mapping \( \text{Id}_{\mathbb{P}}\) is the singleton \( \{\overrightarrow{0}\}\) .

Homotheties with non zero factors As the homothety \( h_{\lambda}\) with a non zero factor \( \lambda\in\mathbb{R}^*\) is one to one and \( h_{\lambda}(\overrightarrow{0})=\overrightarrow{0}\) , then \( \overrightarrow{0}\) is the only vector \( \overrightarrow{u}\) such that \( h_{\lambda}(\overrightarrow{u})=\overrightarrow{0}\) .

Consequently, the kernel of the homothety \( h_{\lambda}\) with a non zero factor \( \lambda\in\mathbb{R}^*\) is the singleton \( \{\overrightarrow{0}\}\) .

The rotations As the rotation \( \rho_{\theta}\) with angle \( \theta\in\mathbb{R}^*\) is one to one and \( \rho_{\theta}(\overrightarrow{0})=\overrightarrow{0}\) , then \( \overrightarrow{0}\) is the only vector \( \overrightarrow{u}\) such that \( \rho_{\theta}(\overrightarrow{u})=\overrightarrow{0}\) .

Consequently, the kernel of the rotation \( \rho_{\theta}\) is the singleton \( \{\overrightarrow{0}\}\) .

The orthogonal projection on the \( x\) axis

Orthogonal projection on the x) axis
Figure 1. Orthogonal projection on the \( x\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_x(\overrightarrow{u})=\begin{bmatrix}x\\0\end{bmatrix}\) .

The kernel \( \ker({p_x})\) of \( p_{x}\) is the set of vectors \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) such that \( x=0\) .

Consequently, \( \ker({p_x})\) is the \( y\) axis.

The orthogonal projection on the \( y\) axis

Orthogonal projection on the y) axis
Figure 2. Orthogonal projection on the \( y\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_y(\overrightarrow{u})=\begin{bmatrix}0\\y\end{bmatrix}\) .

The kernel \( \ker({p_y})\) of \( p_{y}\) is the set of vectors \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) such that \( y=0\) .

Consequently, \( \ker({p_y})\) is the \( x\) axis.

7.1.3 Find the Kernel of a Linear Mapping

Theorem 24

Assume that \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is a linear mapping with matrix \( A=\begin{bmatrix} a&b\\c&d \end{bmatrix}\) in the canonical basis, and consider the determinant \( \det(A)=ad-bc\) of the matrix \( A\) .

Then the following assertions hold:

  1. If \( A=O_{22}\) , the null matrix, then \( \ker(f)=\mathbb{P}\) .

  2. If \( A\ne O_{22}\) and \( \det{A}=0\) , then \( \ker(f)\) is the straight line of equation \( ax+by=0\) , unless \( a=b=0\) , in which case it is the straight line of equation \( cx+dy=0\) .

  3. And if \( \det{A}\ne 0\) , then \( \ker(f)=\{\overrightarrow{0}\}\) , and in that case only, \( f\) is invertible.

The following corollary is a straightforward consequence of the theorem 24.

Corollary 10

The linear mapping \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is invertible if and only if \( \ker(f)=\{\overrightarrow{0}\}\) .

Proof (of the theorem 24)

Assume that \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is a linear mapping with matrix \( A=\begin{bmatrix} a&b\\c&d \end{bmatrix}\) in the canonical basis, and consider the determinant \( \det(A)=ad-bc\) of the matrix \( A\) .

Then the kernel of \( f\) is the set of vectors with cartesian coordinates solutions of the \( 2\times 2\) linear system:

\[ \begin{equation}\left\{\begin{matrix} ax&+&by&=&0\\ cx&+&dy&=&0 \end{matrix}\right.\end{equation} \]

(1)

  1. Assume that \( A=O_{22}\) .

    Then \( f=o\) the null mapping, so that \( \ker(f)=\mathbb{P}\) .

    Moreover, \( f\) is not invertible as it is equal to \( o\) .

  2. Assume that \( A\ne O_{22}\) and \( \det{A}=0\) .

    Then \( a\) , \( b\) , \( c\) and \( d\) are not zero together, and the solution of the sytem 1 is the following.

    • If \( a\) and \( b\) are not zero together, it is the straight line of equation \( ax+by=0\) .

    • And if \( a=b=0\) , it is the straight line of equation \( cx+dy=0\) .

  3. Assume that \( \det{A}\ne 0\) .

    Then the system 1 has a unique solution and, as \( (0,0)\) is a solution, it is the solution and \( \ker(f)=\{\overrightarrow{0}\}\) .

    Moreover, for any vector \( \overrightarrow{v}\) with cartesian coordinates \( (X,Y)\) , the vectors \( \overrightarrow{u})\) such that \( f(\overrightarrow{u}=\overrightarrow{v}\) are the vectors with cartesian coordinates solutions of the \( 2\times 2\) linear system:

    \[ \begin{equation}\left\{\begin{matrix} ax&+&by&=&X\\ cx&+&dy&=&Y \end{matrix}\right.\end{equation} \]

    (2)

As \( \det{A}\ne 0\) , that system has a unique solution, so that \( f\) is one-to-one from \( \mathbb{P}\) onto \( \mathbb{P}\) .

Consequently, \( f\) is invertible.

7.2 The Range of a Linear Mapping

7.2.1 Definition of the Range of a Linear Mapping

Definition 13

We define the range of a linear mapping \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) as the set of all vectors \( \overrightarrow{v}\in\mathbb{P}\) such that \( f(\overrightarrow{u})=\overrightarrow{v}\) for at least one vector \( \overrightarrow{u}\in\mathbb{P}\) :

\[ \text{range}(f)=\{\overrightarrow{v}\in\mathbb{P} : \exists \overrightarrow{u}\in\mathbb{P} \text{ such that }f(\overrightarrow{u})=\overrightarrow{v} \} \]

7.2.2 Examples of Ranges of Linear Mappings

The null mapping \( o\) As for any vector \( \overrightarrow{u}\in\mathbb{P}\) , \( o(\overrightarrow{u})=\overrightarrow{0}\) , the range of the null mapping \( o\) is the singleton \( \{\overrightarrow{0}\}\) .

Reversely, if the range of a linear mapping \( f\) is the singleton \( \{\overrightarrow{0}\}\) then, for any vector \( \overrightarrow{u}\in\mathbb{P}\) , \( f(\overrightarrow{u})=\overrightarrow{0}\) , so that \( f\) is the null mapping \( o\) .

Consequently, the null mapping \( o\) is the only mapping having \( \overrightarrow{u}\in\mathbb{P}\) as range.

The identity mapping \( \text{Id}_{\mathbb{P}}\) As \( \text{Id}_{\mathbb{P}}(\overrightarrow{u})=\overrightarrow{u}\) for any vector \( \overrightarrow{u}\) , then any vector of the plane is mapped by a vector by \( \text{Id}_{\mathbb{P}}\) .

Consequently, the range of the identity mapping \( \text{Id}_{\mathbb{P}}\) is the whole plane \( \mathbb{P}\) .

Homotheties with non zero factors As if \( \lambda\ne 0\) , \( h_{\lambda}(\frac{1}{\lambda}\overrightarrow{v})=\overrightarrow{v}\) for any vector \( \overrightarrow{v}\) , then any vector of the plane is mapped by a vector by \( h_{\lambda}\) .

Consequently, the range of the homothety \( h_{\lambda}\) with a non zero factor \( \lambda\in\mathbb{R}^*\) is the whole plane \( \mathbb{P}\) .

The rotations As if \( \theta\in\mathbb{R}^*\) , \( \rho_{\theta}(\rho_{(-\theta)}(\overrightarrow{v}))=\overrightarrow{v}\) for any vector \( \overrightarrow{v}\) , then any vector of the plane is mapped by a vector by \( \rho_{\theta}\) .

Consequently, the range of the rotation \( \rho_{\theta}\) is the whole plane \( \mathbb{P}\) .

The orthogonal projection on the \( x\) axis

Orthogonal projection on the x) axis
Figure 3. Orthogonal projection on the \( x\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_x(\overrightarrow{u})=\begin{bmatrix}x\\0\end{bmatrix}\) .

The range \( \text{range}({p_x})\) of \( p_{x}\) is the set of vectors \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) such that \( y=0\) .

Consequently, \( \text{range}({p_x})\) is the \( x\) axis.

The orthogonal projection on the \( y\) axis

Orthogonal projection on the y) axis
Figure 4. Orthogonal projection on the \( y\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_y(\overrightarrow{u})=\begin{bmatrix}0\\y\end{bmatrix}\) .

The range \( \text{range}({p_y})\) of \( p_{y}\) is the set of vectors \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) such that \( x=0\) .

Consequently, \( \text{range}({p_y})\) is the \( y\) axis.

7.2.3 Find the Range of a Linear Mapping

Theorem 25

Assume that \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is a linear mapping with matrix \( A=\begin{bmatrix} a&b\\c&d \end{bmatrix}\) in the canonical basis, and consider the determinant \( \det(A)=ad-bc\) of the matrix \( A\) .

Then the following assertions hold:

  1. If \( A=O_{22}\) , the null matrix, then \( \text{range}(f)=\{\overrightarrow{0}\}\) .

  2. If \( A\ne O_{22}\) and \( \det{A}=0\) , then \( \text{range}(f)\) is the straight line of equation \( cx-ay=0\) , unless \( a=c=0\) , in which case it is the straight line of equation \( dx-by=0\) .

  3. And if \( \det{A}\ne 0\) , then \( \text{range}(f)=\mathbb{P}\) , and in that case only, \( f\) is invertible.

The following corollary is a straightforward consequence of the theorem 25.

Corollary 11

The linear mapping \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is invertible if and only if \( \text{range}(f)=\mathbb{P}\) .

Proof (of the theorem 25)

Assume that \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is a linear mapping with matrix \( A=\begin{bmatrix} a&b\\c&d \end{bmatrix}\) in the canonical basis, and consider the determinant \( \det(A)=ad-bc\) of the matrix \( A\) .

Then the range of \( f\) is the set of vectors with cartesian coordinates \( (X,Y)\) such that the following \( 2\times 2\) linear system has at least one solution.

\[ \begin{equation}\left\{\begin{matrix} ax&+&by&=&X\\ cx&+&dy&=&Y \end{matrix}\right.\end{equation} \]

(3)

  1. Assume that \( A=O_{22}\) .

    Then \( f=o\) the null mapping, so that \( \text{range}(f)=\{\overrightarrow{0}\}\) .

    Moreover, \( f\) is not invertible as it is equal to \( o\) .

  2. Assume that \( A\ne O_{22}\) and \( \det{A}=0\) .

    Then \( a\) , \( b\) , \( c\) and \( d\) are not zero together, and that \( ad=bc\) , the system 3 has an infinity of solutions under the following conditions.

    • If \( a\ne 0\) and \( c\ne 0\) , then \( \frac{d}{c}=\frac{b}{a}\) , that we denote \( k\) .

      The system 3 becomes then:

      \[ \begin{equation} \left\{\begin{matrix} &a(x&+&ky)&=&X\\ &c(x&+&ky)&=&Y \end{matrix}\right. \end{equation} \]

      (4)

      Consequently, \( cX-aY=0\) .

      Reversely, if \( cX-aY=0\) , then \( cX=aY\) the first equation of 5 is equivalent to \( ca(x+ky)=cX\) and the second equation of 5 is equivalent to \( ac(x+ky)=aY\) , that is the same equation.

      Consequently, the range of \( f\) is the straight line of equation

      \( cx-ay=0\) .

    • If \( a\ne 0\) and \( c= 0\) , then \( d=0\) because \( ad=bc=0\) .

      The system 3 becomes then:

      \[ \begin{equation} \left\{\begin{matrix} &ax&=&X\\ &0&=&Y \end{matrix}\right. \end{equation} \]

      (5)

      Consequently, \( Y=0\) .

      Reversely, if \( Y=0\) , then the first equation of 5 is equivalent to \( ax=X\) , that has a solution in \( x\) because \( a\ne 0\) , and the second equation of 5 is equivalent to \( 0=0\) , that is always true.

      Consequently, the range of \( f\) is the \( x\) axis.

    • If \( a=c=0\) , \( b\ne 0\) and \( d\ne 0\) and the system 3 becomes:

      \[ \left\{\begin{matrix} &by&=&X\\&dy&=&Y \end{matrix}\right. \]

      That system has at least one solution if and only if \( \frac{X}{b}=\frac{Y}{d}\) , that is equivalent to \( dX-bY=0\) .

      Consequently, the range of \( f\) is the straight line of equation

      \( dx-by=0\) .

    • If \( a=c=b=0\) , then \( d\ne 0\) and the system 3 becomes:

      \[ \left\{\begin{matrix} &0&=&X\\&dy&=&Y \end{matrix}\right. \]

      Consequently, \( X=0\) .

      Reversely, if \( X=0\) , then the first equation of 5 is equivalent to \( 0=0\) , that is always true, and the second equation of 5 is equivalent to \( dy=Y\) , that has a solution in \( y\) because \( d\ne 0\) .

      Consequently, the range of \( f\) is the \( y\) axis.

    • Moreover, under the conditions \( A\ne O_{22}\) and \( \det{A}=0\) , \( f\) is not invertible, as we saw in the proof of the theorem 24.

  3. Assume that \( \det{A}\ne 0\) .

    Then the system 3 has a unique solution for any \( (X,Y)\) , and \( \text{range}(f)=\mathbb{P}\) .

    Moreover, under the condition \( \det{A}\ne 0\) , \( f\) is invertible, as we saw in the proof of the theorem 24.

7.3 The Rank of a Linear Mapping

7.3.1 Definition of the Rank of a Linear Mapping

Definition 14

We define the rank of a linear mapping as the dimension of its range, that is:

  • \( \text{rank}(f)=0\) if \( \text{range}(f)=\{\overrightarrow{0}\}\) .

  • \( \text{rank}(f)=1\) if \( \text{range}(f)\) is a straight line.

  • And \( \text{rank}(f)=0\) if \( \text{range}(f)=\mathbb{P}\) .

7.3.2 Examples of Ranks of Linear Mappings

We use the section 7.2.2 t derive th ranks of the following linear mappings.

The null mapping \( o\) As the range of the null mapping \( o\) is the singleton \( \{\overrightarrow{0}\}\) , \( \text{rank}(o)=0\) .

Moreover, the null mapping \( o\) is the only mapping having \( \overrightarrow{u}\in\mathbb{P}\) as range, it is the only mapping having \( 0\) as rank.

The identity mapping \( \text{Id}_{\mathbb{P}}\) As the range of the identity mapping \( \text{Id}_{\mathbb{P}}\) is the whole plane \( \mathbb{P}\) , \( \text{rank}(\text{Id}_{\mathbb{P}})=2\) .

Homotheties with non zero factors As the range of the homothety \( h_{\lambda}\) with a non zero factor \( \lambda\in\mathbb{R}^*\) is the whole plane \( \mathbb{P}\) , \( \text{rank}(h_{\lambda})=2\) .

The rotations As the range of the rotation \( \rho_{\theta}\) is the whole plane \( \mathbb{P}\) , \( \text{rank}(\rho_{\theta})=2\) .

The orthogonal projection on the \( x\) axis

Orthogonal projection on the x) axis
Figure 5. Orthogonal projection on the \( x\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_x(\overrightarrow{u})=\begin{bmatrix}x\\0\end{bmatrix}\) .

As \( \text{range}({p_x})\) is the \( x\) axis, that is a straight line, \( \text{rank}({p_x})=1\) .

The orthogonal projection on the \( y\) axis

Orthogonal projection on the y) axis
Figure 6. Orthogonal projection on the \( y\) axis

assigns to each vector \( \overrightarrow{u}=\begin{bmatrix}x\\y\end{bmatrix}\) , with \( (x,y)\in\mathbb{R}^2\) real numbers, the projected vector \( p_y(\overrightarrow{u})=\begin{bmatrix}0\\y\end{bmatrix}\) .

As \( \text{range}({p_y})\) is the \( y\) axis, that is a straight line, \( \text{rank}({p_y})=1\) .

7.3.3 Find the Rank of a Linear Mapping

The following theorem is a straightforward consequence of the theorem 25

Theorem 26

Assume that \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is a linear mapping with matrix \( A=\begin{bmatrix} a&b\\c&d \end{bmatrix}\) in the canonical basis, and consider the determinant \( \det(A)=ad-bc\) of the matrix \( A\) .

Then the following assertions hold:

  1. If \( A=O_{22}\) , the null matrix, then \( \text{rank}(f)=0\) .

  2. If \( A\ne O_{22}\) and \( \det{A}=0\) , then \( \text{rank}(f)=1\) .

  3. And if \( \det{A}\ne 0\) , then \( \text{rank}(f)=2\) , and in that case only, \( f\) is invertible.

The following corollary is a straightforward consequence of the theorem 26.

Corollary 12

The linear mapping \( f\;:\;\mathbb{P}\rightarrow\mathbb{P}\) is invertible if and only if \( \text{rank}(f)=2\) .

8 Conclusion

The linear mappings in \( \mathbb{P}\) , that preserve the addition and scalar multiplication of vectors are usefully characterised by their matrices in the canonical basis.

Namely, the following correspondences are met:

  • The addition of linear mappings corresponds the the addition of matrices.

  • The scalar multiplication of linear mappings corresponds to the scalar multiplication of matrices.

  • And the composition of linear mapping corresponds to the matrix multiplication of matrices.

This is useful to determine the sum of linear mappings, their product with a scalar and, last but not least, the composed linear mappings.

Moreover, the kernel, the range and the rank of a linear mapping are easily derived from its matrix.

In particular, the nullity or non nullity of the determinant of that matrix determines if the linear mapping is invertible of not.

I am normally hidden by the status bar