The method of solving systems of equations by matrices that we will look at is based on procedures involving equations that we are familiar with
from previous mathematics courses. The main idea is to reduce a given system of equations to another simpler system that has the same solutions.
Definition12.1.1Solution Set
Given a system of equations involving real variables \(x_1\), \(x_2 \ldots \), \(x_n\), the solution set of the
system is the set of \(n\)tuples in \(\mathbb{R}^n\), \(\left(a_1, a_2, \ldots ,a_n\right)\) such that the substitutions \(x_1= a_1\), \(x_2=
a_2, \ldots\), \(x_n= a_n\) make all the equations true.
In terms of logic, a solution set is a truth set of a system of equations, which is a proposition over \(n\)tuples of real numbers.
In general, if the variables are from a set \(S\), then the solution set will be a subset of \(S^n\). For example, in number theory mathematicians
study Diophantine equations, where the variables can only take on integer values instead of real values.
Definition12.1.2Equivalent Systems of Equations.
Two systems of linear equations are called equivalent if they have the same set of solutions.
Example12.1.3Two equivalent systems
The previous definition tells us that if we know that the system
\[\begin{array}{l}
4 x_1+2 x_2+x_3=1 \\
2 x_1+x_2+x_3=4 \\
2 x_1+2 x_2+x_3=3 \\
\end{array}\]
is equivalent to the system
\[\begin{array}{l}
\text{ }x_1+0 x_2+0x_3=1 \\
0 x_1+x_2\text{ }+0x_3=1 \\
0 x_1+0 x_2\text{ }+x_3= 7 \\
\end{array}\]
then both systems have the solution set \(\{(1, 1, 7)\}\). In other words, the simultaneous values \(x_1=1\), \(x_2= 1\), and \(x_3= 7\) are the only values
of the variables that make all three equations in either system true.
Theorem12.1.4Elementary Operations on Equations.
If any sequence of the following operations is performed on a system of equations, the
resulting system is equivalent to the original system:
Interchange any two equations in the system.
Multiply both sides of any equation by a nonzero constant.
Multiply both sides of any equation by a nonzero constant and add the result to a second equation in the system, with the sum replacing the latter
equation.
Let us now use the above theorem to work out the details of Example 12.1.3 and see how we can arrive at the simpler system.
The original system:
\begin{gather}
\begin{array}{c}
4 x_1+2 x_2+x_3=1 \\
2 x_1+x_2+x_3=4 \\
2 x_1+2 x_2+x_3=3
\end{array}\label{mrow45}\tag{12.1.1}
\end{gather}
Step 1. We will first change the coefficient of \(x_1\) in the first equation to one and then use it as a pivot to obtain 0's for the coefficients of \(x_1\) in Equations 2 and 3.
Multiply Equation 1 by \(\frac{1}{4}\) to obtain
\begin{gather}
\begin{array}{c}
x_1+\frac{x_2}{2}+\frac{x_3}{4}=\frac{1}{4}\\
2 x_1+x_2+x_3=4 \\
2 x_1+2 x_2+x_3=3\\\\
\end{array}\label{mrow46}\tag{12.1.2}
\end{gather}
Multiply Equation 1 by 2 and add the result to Equation 2 to obtain
\begin{gather}
\begin{array}{c}
x_1+\frac{x_2}{2}+\frac{x_3}{4}=\frac{1}{4} \\
0 x_1 + 0 x_2+ \frac{x_3}{2}=\frac{7}{2} \\
2 x_1+2 x_2+x_3=3 \\\\
\end{array}\label{mrow47}\tag{12.1.3}
\end{gather}
Multiply Equation 1 by 2 and add the result to Equation 3 to obtain
\begin{gather}
\begin{array}{c}
x_1+\frac{x_2}{2}+\frac{x_3}{4}=\frac{1}{4} \\
0 x_1 + 0 x_2+ \frac{x_3}{2}=\frac{7}{2} \\
0 x_1+ x_2+ \frac{x_3}{2}=\frac{5}{2} \\\\
\end{array}\label{mrow48}\tag{12.1.4}
\end{gather}
We've explicitly written terms with zero coefficients such as \(0 x_1\) to make a point that all variables can be thought of as being involved in all equations. After this example is complete, we will discontinue this practice in favor of the normal practice of making these terms “disappear.”
Step 2. We would now like to proceed in a fashion analogous to Step 1; namely, multiply the coefficient of \(x_2\) in the second equation by a suitable number so that the result is 1. Then use it as a pivot to obtain 0's as coefficients for \(x_2\) in the first and third equations. This is clearly impossible (Why?), so we will first interchange Equations 2 and 3 and proceed as outlined above.
Exchange Equations 2 and 3 to obtain
\begin{gather}
\begin{array}{c}
x_1+\frac{x_2}{2}+\frac{x_3}{4}=\frac{1}{4} \\
0 x_1+ x_2+ \frac{x_3}{2}=\frac{5}{2} \\
0 x_1 + 0 x_2+ \frac{x_3}{2}=\frac{7}{2} \\\\
\end{array}\label{mrow49}\tag{12.1.5}
\end{gather}
Multiply Equation 2 by \(\frac{1}{2}\) and add
the result to Equation 1 to obtain
\begin{gather}
\begin{array}{c}
x_1+ 0x_2 + 0x_3 = 1 \\
0 x_1+ x_2+ \frac{x_3}{2}=\frac{5}{2} \\
0 x_1 + 0 x_2+ \frac{x_3}{2}=\frac{7}{2} \\\\
\end{array}\label{mrow50}\tag{12.1.6}
\end{gather}
Step 3. Next, we will change the coefficient of \(x_3\) in the third equation to one and then use it as a pivot to obtain 0's for the coefficients of \(x_3\) in Equations 1 and 2. Notice that the coefficient of \(x_3\) is already zero in Equation 1, so we have been saved some work!
Multiply Equation 3 by 2 to obtain
\begin{equation*}
\begin{array}{c}
x_1+ 0x_2 + 0x_3 = 1 \\
0 x_1+ x_2+ \frac{x_3}{2}=\frac{5}{2} \\
0 x_1 + 0 x_2+ x_3=7
\end{array}
\end{equation*}
Multiply Equation 3 by 1/2 and add the result to Equation 2 to obtain
\begin{gather}
\begin{array}{c}
x_1+ 0x_2 + 0x_3 = 1 \\
0 x_1+ x_2+ 0 x_3=1 \\
0 x_1 + 0 x_2+ x_3=7 \\
\end{array}\label{sys121finalsystem}\tag{12.1.7}
\end{gather}
From the system of equations at the end of Step 3, we see that the solution to the original system is \(x_1=1\), \(x_2= 1\), and \(x_3= 7\).
In the above sequence of steps, we note that the variables serve the sole purpose of keeping the coefficients in the appropriate location. This we
can effect by using matrices. The matrix of the original system in our example is
\[\left(
\begin{array}{cccc}
4 & 2 & 1 & 1 \\
2 & 1 & 1 & 4 \\
2 & 2 & 1 & 3 \\
\end{array}
\right)\]
where the matrix of the first three columns is called the coefficient matrix and the complete matrix is referred to as the augmented matrix. Since we are now using matrices to solve the system, we will translate Theorem 12.1.4 into matrix language.
Theorem12.1.5Elementary Row Operations
If any sequence of the following operations is performed on the augmented matrix of a system of equations, the resulting matrix is a system that is equivalent to the original system.
The following operations on a matrix are called elementary row operations:
Exchange any two rows of the matrix.
Multiply any row of the matrix by a nonzero constant.
Multiply any row of the matrix by a nonzero constant and add the result to a second row, with the sum replacing that second row.
Definition12.1.6Row Equivalent Matrices
Two matrices, \(A\) and \(B\), are said to be rowequivalent if one can be obtained from the other by any sequence of zero or more elementary row operations.
If we use the notation \(R_i\) to stand for Row \(i\) of a matrix and \(\longrightarrow\) to stand for row equivalence, then
\[A \overset{c R_i+ R_j}{\longrightarrow }B\]
means that the matrix \(B\) is obtained from the matrix \(A\) by multiplying the Row \(i\) of \(A\) by \(c\) and adding the result to Row \(j\). The operation of multiplying row \(i\) by \(c\) is indicated by
\[A \overset{c R_i}{\longrightarrow }B\]
while exchanging rows \(i\) and \(j\) is denoted by
\[A \overset{R_i\leftrightarrow R_j}{\longrightarrow }B\].
The matrix notation for the system given in our first example, with the subsequent steps, is:
\begin{equation*}
\begin{split}
\left(
\begin{array}{cccc}
4 & 2 & 1 & 1 \\
2 & 1 & 1 & 4 \\
2 & 2 & 1 & 3 \\
\end{array}
\right) & \overset{\frac{1}{4} R_1}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & \frac{1}{2} & \frac{1}{4} & \frac{1}{4} \\
2 & 1 & 1 & 4 \\
2 & 2 & 1 & 3 \\
\end{array}
\right) \text{ }\overset{2 R_1+ R_2}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & \frac{1}{2} & \frac{1}{4} & \frac{1}{4} \\
0 & 0 & \frac{1}{2} & \frac{7}{2} \\
2 & 2 & 1 & 3 \\
\end{array}
\right) \\
& \overset{2 R_1+ R_3}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & \frac{1}{2} & \frac{1}{4} & \frac{1}{4} \\
0 & 0 & \frac{1}{2} & \frac{7}{2} \\
0 & 1 & \frac{1}{2} & \frac{5}{2} \\
\end{array}
\right)\text{ }\overset{R_2\leftrightarrow R_3}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & \frac{1}{2} & \frac{1}{4} & \frac{1}{4} \\
0 & 1 & \frac{1}{2} & \frac{5}{2} \\
0 & 0 & \frac{1}{2} & \frac{7}{2} \\
\end{array}
\right) \\
& \overset{\frac{1}{2} R_2+ R_1}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & 0 & 0 & 1 \\
0 & 1 & \frac{1}{2} & \frac{5}{2} \\
0 & 0 & \frac{1}{2} & \frac{7}{2} \\
\end{array}
\right)\text{ }\overset{2 R_3}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & 0 & 0 & 1 \\
0 & 1 & \frac{1}{2} & \frac{5}{2} \\
0 & 0 & 1 & 7 \\
\end{array}
\right) \text{ }\\
& \overset{\frac{1}{2} R_3+ R_2}{\text{ }\longrightarrow }\text{ }\left(
\begin{array}{cccc}
1 & 0 & 0 & 1 \\
0 & 1 & 0 & 1 \\
0 & 0 & 1 & 7 \\
\end{array}
\right)
\end{split}
\end{equation*}
This again gives us the solution. This procedure is called the GaussJordan elimination method.
It is important to remember when solving any system of equations via this or any similar approach that at any step in the procedure we can rewrite the matrix in “equation format” to help us to interpret the meaning of the augmented matrix.
In our first example we found a unique solution, only one triple, namely \((1,1,7)\), which satisfies all three equations. For a system involving three unknowns, are there any other possible results? To answer this question, let's review some basic facts from analytic geometry.
The graph of a linear equation in threedimensional space is a plane. So geometrically we can visualize the three linear equations as three planes in threespace. Certainly the three planes can intersect in a unique point, as in the first example, or two of the planes could be parallel. If two planes are parallel, there are no common points of intersection; that is, there are no triple of real numbers that will satisfy all three equations. Another possibility is that the three planes could intersect along a common axis or line. In this case, there would be an infinite number of real number triples in \(\mathbb{R}^3\). Yet another possiblity would be if the first two planes intersect in a line, but the line is parallel to, but not on, the third plane, giving us no solution. Finally if all three equations describe the same plane, the solution set would be that plane.
We can generalize these observations. In a system of \(n\) linear equations, \(n\) unknowns, there can be
a unique solution,
no solution, or
an infinite number of solutions.
To illustrate these points, consider the following examples:
Example12.1.7A system with no solutions.
Find all solutions to the system
\[\begin{array}{l}
\text{ }x_1 +3 x_2\text{ }+x_3=2 \\
\text{ }x_1\text{ }+x_2 +5 x_3=4 \\
2 x_1+2 x_2+10 x_3=6 \\
\end{array}\]
The reader can verify that the augmented matrix of this system,
\(\left(
\begin{array}{cccc}
1 & 3 & 1 & 2 \\
1 & 1 & 5 & 4 \\
2 & 2 & 10 & 6 \\
\end{array}
\right)\),
reduces to
\(\left(
\begin{array}{cccc}
1 & 3 & 1 & 2 \\
1 & 1 & 5 & 4 \\
0 & 0 & 0 & 2 \\
\end{array}
\right)\).
We can attept rowreduce this matrix further if we wish. However, any further rowreduction will not substantially change the last row, which, in equation
form, is \(0x_1+ 0x_2+0x_3 = 2\), or simply \(0=2\). It is clear that we cannot find real numbers \(x_1\), \(x_2\), and \(x_3\) that will satisfy
this equation. Hence we cannot find real numbers that will satisfy all three original equations simultaneously. When this occurs, we say that the system has no solution, or the solution set is empty.
Example12.1.8A system with an infinite number of solutions
Next, let's attempt to find all of the solutions to:
\[\begin{array}{l}
\text{ }x_1+6 x_2+2 x_3=1 \\
2 x_1\text{ }+x_2+3 x_3 =2 \\
4 x_1+2 x_2+6 x_3=4 \\
\end{array}\]
The augmented matrix for the system is
\begin{align}
\left(
\begin{array}{cccc}
1 & 6 & 2 & 1 \\
2 & 1 & 3 & 2 \\
4 & 2 & 6 & 4 \\
\end{array}
\right)\label{sys1213startsystem}\tag{12.1.8}
\end{align}
which reduces to
\begin{align}
\left(
\begin{array}{cccc}
1 & 0 & \frac{16}{11} & 1 \\
0 & 1 & \frac{1}{11} & 0 \\
0 & 0 & 0 & 0 \\
\end{array}
\right)\label{sys1213finalsystem}\tag{12.1.9}
\end{align}
If we apply additional elementary row operations to this matrix, it will only become more complicated. In particular, we cannot get a one in the
third row, third column. Since the matrix is in simplest form, we will express it in equation format to help us determine the solution set.
\begin{gather}
\begin{array}{l}
x_1 \quad+\frac{16}{11} x_3 =1 \\
\quad x_2+\frac{1}{11}x_3 =0\\
\quad\quad\quad \quad 0=0 \\
\end{array}\label{sys1213reduced}\tag{12.1.10}
\end{gather}
Any real numbers will satisfy the last equation. However, the first equation can be rewritten as \(x_1 =1\frac{16 }{11}x_3\), which describes the coordinate \(x_1\) in terms of \(x_3\) . Similarly, the second equation gives \(x_1\) in terms of \(x_3\) . A convenient way of listing the solutions
of this system is to use set notation. If we call the solution set of the system \(S\), then
\[S = \left\{\left(1\frac{16}{11}x_3, \frac{1}{11}x_3, x_3\right) \mid x_3\in \mathbb{R}\right\}\].
What this means is that if we wanted to list all solutions, we would replace \(x_3\) by all possible numbers. Clearly, there is an infinite number of solutions, two of which are \((1, 0, 0)\) and \((15, 1, 11)\), when \(x_3\) takes on the values 0 and 11, respectively.
A Word Of Caution: Frequently we may can get “differentlooking” answers to the same problem when a system has an infinite number of solutions. Assume the solutions set in this example is reported to be \(A = \left\{\left(1+16x_2, x_2, 11x_3\right) \mid x_3\in \mathbb{R}\right\}\).
Certainly the result described by \(S\) looks different from that described by \(A\). To see whether they indeed describe the same set,
we wish to determine whether every solution produced in \(S\) can be generated in \(A\). For example, the solution generated by \(S\) when \(x_3=11\) is \((15, 1, 11)\). The same triple can be produced by \(A\) by taking \(x_2= 1\). We must prove that every solution described in \(S\) is described in \(A\) and, conversely, that every solution described in \(A\) is described in \(S\). (See Exercise 6 of this section.)
To summarize the procedure in the GaussJordan technique for solving systems of equations, we attempt to obtain 1's along the main diagonal of the coefficient matrix with 0's above and below the diagonal. We may find in attempting this that this objective cannont be completed, as in the last two examples we have seen. Depending on the way we interpret the results in equation form, we either recognize that no solution exists, or we idenfity “free variables” on which an infinite number of solutions are based. The final matrix forms that we have produced in our examples are referred to as echelon forms.
In practice, larger systems of linear equations are solved using computers. Generally, the GaussJordan algorithm is the most useful; however, slight
variations of this algorithm are also used. The different approaches share many of the same advantages and disadvantages. The two major concerns
of all methods are:
minimizing inaccuracies due to roundoff errors, and
minimizing computer time.
The accuracy of the GaussJordan method can be improved by always choosing the element with the largest absolute value as the pivot element, as in the following algorithm.
Algorithm12.1.9The GaussJordan Algorithm
Given a matrix equation \(A x = b\), where A is \(n \times m\), let \(C\) be the augmented matrix \([A  b]\). The process of rowreducing to echelon form involves performing the following algorithm where \(C[i]\) id the \(i^{\textrm{th}}\) row of \(C\).
i = 1
j = 1

while i <= n and j <= m):
maxi=i

for k = i+1 to n:
\(\quad\)if abs(C[k,j])>abs(C[maxi,j]): then maxi=k

if C[maxi,j] != 0 then:
exchange rows i and maxi
divide each entry in row i by C[i,j]

for u = i+1 to n:
\(\quad\)subtract C[u,j]*C[i] from C[u]
i = i+1
j=j+1
Example12.1.11
If we apply Algorithm 12.1.1 to the system
\[\begin{array}{l}
5 x_1+x_2+2 x_3+x_4=2 \\
3 x_1+x_22 x_3\quad \quad=5 \\
x_1+x_2+3 x_3x_4=1 \\
\end{array}\]
the augmented matrix is
\[\left(
\begin{array}{ccccc}
5 & 1 & 2 & 1 & 2 \\
3 & 1 & 2 & 0 & 5 \\
1 & 1 & 3 & 1 & 1 \\
\end{array}
\right)\]
is reduced to
\[\left(
\begin{array}{ccccc}
1 & 0 & 0 & \frac{1}{2} & \frac{1}{2} \\
0 & 1 & 0 & \frac{3}{2} & \frac{3}{2} \\
0 & 0 & 1 & 0 & 1 \\
\end{array}
\right)\]
Therefore, \(x_4\) is a free variable in the solution and general solution of the system is
\[x =\left(
\begin{array}{c}
x_1 \\
x_2 \\
x_3 \\
x_4 \\
\end{array}
\right)=\text{ }\left(
\begin{array}{c}
\frac{1}{2}\frac{1}{2}x_4 \\
\frac{3}{2}+\frac{3}{2}x_4 \\
1 \\
x_4 \\
\end{array}
\right)\]
This conclusion is easy to see if you revert back to the equations that the final value the reduced matrix represents.
1
Solve the following systems by describing the solution sets completely:
\(\begin{array}{l}
2 x_1+x_2=3 \\
x_1x_2=1 \\
\end{array}\)
\(\begin{array}{l}
2 x_1+x_2+3 x_3=5 \\
4 x_1+x_2+2 x_3=1 \\
8 x_1+2 x_2+4 x_3=2 \\
\end{array}\)
\(\begin{array}{l}
x_1+x_2+2 x_3=1 \\
x_1+2 x_2x_3=1 \\
x_1+3 x_2+x_3=5 \\
\end{array}\)
\(\begin{array}{l}
x_1x_2+3 x_3=7 \\
x_1+3 x_2+x_3=4 \\
\end{array}\)
Answer
\(\{(4/3, 1/3)\}\)
\(\left\{\left(3  0.5x_3, 11  4x_3, x_3 \right) \mid x_3\right\}\)
\(\{(5, 14/5, 8/5)\}\)
\(\left\{\left(6.25  2.5x_3, 0.75 + 0.5x_3 , x_3\right) \mid x_3 \in \mathbb{R}\right\}\)
2
Solve the following systems by describing the solution sets completely:
\(\begin{array}{l}
2 x_1+2 x_2+4 x_3=2 \\
2 x_1+x_2+4 x_3=0 \\
3 x_1+5 x_2+x_3=0 \\
\end{array}\)
\(\begin{array}{l}
2 x_1+x_2+3 x_3=2 \\
4 x_1+x_2+2 x_3=1 \\
8 x_1+2 x_2+4 x_3=4 \\
\end{array}\)
\(\begin{array}{l}
x_1+x_2+2 x_3+x_4=3 \\
x_1x_2+3 x_3x_4=2 \\
3 x_1+3 x_2+6 x_3+3 x_4=9 \\
\end{array}\)
\(\begin{array}{l}
6 x_1+7 x_2+2 x_3=3 \\
4 x_1+2 x_2+x_3=2 \\
6 x_1+x_2+x_3=1 \\
\end{array}\)
\(\begin{array}{l}
x_1+x_2x_3+2 x_4=1 \\
x_1+2 x_2+3 x_3+x_4=5 \\
x_1+3 x_2+2 x_3x_4=1 \\
\end{array}\)
3
Given the final augmented matrices below from the GaussJordan Algorithm, identify the solutions sets. Identify the basic and free variables, and describe the solution set of the original system.
\(\left(
\begin{array}{ccccc}
1 & 0 & 5 & 0 & 1.2 \\
0 & 1 & 4 & 0 & 2.6 \\
0 & 0 & 0 & 1 & 4.5 \\
\end{array}
\right)\)
\(\left(
\begin{array}{cccc}
1 & 0 & 9 & 3 \\
0 & 1 & 0 & 4 \\
0 & 0 & 0 & 1 \\
\end{array}
\right)\)
\(\left(
\begin{array}{cccc}
1 & 0 & 6 & 5 \\
0 & 1 & 2 & 1 \\
0 & 0 & 0 & 0 \\
\end{array}
\right)\)
\(\left(
\begin{array}{ccccc}
1 & 0 & 0 & 3 & 1 \\
0 & 1 & 0 & 2 & 2 \\
0 & 0 & 1 & 1 & 1 \\
\end{array}
\right)\)
Answer
\(\{(1.2, 2.6, 4.5)\}\)
\(\left\{\left(6 x_3+ 5, 2 x_3 + 1, x_3 \right) \mid x_3 \in \mathbb{R}\right\}\)
\(\left\{\left(9 x_3 + 3, 4, x_3 \right) \mid x_3 \in \mathbb{R}\right\}\)
\(\left\{\left(3 x_4 + 1, 2x_4 + 2, x_4 + 1, x_4\right) \mid x_4 \in \mathbb{R}\right\}\)
4
Write out the details of Example 12.1.7.
Write out the details of Example 12.1.8.
Write out the details of Example 12.1.11.
5
Solve the following systems using only mod 5 arithmetic. Your solutions should be \(n\textrm{tuples}\) from \(\mathbb{Z}_5\).
\(\begin{array}{l}
2 x_1+ x_2=3 \\
x_1+4 x_2=1 \\
\end{array}\) (compare your solution to the system in 5(a))
\(\begin{array}{l}
x_1+x_2+2 x_3=1 \\
x_1+2 x_2+4 x_3=4 \\
x_1+3 x_2+3 x_3=0 \\
\end{array}\)
Answer
\(\{(3,0)\}\)
\(\{(3,0,4)\}\)
6
Use the solution set \(S\) of Example 12.1.8 to list three different solutions to the given system. Then show that each of these solutions
can be described by the set \(A\) in the same example.
Prove that \(S = A\).
7
Given a system of \(n\) linear equations in \(n\) unknowns in matrix form \(A X = b\), prove that if \(b\) is a matrix of all
zeros, then the solution set of \(A X = b\) is a subgroup of \(\mathbb{R}^n\).
Answer Proof: Since \(b\) is the \(n\times 1\) matrix of 0{'}s, let's call it \(\pmb{0}\). Let S be the set of solutions to \(A X = \pmb{0}\). If \(X_1\)
and \(X_2\) be in \(S\). Then
\[A\left(X_1 + X_2 \right) = A X_1 + A X_2 = \pmb{0} + \pmb{0} = \pmb{0}\]
so \(X_1+ X_2 \in S\); or in other words, \(S\) is closed under addition in \(\mathbb{R}^n\).
The identity of \(\mathbb{R}^n\) is \pmb{ 0}, which is in \(S\). Finally, let \(X\) be in\(S\). Then
\[A(X) = (A X) =  \pmb{0} = \pmb{0}\]
and so \(X\) is also in \(S\).