5.4. Miscellaneous Applications of Determinants#

In this section we will address the following matters:

  • The determinant as a uniform scale factor for an arbitrary linear transformation from \(\R^n\) to \(\R^n\) .

  • Cramer’s rule. Seemingly the ultimate solution to almost all systems of \(n\) linear equations in \(n\) unknowns.

  • The generalization of the formula

    \[\begin{split} \left[\begin{array}{cc} a & b \\ c & d\end{array} \right]^{-1} = \dfrac{1}{ad-bc} \left[\begin{array}{cc} d & -b \\ -c & a\end{array} \right] \end{split}\]

    to \(n\times n\) matrices.

  • A certain generalization of the cross product to \(n\) dimensions.

5.4.1. Volume and Orientation Revisited#

We have seen in Section 5.1 how determinants arise in the context of areas of parallelograms and volumes of parallelepipeds.

In Section 1.2 we used the dot product to define length, distance and orthogonality in \(\R^n\). Determinants permit to define the concepts of volume and orientation in \(n\) dimensions.

Definition 5.4.1 (Volume in \(\R^n\))

Let \(\{\vect{v_1}, \ldots, \vect{v}_n\}\) be a set of \(n\) vectors in \(\R^n\). The \(n\)-dimensional parallelepiped \(\mathcal{P}\) spanned by \(\vect{v_1}, \ldots, \vect{v}_n\) is the set

\[ \mathcal{P} = \{c_1\vect{v}_1+c_2\vect{v}_2 + \ldots + c_1\vect{v}_n \,|\, c_i \in \R, 0 \leq c_i \leq 1\}. \]

See Figure 5.4.1 for an illustration of such a set in \(\R^2\).

The volume of such a parallelepiped is defined by

\[ \text{Vol}(\mathcal{P}) = |\det{[\,\vect{v_1}\, \,\ldots \,\, \vect{v}_n\,]}|. \]

So, it is the absolute value of a determinant.

Note that if the vectors \(\vect{v}_1, \ldots, \vect{v}_n\) in Definition 5.4.1 are linearly dependent the volume automatically becomes 0.

../_images/Fig-DetExtras-ParPed.svg

Fig. 5.4.1 The parallelepiped generated by two vectors.#

Definition 5.4.2 (Orientation in \(\R^n\))

Suppose the vectors \((\vect{v_1}, \ldots, \vect{v}_n)\) in \(\R^n\) are linearly independent.

Then we say that the ordered set \((\vect{v_1}, \ldots, \vect{v}_n)\) is positively oriented   if   \( \det{[\vect{v_1} \ldots \vect{v}_n]}>0\).

If this determinant is negative the set is called negatively oriented.

For vectors that are linearly dependent we do not define the orientation.

Proposition 5.4.1

Suppose \(T\) is a linear transformation from \(\R^2\) to \(\R^2\), with standard matrix \(A = [\,\vect{a}_1 \,\, \vect{a}_2\,]\). So we have

\[ T(\vect{x}) = A\vect{x}, \quad \text{for} \,\, \vect{x} \,\text{ in }\, \R^2. \]

Let \(R\) be any region in \(\R^2\) for which the area is well-defined, and let \(S\) be the image of \(R\) under \(T\).

Then for the area of \(S\) it holds that

\[ \text{area}(S) = |\det{A}|\cdot \text{area}(R). \]

Proof of Proposition 5.4.1

If the matrix \(A\) is not invertible, the range of \(T\), which is given by \(\text{span}\{\vect{a}_1, \vect{a}_2\}\), is contained in a line. Each region \(R\) is then mapped onto a subset \(S\) that is contained in this line, so

\[ \text{area}(S) = 0 = 0\cdot \text{area}(R) =|\det{A}|\cdot \text{area}(R). \]

Next suppose that \(A\) is invertible. Then the unit grid is mapped onto a grid with as unit region the parallelogram with sides \(\vect{a}_1 = A\vect{e}_1\) and \(\vect{a}_2 = A\vect{e}_2\) . See Figure 5.4.2.

../_images/Fig-DetExtras-StandardGrid.svg

Fig. 5.4.2 The image of the standard grid.#

First we show that the formula holds if \(R\) is the unit square, i.e., the parallelogram generated by \(\vect{e}_1\) and \(\vect{e}_2\). The unit square is mapped onto the parallelogram \(S\) generated by \(T(\vect{e}_1)=\vect{a}_1\) and \(T(\vect{e}_2)=\vect{a}_2\). It follows that

\[ \text{area}(S) = |\det{[\,\vect{a}_1\,\, \vect{a}_2\,]}| = |\det{A}|,\]

and since the area of \(R\) is equal to 1, we have

\[ |\det{A}| = |\det{A}| \cdot 1 = |\det{A}| \cdot \text{area}(R). \]

This then also holds for any square \(R\) with sides of length \(r\) that are parallel to the axes. Namely, such a square has area \(r^2\) and can be described as the square with vertices

\[ \vect{p},\quad \vect{p}+ r\vect{e}_1, \quad \vect{p}+ r\vect{e}_1+r\vect{e}_2 \quad\text{and}\quad \vect{p}+ r\vect{e}_2. \]

These are mapped to

\[ A\vect{p}, A\vect{p}+ rA\vect{e}_1, A\vect{p}+ rA\vect{e}_1+rA\vect{e}_2 \quad\text{and}\quad A\vect{p}+ rA\vect{e}_2. \]

This is a parallelogram with sides \(rA\vect{e}_1 = r\vect{a}_1\) and \(rA\vect{e}_2 =r \vect{a}_2\), which has area

\[ \text{area}(S) = |\det{[\,r\vect{a}_1\,\, r\vect{a}_2\,]}| = r^2 |\det{A}| = |\det{A}|\cdot \text{area}(R). \]

See Figure 5.4.3

../_images/Fig-DetExtras-ImageOfSquare.svg

Fig. 5.4.3 The image of a square with ‘corner’ \(\vect{p}\) and side length \(r\).#

For a general (reasonable) region \(R\) we sketch the idea and omit the technical details.

The region \(R\) can be approximated arbitrarily close by a collection of smaller and smaller squares \(R_i\) of which the interiors do not overlap. See Figure 5.4.4. The limit of the areas of these approximations when the grids get finer and finer gives the area of \(R\).

../_images/Fig-DetExtras-Subdivision.svg

Fig. 5.4.4 Approximating a region by smaller and smaller squares.#

The formula holds for each of the \(R_i\). Since \(T\) is one-to-one, the images \(S_i = T(R_i)\) will not overlap either, and the images taken together will approximate the image \(S = T(R)\) as well. We deduce that

\[\begin{split} \begin{array}{rcl} \text{area}(S) \approx \sum \text{area}(S_i) &=& \sum |\det{A}|\cdot \text{area}(R_i) \\ &=& |\det{A}| \sum \text{area}(R_i) \approx |\det{A}|\cdot \text{area}(R). \end{array} \end{split}\]

By taking an appropriate limit one can show that in fact

\[ \text{area}(S) = |\det{A}|\cdot \text{area}(R). \]

Proposition 5.4.3 can be generalized to higher dimensions.
For \(n = 3\) area becomes volume, and for higher dimensions we use the definition of \(n\)-dimensional volume as in Definition 5.4.1.

Proposition 5.4.2

Suppose \(T\) is a linear transformation from \(\R^n\) to \(\R^n\), with standard matrix \(A\).

Then for any region \(S\) in \(\R^n\) for which the volume is well-defined, it holds that

\[ \text{Vol}(S) = |\det{A}|\cdot \text{Vol}(R), \]

where \(S\) is the image of \(R\) under \(T\).

Proof of Proposition 5.4.2


If \(R\) is the \(n\)-dimensional parallelepiped \(\mathcal{P}\) generated by \(\{\vect{v}_1, \ldots, \vect{v}_n\}\) we have that \(T(\mathcal{P})\) is generated by \(\{T(\vect{v}_1), \ldots, T(\vect{v}_n)\}\).

Then

\[\begin{split} \begin{array}{rcl} \text{Vol}(T(\mathcal{P})) &=& |\det{[\,T(\vect{v_1})\, \,\ldots \,\, T(\vect{v}_n)\,]}| \\ &=& |\det{[\,A(\vect{v_1})\, \,\ldots \,\, A(\vect{v}_n)\,]}| \\ &=& |\det{\left(A [\,\vect{v_1}\, \,\ldots \,\, \vect{v}_n\,]\right)}|\\ &=& |\det{A} \det{[\,\vect{v_1}\, \,\ldots \,\, \vect{v}_n\,]}|\\ &=& |\det{A}| \text{Vol}(\mathcal{P}). \end{array} \end{split}\]

For a more general region \(R\) we would again have to work with approximations/subdivisions like in the proof of Proposition 5.4.3. Then we would first have to extend the definition of \(n\)-dimensional volume. We will not pursue that track.

To conclude our interpretation of the determinant of \(A\) regarding the linear transformation \(T(\vect{x}) = A\vect{x}\) we look at the orientation.

Proposition 5.4.3

Suppose \(A = [\,\vect{a}_1\,\,\vect{a}_2\,\,\ldots\,\,\vect{a}_n\, ]\) is the standard matrix of the linear transformation \(T: \R^n \to R^n\). So we have

\[ T(\vect{x}) = A\vect{x}. \]

Suppose \((\vect{v}_1,\,\vect{v}_2,\,\ldots\,,\,\vect{v}_n)\) is an ordered set of vectors in \(\R^n\).

Then the following holds.

If \(\det{A} > 0\) the set \(\big(T(\vect{v}_1),\,T(\vect{v}_2),\,\ldots\,,\,T(\vect{v}_n)\big)\) has the same orientation as the set \((\vect{v}_1,\,\vect{v}_2,\,\ldots\,,\,\vect{v}_n)\).

If \(\det{A} < 0\) the set \(\big(T(\vect{v}_1),\,T(\vect{v}_2),\,\ldots\,,\,T(\vect{v}_n)\big)\) has the opposite orientation as the set \((\vect{v}_1,\,\vect{v}_2,\,\ldots\,,\,\vect{v}_n)\).

In short: the transformation \(T(\vect{x}) = A\vect{x}\) preserves the orientation if
\(\det{A} > 0\) and reverses the orientation if \(\det{A} < 0\).

If the determinant is 0, then the set \(\{T(\vect{v}_1), \ldots,T(\vect{v}_n) \}\) will be linearly dependent, and for such a set the orientation is not defined.

Proof of Proposition 5.4.3

This too follows immediately from the product rule of determinants.

\[\begin{split} \begin{array}{rcl} \det{\left[\,T(\vect{v}_1)\,\,T(\vect{v}_2)\,\,\ldots\,\,T(\vect{v}_n)\, \right]} &=& \det{\left[\,A\vect{v}_1\,\,A\vect{v}_2\,\,\ldots\,\,A\vect{v}_n\, \right]} \\ &=& \det{\big(A\left[\,\vect{v}_1\,\,\vect{v}_2\,\,\ldots\,\,\vect{v}_n\, \right]\big)} \\ &=& \det{A}\cdot\det{\left[\,\vect{v}_1\,\,\vect{v}_2\,\,\ldots\,\,\vect{v}_n\, \right]}. \end{array} \end{split}\]

A nice illustration of what this means in \(\R^2\) is given by the following example.

Example 5.4.1

Consider the two linear transformations from \(\R^2\) to \(\R^2\) with matrices

\[\begin{split} A = \begin{bmatrix} 1 & 3 \\ 3 & 1 \end{bmatrix}, \quad B = \begin{bmatrix} 3 & 1 \\ 1 & 3 \end{bmatrix}. \end{split}\]

Note that

\[ \det{A} = -8 < 0 \quad \text{and} \quad \det{B} = 8 > 0. \]

Figure 5.4.5 visualizes what is going on.

../_images/Fig-DetExtras-Orientation.svg

Fig. 5.4.5 Images under transformations with negative and positive determinant.#

The images of a unit vector that rotates counterclockwise under transformation \(A\) move around clockwise, i.e., in the opposite orientation/direction. Under transformation \(B\) the images will go around the origin counterclockwise, i.e., in the same direction as the original vectors.

5.4.2. Cramer’s rule#

We first introduce a new notation that will help to simplify formulas later.

Definition 5.4.3

Let \(A\) be an \(n\times n\) matrix, and \(\vect{v}\) a vector in \(\R^n\). Then \(A^{(i)}(\vect{v})\) denotes the matrix that results when the \(i\)th column of \(A\) is replaced by the vector \(\vect{v}\).

Example 5.4.2

For the matrix \(A = \begin{bmatrix} 1 & 3 & 1 \\ 1 & 4 & 2 \\ 3 & 1 & 5 \end{bmatrix}\) and the vector \(\vect{v} = \begin{bmatrix} \class{blue}6 \\ \class{blue}7 \\ \class{blue}8 \end{bmatrix}\) we have that

\[\begin{split} A^{(2)}(\vect{v}) = \begin{bmatrix} 1 & \class{blue}6 & 1 \\ 1 & \class{blue}7 & 2 \\ 3 & \class{blue}8 & 5 \end{bmatrix}. \end{split}\]

Suppose that \(A\) is an invertible \(n \times n\) matrix. Then we know that the linear system \(A\vect{x} = \vect{b}\) has a unique solution for each \(\vect{b}\) in \(\R^n\). And we also know that the determinant of \(A\) is not equal to zero.
The next proposition gives a ‘ready made’ formula for the solution.

Theorem 5.4.1 (Cramer’s Rule)

Suppose \(A\) is an invertible \(n \times n\) matrix, and \(\vect{b}\) a vector in \(\R^n\). The entries of \(x_i\) of the unique solution \(\vect{x}\) of the linear system

\[ A\vect{x} = \vect{b} \]

are given by

(5.4.1)#\[x_i = \dfrac{\det{\left(A^{(i)}(\vect{b})\right)}}{\det{A}}.\]

Example 5.4.3

We use Cramer’s rule to solve the system

\[\begin{split} \left\lbrace \begin{array}{rcc} x_1 + 2x_2 + x_3 & = & 3 \\ x_1 - x_2 + 2x_3 & = & 4 \\ 3x_1 + x_2 -5x_3 & = & 1 \end{array} \right. \quad\text{i.e.} \quad \begin{bmatrix} 1 & 2 & 1 \\ 1 & -1 & 2 \\ 3 & 1 & -5 \end{bmatrix} \left[\begin{array}{c} x_1 \\ x_2 \\ x_3 \end{array} \right] = \begin{bmatrix}3 \\ 4 \\ 1 \end{bmatrix}. \end{split}\]

First of all, the determinant of \(A\) can be computed as follows (in the first step we use column reduction, with the boxed 1 as a pivot):

\[\begin{split} \left|\begin{array}{ccc} \fbox{1} & 2 & 1 \\ 1 & -1 & 2 \\ 3 & 1 & -5 \end{array} \right|= \left|\begin{array}{ccc} 1 & 0 & 0 \\ 1 & -3 & 1 \\ 3 & -5 & -8 \end{array} \right|= \left|\begin{array}{cc} -3 & 1 \\ -5 & -8 \end{array} \right|= 29 \neq 0, \end{split}\]

so the coefficient matrix is invertible and consequently the system has a unique solution.

According to Cramer’s rule we find the first entry of the solution as follows (again we use the boxed 1 as a pivot):

\[\begin{split} x_1 = \dfrac{\begin{vmatrix} 3 & 2 & 1 \\ 4 & -1 & 2 \\ 1 & \fbox{1} & -5 \end{vmatrix}}{29} = \dfrac{\begin{vmatrix} 1 & 0 & 11 \\ 5 & 0 & -3 \\ 1 & 1 & -5 \end{vmatrix}}{29} = \dfrac{-\begin{vmatrix} 1 & 11 \\ 5 & -3 \end{vmatrix}}{29} = \dfrac{58}{29} = 2. \end{split}\]

Likewise we can compute the other two entries of the solution.

\[\begin{split} x_2 = \dfrac{\begin{vmatrix} 1 & 3 & 1 \\ 1 & 4 & 2 \\ 3 & 1 & -5 \end{vmatrix}}{29} = 0 \quad \text{and} \quad x_2 = \dfrac{\begin{vmatrix} 1 & 2 & 3 \\ 1 & -1 & 4 \\ 3 & 1 & 1 \end{vmatrix}}{29} = 1. \end{split}\]

The following proof of Cramer’s rule rests rather nicely on properties of the determinant function. But feel free to skip it.

Remark 5.4.1

Cramer’s formula seems the solution to all your linear systems. However, it has its drawbacks.

Disclaimer 1   Cramer’s formula can only be used for a square linear system with an invertible matrix.

Disclaimer 2   For a system with two equations in two unknowns Cramer’s rule may come in handy, but for solving larger systems it is highly inefficient. For instance, for a system of four equations in four unknowns, to find the solution using Cramer’s rule, one needs to compute five \(4 \times 4\) determinants. The good-old method using the augmented matrix \([\,A\,|\,\vect{b}\,]\) only asks for one row reduction process.

5.4.3. The inverse of a matrix in terms of determinants#

As an interesting corollary of Cramer’s Theorem we can give a ready-made formula for the inverse of an invertible matrix. The following proposition considers the notation of the previous section for a special case.

Proposition 5.4.4

Let \(A\) be an \(n\times n\) matrix, and \(\vect{e}_j\) the \(j\)th vector of the standard basis of \(\R^n\). Then

\[ \det{(A^{(i)}(\vect{e}_j))} = (-1)^{j+i} \det{A_{ji}} = C_{ji}, \]

where \(A_{ji}\) is the submatrix and \(C_{ji} = (-1)^{j+i} \det{\left(A_{ji}\right)} \) the cofactor as introduced in the definition of the \(n \times n\) determinant (Definition 5.2.2).

The following example serves as an illustration of what is going on here.

Example 5.4.4

Let \(A = \left[\begin{array}{rrrr} a_{11} &a_{12} &a_{13} &a_{14} \\ a_{21} &a_{22} &a_{23} &a_{24} \\ a_{31} &a_{32} &a_{33} &a_{34} \\ a_{41} &a_{42} &a_{43} &a_{44} \end{array} \right] \) be any \(4 \times 4\) matrix.

Then   \( A^{(4)}(\vect{e}_2) = \left[\begin{array}{rrrr} a_{11} &a_{12} &a_{13} &0 \\ a_{21} &a_{22} &a_{23} &1 \\ a_{31} &a_{32} &a_{33} &0 \\ a_{41} &a_{42} &a_{43} &0 \end{array} \right].\)

Expanding along the fourth column gives

\[\begin{split} \det{(A^{(4)}(\vect{e}_2))} = \left|\begin{array}{rrrr} a_{11} &a_{12} &a_{13} &0 \\ a_{21} &a_{22} &a_{23} &1 \\ a_{31} &a_{32} &a_{33} &0 \\ a_{41} &a_{42} &a_{43} &0 \end{array} \right|= (-1)^{(2+4)} \left|\begin{array}{rrr} a_{11} &a_{12} &a_{13} \\ a_{31} &a_{32} &a_{33} \\ a_{41} &a_{42} &a_{43} \end{array} \right|= C_{24}. \end{split}\]

Proposition 5.4.5

If \(A\) is an invertible \(n \times n\) matrix then the inverse \(B\) of \(A\) is given by

(5.4.3)#\[\begin{split}B = \dfrac{1}{\det{A}} \left[\begin{array}{ccccc} C_{11} &C_{21} &C_{31} & \ldots &C_{n1} \\ C_{12} &C_{22} &C_{32} & \ldots &C_{n2} \\ C_{13} &C_{23} &C_{33} & \ldots &C_{n3} \\ \vdots & \vdots &\vdots & \ddots & \vdots \\ C_{1n} &C_{2n} &C_{3n} & \ldots &C_{nn} \\ \end{array} \right].\end{split}\]

Proof of Proposition 5.4.5

The \(j\)th column \(\vect{b}_j\) of \(B = A^{-1}\) is the solution of the linear system \(A\vect{x} = \vect{e}_j\).

Cramer’s rule then gives that \(b_{ij}\), the \(i\)th entry of this column, is equal to

\[ b_{ij} = \dfrac{\det{\left(A^{(i)}(\vect{e}_j)\right)}}{\det{A}} = \dfrac{C_{ji}}{\det{A}}. \]

For the last step we used Proposition 5.4.4.

Definition 5.4.4

For an \(n \times n\) matrix \(A\) the matrix

\[\begin{split} \left[\begin{array}{ccccc} C_{11} &C_{12} &C_{13} & \ldots &C_{1n} \\ C_{21} &C_{22} &C_{23} & \ldots &C_{2n} \\ C_{31} &C_{32} &C_{33} & \ldots &C_{3n} \\ \vdots & \vdots &\vdots & \ddots & \vdots \\ C_{n1} &C_{n2} &C_{n3} & \ldots &C_{nn} \\ \end{array} \right] \end{split}\]

is called its cofactor matrix.

The adjugate matrix of \(A\) is defined as the transpose of the cofactor matrix. So

\[\begin{split} \text{Adj}(A) = \left[\begin{array}{ccccc} C_{11} &C_{21} &C_{31} & \ldots &C_{n1} \\ C_{12} &C_{22} &C_{32} & \ldots &C_{n2} \\ C_{13} &C_{23} &C_{33} & \ldots &C_{n3} \\ \vdots & \vdots &\vdots & \ddots & \vdots \\ C_{1n} &C_{2n} &C_{3n} & \ldots &C_{nn} \\ \end{array} \right] . \end{split}\]

Thus Proposition 5.4.5 states that

\[ A^{-1} = \dfrac{1}{\det{A}} \text{Adj}(A), \]

provided that \(A\) is invertible. In fact a slightly more general formula holds for any square matrix.

Proposition 5.4.6

For any square matrix \(A\) the following identity holds:

\[ A\cdot\text{Adj}(A) = \text{Adj}(A)\cdot A = (\det{A})\cdot I. \]

For clarity we used dots to indicate products. Note that the first two products are matrix products and the third product is a scalar times a matrix.

The proof we think, is short and instructive.

For \(n = 2\) Proposition 5.4.5 gives us back the old formula for the inverse. That is, if we define the determinant of a \(1 \times 1\) matrix \(A = [a]\) as the number \(a\) .

For an arbitrary invertible \(3 \times 3\) matrix \(A=\left[\begin{array}{ccc} a_{11} &a_{12} &a_{13} \\ a_{21} &a_{22} &a_{23} \\ a_{31} &a_{32} &a_{33} \end{array} \right] \) the formula yields

\[\begin{split} A^{-1} = \dfrac{1}{\begin{vmatrix} a_{11} &a_{12} &a_{13} \\ a_{21} &a_{22} &a_{23} \\ a_{31} &a_{32} &a_{33} \end{vmatrix}} \left[\begin{array}{ccc} \begin{vmatrix} a_{22} & a_{23} \\ a_{32} & a_{33} \end{vmatrix} & - \begin{vmatrix} a_{12} & a_{13} \\ a_{32} & a_{33} \end{vmatrix} & \begin{vmatrix} a_{12} & a_{13} \\ a_{22} & a_{23} \end{vmatrix} \\ - \begin{vmatrix} a_{21} & a_{23} \\ a_{31} & a_{33} \end{vmatrix} & \begin{vmatrix} a_{11} & a_{13} \\ a_{31} & a_{33} \end{vmatrix} & - \begin{vmatrix} a_{21} & a_{31} \\ a_{22} & a_{32} \end{vmatrix} \\ \begin{vmatrix} a_{21} & a_{22} \\ a_{31} & a_{32} \end{vmatrix} & - \begin{vmatrix} a_{11} & a_{31} \\ a_{12} & a_{32} \end{vmatrix} & \begin{vmatrix} a_{11} & a_{12} \\ a_{21} & a_{22} \end{vmatrix} \end{array} \right]. \end{split}\]

Remark 5.4.2

Like Cramer’s rule, the formula for the inverse is highly inefficient. The comparison between the efforts required to compute the inverse via the adjugate matrix versus row reduction of the augmented matrix \([\,A\,|\,I\,]\) works out rather favorably for the latter.

5.4.4. Determinant and cross product#

In Section 1.3 the cross product of two vectors \(\mathbf{u}\) and \(\mathbf{v}\) in \(\R^3\) is defined. It is the unique vector \(\mathbf{w}\) that is (1) orthogonal to \(\mathbf{u}\) and \(\mathbf{v}\), with (2) length equal to the area of the parallelogram with sides \(\mathbf{u}\) and \(\mathbf{v}\), and (3) such that the triple \((\mathbf{u},\mathbf{v},\mathbf{w})\) is ‘righthanded’ (= positively oriented).

In Section 5.1 we defined the determinant of the ordered set \((\vect{a},\vect{b},\vect{c})\) in \(\R^3\) via

\[\begin{split} \begin{array}{rcl} \det{(\vect{a},\vect{b},\vect{c})} &=& (\vect{a}\times\vect{b})\ip\vect{c} = \left|\begin{array}{ccc} a_1 & b_1 &c_1 \\ a_2 & b_2 &c_2 \\ a_3 & b_3 & c_3 \end{array}\right|\\ &=& \left|\begin{array}{cc} a_2 & a_3 \\b_2 & b_3 \end{array}\right| c_1 - \left|\begin{array}{cc} a_1 & a_3 \\b_1 & b_3 \end{array}\right| c_2 + \left|\begin{array}{cc} a_1 & a_2 \\b_1 & b_2 \end{array}\right| c_3. \end{array} \end{split}\]

Conversely, we can write the cross product in terms containing determinants.

(5.4.4)#\[\begin{split}\begin{array}{rcl} \left[\begin{array}{c} a_1 \\ a_2 \\ a_3 \end{array}\right] \times \left[\begin{array}{c}b_1 \\ b_2 \\ b_3 \end{array}\right] &=& \left[\begin{array}{c}a_2b_3-a_3b_2 \\ a_3b_1 - a_1b_3 \\ a_2b_1-a_2b_1 \end{array}\right] \\ &=& \left|\begin{array}{cc} a_2 & b_2 \\a_3 & b_3 \end{array}\right|\vect{e}_1 - \left|\begin{array}{cc} a_1 & b_1 \\ a_3 & b_3 \end{array}\right|\vect{e}_2 + \left|\begin{array}{cc} a_1 & b_1 \\a_2 & b_2 \end{array}\right|\vect{e}_3. \end{array}\end{split}\]

The last expression can formally be written as

\[\begin{split} \left|\begin{array}{ccc} a_1 & b_1 &\vect{e}_1 \\ a_2 & b_2 &\vect{e}_2 \\ a_3 & b_3 & \vect{e}_3 \end{array}\right|. \end{split}\]

In exactly the same fashion, we can, for \(n-1\) vectors \(\vect{a}_1, \ldots, \vect{a}_{n-1}\) in \(\R^n\), say

\[\begin{split} \vect{a}_1 = \left[\begin{array}{c} a_{11} \\ a_{21} \\ \vdots \\ a_{n1} \end{array}\right], \quad \vect{a}_2 = \left[\begin{array}{c} a_{12} \\ a_{22} \\ \vdots \\ a_{n2} \end{array}\right], \quad \ldots \quad , \quad \vect{a}_{n-1} = \left[\begin{array}{c} a_{1,(n-1)} \\ a_{2(n-1)} \\ \vdots \\ a_{n,(n-1)} \end{array}\right] \end{split}\]

define

(5.4.5)#\[\begin{split}\vect{a}^{\ast}_n = \vect{N}(\vect{a}_1, \ldots, \vect{a}_{n-1}) = \left|\begin{array}{ccccc} a_{11} & a_{12} & \ldots & a_{1,(n-1)} & \vect{e}_1 \\ a_{21} & a_{22} & \ldots & a_{2,(n-1)} & \vect{e}_2 \\ \vdots & \vdots & & \vdots & \vdots \\ a_{n1} & a_{n2} & \ldots & a_{n,(n-1)} & \vect{e}_n \end{array}\right|.\end{split}\]

Here \(\vect{e}_1, \ldots , \vect{e}_n\) denote the vectors of the standard basis for \(\R^n\).

With some effort it can be shown that the following properties hold.

Proposition 5.4.7

Suppose that \(\vect{a}_1, \ldots, \vect{a}_{n-1}\) are vectors in \(\R^n\) and \(\vect{a}^{\ast}_n\) is defined as in Equation (5.4.5). Then the following properties hold.

  1. \(\vect{a}^{\ast}_n \perp \vect{a}_i\), for \(i = 1,2,\ldots, n-1\) .

  2. \( \{\vect{a}_1, \, \ldots, \,\vect{a}_{n-1}\}\) is linearly dependent if and only if \(\det{\left[\,\vect{a}_1\,\, \ldots\,\, \vect{a}_{n-1}\,\, \vect{a}^{\ast}_n\,\right] } = 0\) .

  3. If \( \{\vect{a}_1, \ldots, \vect{a}_{n-1}\}\) is linearly independent, then \(\det{\left[\,\vect{a}_1, \ldots, \vect{a}_{n-1}, \vect{a}^{\ast}_n\,\right] } > 0\) .

  4. The norm of the vector \(\vect{a}^{\ast}_n\) is equal to the \((n-1)\)-dimensional volume of the \((n-1)\)-dimensional parallelepiped generated by \(\vect{a}_1, \ldots, \vect{a}_{n-1}\) .

For an independent set of vectors \(\{\vect{a}_1, \ldots, \vect{a}_{n-1}\}\) in \(\R^n\), the properties of Proposition 5.4.7 uniquely determine \(\vect{a}^{\ast}_n\) as the vector \(\vect{v}\) that is orthogonal to \( \vect{a}_1, \ldots, \vect{a}_{n-1}\), has a prescribed length, and makes the ordered set \((\vect{a}_1, \ldots, \vect{a}_{n-1}, \vect{v}) \) positively oriented.
For a linearly dependent set of vectors property iv. implies that \(\vect{a}^{\ast}_n = \vect{0}\).

Example 5.4.5

For \(n = 2\) we get, for an arbitrary vector \(\vect{v} = \left[\begin{array}{c} a \\ b \end{array}\right] \neq \left[\begin{array}{c} 0\\0 \end{array}\right] \):

\[\begin{split} \vect{w} = \vect{N}\left(\vect{v}\right) = \left|\begin{array}{cc} a & \vect{e}_1\\ b & \vect{e}_2 \end{array}\right|= a\vect{e}_1 - b\vect{e}_2 = \left[\begin{array}{c} -b \\ a \end{array}\right|. \end{split}\]

This is indeed a vector orthogonal to \(\vect{v}\) with the same ‘one-dimensional volume’, i.e., length, as the vector \(\vect{v}\).

Moreover, \(\left(\vect{v}, \vect{w}\right) = \left(\left[\begin{array}{c} a \\ b \end{array}\right] , \left[\begin{array}{c} -b \\ a \end{array}\right] \right) \) is positively oriented, as can be seen by making a sketch.

This shows that the construction also works in \(\R^2\).

Example 5.4.6

We will find the vector \(\vect{n} = N(\vect{a}_1, \vect{a}_2, \vect{a}_3)\) for the columns of the matrix

\[\begin{split} A = \left[\begin{array}{ccc} 1 & 1 & 3 \\ 1 & -1 & 1 \\ 1 & 1 & -3 \\ -1 & 1 & 1 \end{array}\right]. \end{split}\]

The first entry \(n_1\) is computed as

\[\begin{split} n_1 = (-1)^{1+4}\left|\begin{array}{ccc} 1 & -1 & 1 \\ 1 & 1 & -3 \\ -1 & 1 & 1 \end{array}\right| = -\left|\begin{array}{ccc} 1 & -1 & 1 \\ 0 & 2 & -4 \\ 0 & 0 & 2 \end{array}\right| = -4. \end{split}\]

All in all we find

\[\begin{split} \vect{n} = \left[\begin{array}{c} -4 \\ 12 \\ 4 \\ 12 \end{array}\right] = (-4)\left[\begin{array}{c} 1 \\ -3 \\ -1 \\ -3 \end{array}\right]. \end{split}\]

By taking inner products, or by computing \(A^T\vect{n}\), it is checked that indeed \(\vect{n} \perp \vect{a}_i\) for each column \(\vect{a}_i\). So property i. of Proposition 5.4.7 is satisfied.

Since the three columns are orthogonal, the ‘rectangular box’ in \(\R^4\) they generate will have 3d-volume

\[ \norm{\vect{a}_1} \cdot \norm{\vect{a}_2} \cdot \norm{\vect{a}_3} = \sqrt{4}\cdot \sqrt{4}\cdot \sqrt{20} = 8\sqrt{5}. \]

This is indeed equal to

\[ \norm{\vect{n}} = \sqrt{4^2+12^2+4^2+12^2} = \sqrt{320}, \]

so property iv. is satisfied too.

We end the chapter with a proof of Proposition 5.4.7.
So, if you are interested, push the button on the right.

5.4.5. Grasple Exercises#

Grasple Exercise 5.4.1

https://embed.grasple.com/exercises/ddb8daf3-3773-44c9-8df0-fe3084a6e7c4?id=93170

To compute the area of a triangle with sides \(\vect{u}\) and \(\vect{v}\) in the plane.

Grasple Exercise 5.4.2

https://embed.grasple.com/exercises/8d7a0672-6283-4bb1-9b43-b41a03067e40?id=93171

To find a point \(C\) on a line, such that  the area of a triangle \(ABC\) has a given value.

Grasple Exercise 5.4.3

https://embed.grasple.com/exercises/62770bc4-da31-4212-a713-bb2843b0e580?id=93172

Which points lie on the same side of a plane?

Grasple Exercise 5.4.4

https://embed.grasple.com/exercises/f787e084-9a77-40b4-b755-97890b98cfb6?id=93176

To solve a 3x3 system using Cramer’s rule.

Grasple Exercise 5.4.5

https://embed.grasple.com/exercises/3add427a-88a3-4da0-8f0a-2bf8bb8781dd?id=93179

Finding two entries in the inverse of a 4x4 matrix (using the adjoint matrix).

Grasple Exercise 5.4.6

https://embed.grasple.com/exercises/8d4a98f7-50ac-4705-8b34-680b7b8395d9?id=93181

To find a vector orthogonal to \(\vect{v}_1,\vect{v}_2,\vect{v}_3\) in \(\mathbb{R}^4\), with good orientation.

Grasple Exercise 5.4.7

https://embed.grasple.com/exercises/bc3df113-95b3-470a-a730-3ad8faab08f5?id=93183

To compute the normal vector \(N(\vect{a}_1,\vect{a}_2,\vect{a}_3)\) as in Subsection 5.4.4