The determinant
2a. Matrix inversion
We have seen in the previous chapter that most of the interesting maps [math]f:\mathbb R^N\to\mathbb R^N[/math] that we know, such as the rotations, symmetries and projections, are linear, and can be written in the following form, with [math]A\in M_N(\mathbb R)[/math] being a square matrix:
In this chapter we develop more general theory for such linear maps. We are mostly motivated by the following fundamental result, which has countless concrete applications, and which is actually at the origin of the whole linear algebra theory:
Any linear system of equations
With linear algebra conventions, our system reads:
Thus, we are led to the conclusions in the statement.
In practice, we are led to the question of inverting the matrices [math]A\in M_N(\mathbb R)[/math]. And this is the same question as inverting the linear maps [math]f:\mathbb R^N\to\mathbb R^N[/math], due to:
A linear map [math]f:\mathbb R^N\to\mathbb R^N[/math], written as
This is something that we basically know, coming from the fact that, with the notation [math]f_A(v)=Av[/math], we have the following formula:
Thus, we are led to the conclusion in the statement.
In order to study invertibility questions, for matrices or linear maps, let us begin with some examples. In the simplest case, in 2 dimensions, the result is as follows:
We have the following inversion formula, for the [math]2\times2[/math] matrices:
We have two assertions to be proved, the idea being as follows:
(1) As a first observation, when [math]ad-bc=0[/math] we must have, for some [math]\lambda\in\mathbb R[/math]:
Thus our matrix must be of the following special type:
But in this case the columns are proportional, and so the linear map associated to the matrix is not invertible, and so the matrix itself is not invertible either.
(2) When [math]ad-bc\neq 0[/math], let us look for an inversion formula of the following type:
We must therefore solve the following equations:
The obvious solution here is as follows:
Thus, we are led to the formula in the statement.
In order to deal now with the inversion problem in general, for the arbitrary matrices [math]A\in M_N(\mathbb R)[/math], we will use the same method as the one above, at [math]N=2[/math]. Let us write indeed our matrix as follows, with [math]v_1,\ldots,v_N\in\mathbb R^N[/math] being its column vectors:
We know from the general results from chapter 1 that, in order for [math]A[/math] to be invertible, the vectors [math]v_1,\ldots,v_N[/math] must be linearly independent. Thus, following the observations (1) from the above proof of Theorem 2.3, we are led into the question of understanding when a family of vectors [math]v_1,\ldots,v_N\in\mathbb R^N[/math] are linearly independent.
In order to deal with this latter question, let us introduce the following notion:
Associated to any vectors [math]v_1,\ldots,v_N\in\mathbb R^N[/math] is the volume
Here the volume is taken in the standard [math]N[/math]-dimensional sense. At [math]N=1[/math] this volume is a length, at [math]N=2[/math] this volume is an area, at [math]N=3[/math] this is the usual 3D volume, and so on. In general, the volume of a body [math]X\subset\mathbb R^N[/math] is by definition the number [math]vol(X)\in[0,\infty][/math] of copies of the unit cube [math]C\subset\mathbb R^N[/math] which are needed for filling [math]X[/math], when allowing this unit cube to be divided into smaller cubes, for the needs of the filling operation.
In order to compute this volume we can use various geometric techniques, and we will see soon that, in what regards the case that we are interested in, namely that of the parallelepipeds [math]P\subset\mathbb R^N[/math], we can basically compute here everything, just by using very basic geometric techniques, essentially based on the Thales theorem.
In relation with our inversion problem, we have the following statement:
The quantity [math]{\rm det}^+[/math] that we constructed, regarded as a function of the corresponding square matrices, formed by column vectors,
This follows from Theorem 2.2, and from the general results from chapter 1, which tell us that a matrix [math]A\in M_N(\mathbb R)[/math] is invertible precisely when its column vectors [math]v_1,\ldots,v_N\in\mathbb R^N[/math] are linearly independent. But this latter condition is equivalent to the fact that we must have the following strict inequality:
Thus, we are led to the conclusion in the statement.
Summarizing, all this leads us into the explicit computation of [math]{\rm det}^+[/math]. As a first observation, in 1 dimension we obtain the absolute value of the real numbers:
In 2 dimensions now, the computation is non-trivial, and we have the following result, making the link with our main result so far, namely Theorem 2.3:
In [math]2[/math] dimensions we have the following formula,
We must show that the area of the parallelogram formed by [math]\binom{a}{c},\binom{b}{d}[/math] equals [math]|ad-bc|[/math]. We can assume [math]a,b,c,d \gt 0[/math] for simplifying, the proof in general being similar. Moreover, by switching if needed the vectors [math]\binom{a}{c},\binom{b}{d}[/math], we can assume that we have:
According to these conventions, the picture of our parallelogram is as follows:
Now let us slide the upper side downwards left, until we reach the [math]Oy[/math] axis. Our parallelogram, which has not changed its area in this process, becomes:
We can further modify this parallelogram, once again by not altering its area, by sliding the right side downwards, until we reach the [math]Ox[/math] axis:
Let us compute now the area. Since our two sliding operations have not changed the area of the original parallelogram, this area is given by:
In order to compute the quantity [math]x[/math], observe that in the context of the first move, we have two similar triangles, according to the following picture:
Thus, we are led to the following equation for the number [math]x[/math]:
By solving this equation, we obtain the following value for [math]x[/math]:
Thus the area of our parallelogram, or rather of the final rectangle obtained from it, which has the same area as the original parallelogram, is given by:
Thus, we are led to the conclusion in the statement.
2b. The determinant
All the above is very nice, and we obviously have a beginning of theory here. However, when looking carefully, we can see that our theory has a weakness, because:
- In 1 dimension the number [math]a[/math], which is the simplest function of [math]a[/math] itself, is certainly a better quantity than the number [math]|a|[/math].
- In 2 dimensions the number [math]ad-bc[/math], which is linear in [math]a,b,c,d[/math], is certainly a better quantity than the number [math]|ad-bc|[/math].
So, let us upgrade now our theory, by constructing a better function, which does the same job, namely checking if the vectors are proportional, of the following type:
That is, we would like to have a clever, signed version of [math]\det^+[/math], satisfying:
In order to do this, we must come up with a way of splitting the systems of vectors [math]v_1,\ldots,v_N\in\mathbb R^N[/math] into two classes, call them positive and negative. And here, the answer is quite clear, because a bit of thinking leads to the following definition:
A system of vectors [math]v_1,\ldots,v_N\in\mathbb R^N[/math] is called:
- Oriented, if one can continuously pass from the standard basis to it.
- Unoriented, otherwise.
The associated sign is [math]+[/math] in the oriented case, and [math]-[/math] in the unoriented case.
As a first example, in 1 dimension the basis consists of the single vector [math]e=1[/math], which can be continuously deformed into any vector [math]a \gt 0[/math]. Thus, the sign is the usual one:
Thus, in connection with our original question, we are definitely on the good track, because when multiplying [math]|a|[/math] by this sign we obtain [math]a[/math] itself, as desired:
In 2 dimensions now, the explicit formula of the sign is as follows:
We have the following formula, valid for any [math]2[/math] vectors in [math]\mathbb R^2[/math],
According to our conventions, the sign of [math]\binom{a}{c},\binom{b}{d}[/math] is as follows:
(1) The sign is [math]+[/math] when these vectors come in this order with respect to the counterclockwise rotation in the plane, around 0.
(2) The sign is [math]-[/math] otherwise, meaning when these vectors come in this order with respect to the clockwise rotation in the plane, around 0.
If we assume now [math]a,b,c,d \gt 0[/math] for simplifying, we are left with comparing the angles having the numbers [math]c/a[/math] and [math]d/b[/math] as tangents, and we obtain in this way:
But this gives the formula in the statement. The proof in general is similar.
Once again, in connection with our original question, we are on the good track, because when multiplying [math]|ad-bc|[/math] by this sign we obtain [math]ad-bc[/math] itself, as desired:
Let us look as well into the case [math]N=3[/math]. Things here are quite complicated, and we will discuss this later on. However, we have the following basic result:
Consider the standard basis of [math]\mathbb R^3[/math], namely:
- [math]sgn(e_1,e_2,e_3)=+[/math].
- [math]sgn(e_1,e_3,e_2)=-[/math].
- [math]sgn(e_2,e_1,e_3)=-[/math].
- [math]sgn(e_2,e_3,e_1)=+[/math].
- [math]sgn(e_3,e_1,e_2)=+[/math].
- [math]sgn(e_3,e_2,e_1)=-[/math].
In each case the problem is whether one can continuously pass from [math](e_1,e_2,e_3)[/math] to the basis in statement, and the computations can be done as follows:
(1) In three of the cases under investigation, namely (2,3,6), one of the vectors is unchanged, and the other two are switched. Thus, we are more or less in 2 dimensions, and since the switch here clearly corresponds to [math]-[/math], the sign in these cases is [math]-[/math].
(2) As for the remaining three cases, namely (1,4,5), here the sign can only be [math]+[/math], since things must be 50-50 between [math]+[/math] and [math]-[/math], say by symmetry reasons. And this is indeed the case, because what we have here are rotations of the standard basis.
As already mentioned, we will be back to this later, with a general formula for the sign in 3 dimensions. This formula is quite complicated, the idea being that of making out of the [math]3\times3=9[/math] entries of our vectors a certain quantity, somewhat in the spirit of the one in Proposition 2.8, and then taking the sign of this quantity.
At the level of the general results now, we have:
The orientation of a system of vectors changes as follows:
- If we switch the sign of a vector, the associated sign switches.
- If we permute two vectors, the associated sign switches as well.
Both these assertions are clear from the definition of the sign, because the two operations in question change the orientation of the system of vectors.
With the above notion in hand, we can now formulate:
The determinant of [math]v_1,\ldots,v_N\in\mathbb R^N[/math] is the signed volume
In other words, we are upgrading here Definition 2.4, by adding a sign to the quantity [math]{\rm det}^+[/math] constructed there, as to potentially reach to good additivity properties:
In relation with our original inversion problem for the square matrices, this upgrade does not change what we have so far, and we have the following statement:
The quantity [math]\det[/math] that we constructed, regarded as a function of the corresponding square matrices, formed by column vectors,
We know from Theorem 2.5 that a matrix [math]A\in M_N(\mathbb R)[/math] is invertible precisely when [math]{\rm det}^+(A)=|\det A|[/math] is strictly positive, and this gives the result.
In the matrix context, we will often use the symbol [math]|\,.\,|[/math] instead of [math]\det[/math]:
Let us try now to compute the determinant. In 1 dimension we have of course the formula [math]\det(a)=a[/math], because the absolute value fits, and so does the sign:
In 2 dimensions now, we have the following result:
In [math]2[/math] dimensions we have the following formula,
According to our definition, to the computation in Theorem 2.6, and to sign formula from Proposition 2.8, the determinant of a [math]2\times2[/math] matrix is given by:
Thus, we have obtained the formula in the statement.
2c. Basic properties
In order to discuss now arbitrary dimensions, we will need a number of theoretical results. Here is a first series of formulae, coming straight from the definitions:
The determinant has the following properties:
- When multiplying by scalars, the determinant gets multiplied as well:
[[math]] \det(\lambda_1v_1,\ldots,\lambda_Nv_N)=\lambda_1\ldots\lambda_N\det(v_1,\ldots,v_N) [[/math]]
- When permuting two columns, the determinant changes the sign:
[[math]] \det(\ldots,u,\ldots,v,\ldots)=-\det(\ldots,v,\ldots,u,\ldots) [[/math]]
- The determinant [math]\det(e_1,\ldots,e_N)[/math] of the standard basis of [math]\mathbb R^N[/math] is [math]1[/math].
All this is clear from definitions, as follows:
(1) This follows from definitions, and from Proposition 2.10 (1).
(2) This follows as well from definitions, and from Proposition 2.10 (2).
(3) This is clear from our definition of the determinant.
As an application of the above result, we have:
The determinant of a diagonal matrix is given by:
The formula in the statement is clear by using the rules (1) and (3) in Theorem 2.14 above, which in matrix terms give:
As for the last assertion, this is rather a remark.
The above result is very useful, and we will see in a moment that, more generally, the determinant of any diagonalizable matrix is the product of its eigenvalues.
In order to reach now to a more advanced theory, let us adopt the linear map point of view. In this setting, the definition of the determinant reformulates as follows:
Given a linear map, written as [math]f(v)=Av[/math], its “inflation coefficient”, obtained as the signed volume of the image of the unit cube, is given by:
The only non-trivial thing in all this is the fact that the inflation coefficient [math]I_f[/math], as defined above, is independent of the choice of the parallelepiped. But this is a generalization of the Thales theorem, which follows from the Thales theorem itself.
As a first application of the above linear map viewpoint, we have:
We have the following formula, valid for any matrices [math]A,B[/math]:
The decomposition formula in the statement follows by using the associated linear maps, which multiply as follows:
Indeed, when computing the determinant, by using the “inflation coefficient” viewpoint from Theorem 2.16, we obtain the same thing on both sides. As for the formula [math]\det(AB)=\det(BA)[/math], this is clear from the first formula, which is symmetric in [math]A,B[/math].
Getting back now to explicit computations, we have the following key result:
The determinant of a diagonalizable matrix
We know that a diagonalizable matrix can be written in the form [math]A=PDP^{-1}[/math], with [math]D=diag(\lambda_1,\ldots,\lambda_N)[/math]. Now by using Theorem 2.17, we obtain:
Thus, we are led to the formula in the statement.
Here is another important result, which is very useful for diagonalization:
The eigenvalues of a matrix [math]A\in M_N(\mathbb R)[/math] are the roots of
We have the following computation, using the fact that a linear map is bijective precisely when the determinant of the associated matrix is nonzero:
Thus, we are led to the conclusion in the statement.
Here are now some other computations, once again in arbitrary dimensions:
We have the following results:
- The determinant of an orthogonal matrix must be [math]\pm1[/math].
- The determinant of a projection must be [math]0[/math] or [math]1[/math].
These are elementary results, the idea being as follows:
(1) Here the determinant must be indeed [math]\pm1[/math], because the orthogonal matrices map the unit cube to a copy of the unit cube.
(2) Here the determinant is in general 0, because the projections flatten the unit cube, unless we have the identity, where the determinant is 1.
In general now, at the theoretical level, we have the following key result:
The determinant has the additivity property
This follows by doing some elementary geometry, in the spirit of the computations in the proof of Theorem 2.6, as follows:
(1) We can either use the Thales theorem, and then compute the volumes of all the parallelepipeds involved, by using basic algebraic formulae.
(2) Or we can solve the problem in “puzzle” style, the idea being to cut the big parallelepiped, and then recover the small ones, after some manipulations.
(3) We can do as well something hybrid, consisting in deforming the parallelepipeds involved, without changing their volumes, and then cutting and gluing.
As a basic application of the above result, we have:
We have the following results:
- The determinant of a diagonal matrix is the product of diagonal entries.
- The same is true for the upper triangular matrices.
- The same is true for the lower triangular matrices.
All this can be deduced by using our various general formulae, as follows:
(1) This is something that we already know, from Theorem 2.15.
(2) This follows by using Theorem 2.14 and Theorem 2.21, then (1), as follows:
(3) This follows as well from Theorem 2.14 and Theorem 2.21, then (1), by proceeding this time from right to left, from the last column towards the first column.
We can see from the above that the rules in Theorem 2.14 and Theorem 2.21 are quite powerful, taken altogether. For future reference, let us record these rules:
The determinant has the following properties:
- When adding two columns, the determinants get added:
[[math]] \det(\ldots,u+v,\ldots) =\det(\ldots,u,\ldots) +\det(\ldots,v,\ldots) [[/math]]
- When multiplying columns by scalars, the determinant gets multiplied:
[[math]] \det(\lambda_1v_1,\ldots,\lambda_Nv_N)=\lambda_1\ldots\lambda_N\det(v_1,\ldots,v_N) [[/math]]
- When permuting two columns, the determinant changes the sign:
[[math]] \det(\ldots,u,\ldots,v,\ldots)=-\det(\ldots,v,\ldots,u,\ldots) [[/math]]
- The determinant [math]\det(e_1,\ldots,e_N)[/math] of the standard basis of [math]\mathbb R^N[/math] is [math]1[/math].
This is something that we already know, which follows by putting together the various formulae from Theorem 2.14 and Theorem 2.21.
As an important theoretical result now, which will ultimately lead to an algebraic reformulation of the whole determinant problematics, we have:
The determinant of square matrices is the unique map
This can be done in two steps, as follows:
(1) Our first claim is that any map [math]\det':M_N(\mathbb R)\to\mathbb R[/math] satisfying the conditions in Theorem 2.23 must coincide with [math]\det[/math] on the upper triangular matrices. But this is clear from the proof of Theorem 2.22, which only uses the rules in Theorem 2.23.
(2) Our second claim is that we have [math]\det'=\det[/math], on all matrices. But this can be proved by putting the matrix in upper triangular form, by using operations on the columns, in the spirit of the manipulations from the proof of Theorem 2.22.
Here is now another important theoretical result:
The determinant is subject to the row expansion formula
This follows from the fact that the formula in the statement produces a certain function [math]\det:M_N(\mathbb R)\to\mathbb R[/math], which has the 4 properties in Theorem 2.23.
We can expand as well over the columns, as follows:
The determinant is subject to the column expansion formula
This follows by using the same argument as for the rows.
We can now complement Theorem 2.23 with a similar result for the rows:
The determinant has the following properties:
- When adding two rows, the determinants get added:
[[math]] \det\begin{pmatrix}\vdots\\ u+v\\ \vdots\end{pmatrix} =\det\begin{pmatrix}\vdots\\ u\\ \vdots\end{pmatrix} +\det\begin{pmatrix}\vdots \\ v\\ \vdots\end{pmatrix} [[/math]]
- When multiplying row by scalars, the determinant gets multiplied:
[[math]] \det\begin{pmatrix}\lambda_1v_1\\ \vdots\\ \lambda_Nv_N\end{pmatrix} =\lambda_1\ldots\lambda_N\det\begin{pmatrix}v_1\\ \vdots\\ v_N\end{pmatrix} [[/math]]
- When permuting two rows, the determinant changes the sign.
This follows indeed by using the using various formulae established above, and is best seen by using the column expansion formula from Theorem 2.26.
We can see from the above that the determinant is the subject to many interesting formulae, and that some of these formulae, when taken altogether, uniquely determine it. In all this, what is the most luminous is certainly the definition of the determinant as a volume. As for the second most luminous of our statements, this is Theorem 2.24, which is something a bit abstract, but both beautiful and useful. So, as a final theoretical statement now, here is an alternative reformulation of Theorem 2.24:
The determinant of the systems of vectors
This is a fancy reformulation of Theorem 2.24, with the various properties of [math]\det[/math] from the statement being those from Theorem 2.23.
As a conclusion to all this, we have now a full theory for the determinant, and we can freely use all the above results, definitions and theorems alike, and even start forgetting what is actually definition, and what is theorem.
2d. Sarrus and beyond
As a first application of the above methods, we can now prove:
The determinant of the [math]3\times3[/math] matrices is given by
Here is the computation, using Theorem 2.25:
Thus, we obtain the formula in the statement.
As a first application, let us go back to the inversion problem for the [math]3\times3[/math] matrices, that we left open in the above. We can now solve this problem, as follows:
The inverses of the [math]3\times3[/math] matrices are given by
We can use here the same method as for the [math]2\times 2[/math] matrices. To be more precise, in order for the matrix to be invertible, we must have:
The trick now is to look for solutions of the following problem:
We know from Theorem 2.29 that the determinant is given by:
But this leads, via some obvious choices, to the following solution:
Thus, by rescaling, we obtain the formula in the statement.
In fact, we can now fully solve the inversion problem, as follows:
The inverse of a square matrix, having nonzero determinant,
This follows indeed by using the row expansion formula from Theorem 2.25, which in terms of the matrix [math]A^{-1}[/math] in the statement reads [math]AA^{-1}=1[/math].
In practice, the above result leads to the following algorithm, which is quite easy to memorize, for computing the inverse:
(1) Delete rows and columns, and compute the corresponding determinants.
(2) Transpose, and add checkered signs.
(3) Divide by the determinant.
Observe that this generalizes our previous computations at [math]N=2,3[/math]. As an illustration, consider an arbitrary [math]2\times2[/math] matrix, written as follows:
By deleting rows and columns we obtain [math]1\times1[/math] matrices, and so the matrix formed by the determinants [math]\det(A^{(ij)})[/math] is as follows:
Now by transposing, adding checkered signs and dividing by [math]\det A[/math], we obtain:
Similarly, at [math]N=3[/math] what we obtain is the Sarrus formula, from Theorem 2.29.
As a new application now, let us record the following result, at [math]N=4[/math]:
The determinant of the [math]4\times4[/math] matrices is given by
The formula for the determinant follows by developing over the first row, then by using the Sarrus formula, for each of the 4 smaller determinants which appear:
As for the formula of the inverse, this is something that we already know.
Let us discuss now the general formula of the determinant, at arbitrary values [math]N\in\mathbb N[/math] of the matrix size, generalizing those that we have at [math]N=2,3,4[/math]. We will need:
A permutation of [math]\{1,\ldots,N\}[/math] is a bijection, as follows:
There are many possible notations for the permutations, the basic one consisting in writing the numbers [math]1,\ldots,N[/math], and below them, their permuted versions:
Another method, which is faster, is by denoting the permutations as diagrams, acting from top to bottom:
Here are some basic properties of the permutations:
The permutations have the following properties:
- There are [math]N![/math] of them.
- They are stable by composition, and inversion.
In order to construct a permutation [math]\sigma\in S_N[/math], we have:
-- [math]N[/math] choices for the value of [math]\sigma(N)[/math]. -- [math](N-1)[/math] choices for the value of [math]\sigma(N-1)[/math]. -- [math](N-2)[/math] choices for the value of [math]\sigma(N-2)[/math]. [math]\vdots[/math] -- and so on, up to 1 choice for the value of [math]\sigma(1)[/math].
Thus, we have [math]N![/math] choices, as claimed. As for the second assertion, this is clear.
We will need the following key result:
The permutations have a signature function
- As [math](-1)^c[/math], where [math]c[/math] is the number of inversions.
- As [math](-1)^t[/math], where [math]t[/math] is the number of transpositions.
- As [math](-1)^o[/math], where [math]o[/math] is the number of odd cycles.
- As [math](-1)^x[/math], where [math]x[/math] is the number of crossings.
- As the sign of the corresponding permuted basis of [math]\mathbb R^N[/math].
{{{4}}}
We can now formulate a key result, as follows:
We have the following formula for the determinant,
This follows by recurrence over [math]N\in\mathbb N[/math], as follows:
(1) When developing the determinant over the first column, we obtain a signed sum of [math]N[/math] determinants of size [math](N-1)\times(N-1)[/math]. But each of these determinants can be computed by developing over the first column too, and so on, and we are led to the conclusion that we have a formula as in the statement, with [math]\varepsilon(\sigma)\in\{-1,1\}[/math] being certain coefficients.
(2) But these latter coefficients [math]\varepsilon(\sigma)\in\{-1,1\}[/math] can only be the signatures of the corresponding permutations [math]\sigma\in S_N[/math], with this being something that can be viewed again by recurrence, with either of the definitions (1-5) in Theorem 2.35 for the signature.
The above result is something quite tricky, and in order to get familiar with it, there is nothing better than doing some computations. As a first, basic example, in 2 dimensions we recover the usual formula of the determinant, the details being as follows:
In 3 dimensions now, we recover the Sarrus formula:
Observe that the triangles in the Sarrus formula correspond to the permutations of [math]\{1,2,3\}[/math], and their signs correspond to the signatures of these permutations:
Also, in 4 dimensions, we recover the formula that we already know, as follows:
The determinant of the [math]4\times4[/math] matrices is given by
We can indeed recover this formula as well as a particular case of Theorem 2.36. To be more precise, the permutations in the statement are listed according to the lexicographic order, and the computation of the corresponding signatures is something elementary, by using the various rules from Theorem 2.35.
As another application, we have the following key result:
We have the formula
This follows from the formula in Theorem 2.36. Indeed, we have:
Thus, we are led to the formula in the statement.
Good news, this is the end of the general theory that we wanted to develop. We have now in our bag all the needed techniques for computing the determinant.
Here is however a nice and important example of a determinant, whose computation uses some interesting new techniques, going beyond what has been said above:
We have the Vandermonde determinant formula
By expanding over the columns, we see that the determinant in question, say [math]D[/math], is a polynomial in the variables [math]x_1,\ldots,x_N[/math], having degree [math]N-1[/math] in each variable. Now observe that when setting [math]x_i=x_j[/math], for some indices [math]i\neq j[/math], our matrix will have two identical columns, and so its determinant [math]D[/math] will vanish:
But this gives us the key to the computation of [math]D[/math]. Indeed, [math]D[/math] must be divisible by [math]x_i-x_j[/math] for any [math]i\neq j[/math], and so we must have a formula of the following type:
Moreover, since the product on the right is, exactly as [math]D[/math] itself, a polynomial in the variables [math]x_1,\ldots,x_N[/math], having degree [math]N-1[/math] in each variable, we conclude that the quantity [math]c[/math] must be a constant, not depending on any of the variables [math]x_1,\ldots,x_N[/math]:
In order to finish the computation, it remains to find the value of this constant [math]c[/math]. But this can be done for instance by recurrence, and we obtain:
Thus, we are led to the formula in the statement.
Getting back now to generalities, and to what we want to do with our linear algebra theory, now that we are experts in the computation of the determinant, we should investigate the next problem, namely the diagonalization one.
And here, we know from Theorem 2.19 that the eigenvalues of a matrix [math]A\in M_N(\mathbb R)[/math] appear as roots of the characteristic polynomial:
Thus, with the determinant theory developed above, we can in principle compute these eigenvalues, and solve the diagonalization problem afterwards.
The problem, however, is that certain real matrices can have characteristic polynomials of type [math]P(x)=x^2+1[/math], and this suggests that these matrices might be not diagonalizable over [math]\mathbb R[/math], but be diagonalizable over [math]\mathbb C[/math] instead. And so, before getting into diagonalization problems, we must upgrade our theory, and talk about complex matrices. We will do this in the next chapter, and afterwards, we will go back to the diagonalization problem.
General references
Banica, Teo (2024). "Linear algebra and group theory". arXiv:2206.09283 [math.CO].