General theory
2a. Linear algebra
You probably know from linear algebra that the important question regarding any matrix [math]d\in M_N(\mathbb R)[/math] is its diagonalization. To be more precise, the first question is that of computing the eigenvalues, which are usually complex numbers, [math]\lambda\in\mathbb C[/math]:
Then comes the computation of the eigenvectors, which are usually complex vectors too, [math]v\in\mathbb C^N[/math], and then the diagonalization problem, amounting in finding a new basis of [math]\mathbb C^N[/math], formed by eigenvectors. This basis can exist or not, and in case it exists, we reach to a formula as follows, with [math]\Lambda=diag(\lambda)[/math], and [math]P[/math] being the basis change matrix:
In the case of the graphs, or rather of the adjacency matrices [math]d\in M_N(0,1)[/math] of the graphs, these notions are quite important and intuitive, as shown by:
The eigenvectors of [math]d\in M_N(0,1)[/math], with eigenvalue [math]\lambda[/math], can be identified with the functions [math]f[/math] satisfying the following condition:
That is, the value of [math]f[/math] at each vertex must be the rescaled average, over the neighbors.
We have indeed the following computation, valid for any vector [math]f[/math]:
Thus, we are led to the conclusion in the statement.
The above result is quite interesting, and as an illustration, when assuming that our graph is [math]k[/math]-regular, for the particular value [math]\lambda=k[/math], the eigenvalue condition reads:
Thus, we can see here a relation with harmonic functions. There are many things that can be said here, and we will be back to this later, when talking Laplace operators.
But let us pause now our study of graphs, and go back to linear algebra. Taking the scalars to be complex numbers, which is something very standard, we have the following general result, that you surely know about, from mathematics or physics classes:
Given a matrix [math]d\in M_N(\mathbb C)[/math], consider its characteristic polynomial
then factorize this polynomial, by computing its complex roots, with multiplicities,
There are many things going on here, the idea being as follows:
(1) Given [math]d\in M_N(\mathbb C)[/math], for any eigenvalue [math]\lambda\in\mathbb C[/math] we can define the corresponding eigenspace as being the vector space formed by the corresponding eigenvectors:
These spaces [math]E_\lambda[/math] are easily seen to be in direct sum position, in the sense that given vectors [math]v_1\in E_{\lambda_1},\ldots,v_k\in E_{\lambda_k}[/math] coming from different eigenvalues [math]\lambda_1,\ldots,\lambda_k[/math], we have:
In particular we have the following estimate, with sum over all the eigenvalues, and our matrix is diagonalizable precisely when we have equality:
(2) Next, consider the characteristic polynomial of our matrix, given by:
Our claim is that the eigenvalues of [math]d[/math] are then the roots of [math]P[/math]. Indeed, this follows from the following computation, using the elementary fact that a linear map is bijective precisely when the determinant of the associated matrix is nonzero:
(3) Our next claim, which will lead to the result in the statement, is that we have an inequality as follows, where [math]m_\lambda[/math] is the multiplicity of [math]\lambda[/math], viewed as root of [math]P[/math]:
Indeed, for any eigenvalue [math]\lambda[/math], consider the dimension [math]n_\lambda=\dim(E_\lambda)[/math] of the corresponding eigenspace. By changing the basis of [math]\mathbb C^N[/math], as for [math]E_\lambda[/math] to be spanned by the first [math]n_\lambda[/math] basis elements, our matrix becomes as follows, with [math]e[/math] being a certain smaller matrix:
We conclude that the characteristic polynomial of [math]d[/math] is of the following form:
Thus the multiplicity of [math]\lambda[/math], as root of [math]P[/math], satisfies [math]m_\lambda\geq n_\lambda[/math], which proves our claim.
(4) Getting now to what we wanted to prove, as a first observation, what is said in the theorem is well formulated, thanks to what we found above. Next, by summing the inequalities [math]\dim(E_\lambda)\leq m_\lambda[/math] found in (3), we obtain an inequality as follows:
On the other hand, we know from (1) that our matrix is diagonalizable precisely when we have global equality. Thus, we are led to the conclusion in the statement.
In practice, diagonalizing a matrix remains something quite complicated. Let us record as well a useful, algorithmic version of the above result, as follows:
The square matrices [math]d\in M_N(\mathbb C)[/math] can be diagonalized as follows:
- Compute the characteristic polynomial.
- Factorize the characteristic polynomial.
- Compute the eigenvectors, for each eigenvalue found.
- If there are no [math]N[/math] eigenvectors, [math]d[/math] is not diagonalizable.
- Otherwise, [math]d[/math] is diagonalizable, [math]d=P\Lambda P^{-1}[/math].
This is an informal reformulation of Theorem 2.2, with (4) referring to the total number of linearly independent eigenvectors found in (3), and with [math]d=P\Lambda P^{-1}[/math] in (5) being the usual diagonalization formula, with [math]P,\Lambda[/math] being as before.
Getting now towards our graph problematics, here is a key result:
Any matrix [math]d\in M_N(\mathbb C)[/math] which is self-adjoint, [math]d=d^*[/math], is diagonalizable, with the diagonalization being of the following type,
As a first remark, the converse trivially holds, because if we take a matrix of the form [math]d=U\Lambda U^*[/math], with [math]U[/math] unitary and [math]\Lambda[/math] diagonal and real, then we have:
In the other sense now, assume that [math]d[/math] is self-adjoint, [math]d=d^*[/math]. Our first claim is that the eigenvalues are real. Indeed, assuming [math]dv=\lambda v[/math], we have:
Thus we obtain [math]\lambda\in\mathbb R[/math], as claimed. Our next claim now is that the eigenspaces corresponding to different eigenvalues are pairwise orthogonal. Assume indeed that:
We have then the following computation, using [math]\lambda,\mu\in\mathbb R[/math]:
Thus [math]\lambda\neq\mu[/math] implies [math]v\perp w[/math], as claimed. In order now to finish, it remains to prove that the eigenspaces span [math]\mathbb C^N[/math]. For this purpose, we will use a recurrence method. Let us pick an eigenvector, [math]dv=\lambda v[/math]. Assuming [math]v\perp w[/math], we have:
Thus, if [math]v[/math] is an eigenvector, then the vector space [math]v^\perp[/math] is invariant under [math]d[/math]. In order to do the recurrence, it still remains to prove that the restriction of [math]d[/math] to the vector space [math]v^\perp[/math] is self-adjoint. But this comes from a general property of the self-adjoint matrices, that we will explain now. Our claim is that for any matrix [math]d\in M_N(\mathbb C)[/math], we have:
Indeed, this follows from the following computation:
But this shows that the restriction of [math]d[/math] to any invariant subspace, and in particular to [math]v^\perp[/math], is self-adjoint. Thus, we can proceed by recurrence, and we obtain the result.
In what concerns us, in relation with our graph problems, we will rather need the real version of the above result, which is also something well-known, as follows:
Any matrix [math]d\in M_N(\mathbb R)[/math] which is symmetric, [math]d=d^t[/math], is diagonalizable, with the diagonalization being of the following type,
As before, the converse trivially holds, because if we take a matrix of the form [math]d=U\Lambda U^t[/math], with [math]U[/math] orthogonal, and [math]\Lambda[/math] diagonal and real, then we have:
In the other sense now, this follows from Theorem 2.4, and its proof. Indeed, we know from there that the eigenvalues are real, and in what concerns the passage matrix, the arguments there carry over to the real case, and show that this matrix is real too.
With the above results in hand, time now to get back to graphs. We have here the following particular case of Theorem 2.5, with the important drawback however that in what concerns the “converse holds too” part, that is unfortunately gone:
The adjacency matrix [math]d\in M_N(0,1)[/math] of any graph is diagonalizable, with the diagonalization being of the following type,
Here the first assertion follows from Theorem 2.5, because [math]d[/math] is by definition real and symmetric. As for the last assertion, this deserves some explanations:
(1) Generally speaking, in analogy with the last assertions in Theorem 2.4 and Theorem 2.5, which are something extremely useful, we would like to know under which assumptions on a rotation [math]U\in O_N[/math], and on a diagonal matrix [math]\Lambda\in M_N(\mathbb R)[/math], the real symmetric matrix [math]d=U\Lambda U^t[/math] has 0-1 entries, and 0 on the diagonal.
(2) Unfortunately, both these questions are obviously difficult, there is no simple answer to them, and things are like that. So, gone the possibility of a converse. However, as a small consolation, we can make the remark that, with [math]d=U\Lambda U^t[/math], we have:
Thus we have at least [math]Tr(\Lambda)=0[/math], as a necessary condition on [math](U,\Lambda)[/math], as stated.
In view of the above difficulties with the bijectivity, it is perhaps wise to formulate as well the graph particular case of Theorem 2.4. The statement here is as follows:
The adjacency matrix [math]d\in M_N(0,1)[/math] of any graph is diagonalizable, with the diagonalization being of the following type,
This follows from Theorem 2.4, via the various remarks from the proof of Theorem 2.5 and Theorem 2.6. But the simplest is to say that the statement itself is just a copy of Theorem 2.6, with [math]U\in O_N[/math] replaced by the more general [math]U\in U_N[/math].
All the above is useful, and we will use these results on a regular basis, in what follows. However, before getting into more concrete things, let us formulate: \begin{problem} Find a geometric proof of Theorem 2.6, or of Theorem 2.7, based on the interpretation of eigenvalues and eigenvectors from Theorem 2.1. \end{problem} This question looks quite reasonable, at a first glance, after all what we have in Theorem 2.1 is all nice and gentle material, so do we really need all the above complicated linear algebra machinery in order to deal with all this. However, at a second look, meaning after studying some examples, the problem suddenly looks very complicated. So, homework for you, in case I forget to assign this, to come back to this problem, later.
2b. The simplex
As an illustration for the above, let us diagonalize the adjacency matrix of the simplest graph that we know, namely the [math]N[/math]-simplex. Let us start with:
The adjacency matrix of the [math]N[/math]-simplex, having [math]0[/math] on the diagonal and [math]1[/math] elsewhere, is in matrix form:
Here the first assertion is clear from definitions, and the second assertion is clear too. As for the last assertion, observe first that with [math]P=\mathbb I/N[/math] we have:
Thus [math]P[/math] is a projection, and since we obviously have [math]P=P^t[/math], this matrix is an orthogonal projection. In order to find now the image, observe that for any vector [math]v\in\mathbb R^N[/math] we have the following formula, with [math]a\in\mathbb R[/math] being the average of the entries of [math]v[/math]:
We conclude that the image of [math]P[/math] is the vector space [math]\mathbb R\xi[/math], with [math]\xi\in\mathbb R^N[/math] being the all-1 vector, and so that [math]P[/math] is the orthogonal projection on [math]\xi[/math], as claimed.
The above is very nice, in particular with [math]d=NP-1[/math] basically diagonalizing [math]d[/math] for us. However, thinking a bit, when it comes to explicitly diagonalize [math]d[/math], or, equivalently, [math]P[/math] or [math]\mathbb I[/math], things are quite tricky, and we run into the following strange problem: \begin{problem} In order to diagonalize [math]d[/math], we need solutions for
and there are no standard such solutions, over the reals. \end{problem} So, this is the problem that we face, which might look a bit futile at a first glance, and in order for you to take me seriously here, let us work out some particular cases. At [math]N=2[/math] things are quickly solved, with the diagonalization being as follows, and with the passage matrix being easy to construct, I agree with you on this:
However, things become suddenly complicated at [math]N=3[/math], where I challenge you to find the passage matrix for the following diagonalization:
In the general case, [math]N\in\mathbb N[/math], the problem does not get any simpler, again with the challenge for you to find the passage matrix for the following diagonalization:
In short, you got my point, Problem 2.10 is something real. Fortunately the complex numbers come to the rescue, via the following standard and beautiful result:
The roots of unity, [math]\{w^k\}[/math] with [math]w=e^{2\pi i/N}[/math], have the property
There are several possible proofs for this, as follows:
(1) Nice proof. The numbers to be summed, when written more conveniently as [math](w^s)^k[/math] with [math]k=0,\ldots,N-1[/math], form a certain regular polygon in the plane [math]P_s[/math]. Thus, if we denote by [math]C_s[/math] the barycenter of this polygon, we have the following formula:
Now observe that in the case [math]N\slash\hskip-1.6mm|\,s[/math] our polygon [math]P_s[/math] is non-degenerate, circling along the unit circle, and having center [math]C_s=0[/math]. As for the case [math]N|s[/math], here the polygon is degenerate, lying at 1, and having center [math]C_s=1[/math]. Thus, we have the following formula, as claimed:
(2) Ugly proof. Normally we won't bother with ugly proofs in this book, but these being mathematics too, at least in theory, here is the computation, for [math]N\slash\hskip-1.6mm|\,s[/math]:
Thus, we are led again to the formula in the statement.
Summarizing, we have the solution to our problem. In order now to finalize, let us start with the following definition, inspired by what happens in Theorem 2.11:
The Fourier matrix [math]F_N[/math] is the following matrix, with [math]w=e^{2\pi i/N}[/math]:
Here the name comes from the fact that [math]F_N[/math] is the matrix of the discrete Fourier transform, that over the cyclic group [math]\mathbb Z_N[/math], and more on this later, when talking Fourier analysis. As a first example now, at [math]N=2[/math] the root of unity is [math]w=-1[/math], and with indices as above, namely [math]i,j\in\{0,1\}[/math], taken modulo 2, our Fourier matrix is as follows:
At [math]N=3[/math] now, the root of unity is [math]w=e^{2\pi i/3}[/math], and the Fourier matrix is:
At [math]N=4[/math] now, the root of unit is [math]w=i[/math], and the Fourier matrix is:
Finally, at [math]N=5[/math] the root of unity is [math]w=e^{2\pi i/5}[/math], and the Fourier matrix is:
You get the point. Getting back now to the diagonalization problem for the flat matrix [math]\mathbb I[/math], this can be solved by using the Fourier matrix [math]F_N[/math], in the following way:
The flat matrix [math]\mathbb I[/math] diagonalizes as follows,
According to the last assertion in Proposition 2.9, we are left with finding the 0-eigenvectors of [math]\mathbb I[/math], which amounts in solving the following equation:
For this purpose, we can use the root of unity [math]w=e^{2\pi i/N}[/math], and more specifically, the following standard formula, coming from Theorem 2.11:
This formula shows that for [math]j=1,\ldots,N-1[/math], the vector [math]v_j=(w^{ij})_i[/math] is a 0-eigenvector. Moreover, these vectors are pairwise orthogonal, because we have:
Thus, we have our basis [math]\{v_1,\ldots,v_{N-1}\}[/math] of 0-eigenvectors, and since the [math]N[/math]-eigenvector is [math]\xi=v_0[/math], the passage matrix [math]P[/math] that we are looking is given by:
But this is precisely the Fourier matrix, [math]P=F_N[/math]. In order to finish now, observe that the above computation of [math] \lt v_i,v_j \gt [/math] shows that [math]F_N/\sqrt{N}[/math] is unitary, and so:
Thus, we are led to the diagonalization formula in the statement.
By substracting now [math]-1[/math] from everything, we can formulate a final result, as follows:
The adjacency matrix of the simplex diagonalizes as follows,
This follows as said above, from what we have in Proposition 2.13, by substracting [math]-1[/math] from everything. Alternatively, if you prefer a more direct proof, this follows from the various computations in the proof of Proposition 2.13.
The above result was something quite tricky, and we will come back regularly to such things, in what follows. For the moment, let us formulate an interesting conclusion: \begin{conclusion} Theorem 2.7, telling us that [math]d[/math] diagonalizes over [math]\mathbb C[/math], is better than the stronger Theorem 2.6, telling us that [math]d[/math] diagonalizes over [math]\mathbb R[/math]. \end{conclusion} And isn't this surprising. But after some thinking, after all no surprise, because the graphs, as we defined them in the beginning of chapter 1, are scalarless objects. So, when needing a field for studying them, we should just go with the mighty [math]F=\mathbb C[/math].
By the way, regarding complex numbers, time to recommend some reading. Mathematically the book of Rudin [1] is a must-read, and a pleasure to read. However, if you want to be really in love with complex numbers, and with this being an enormous asset, no matter what mathematics you want to do, nothing beats some physics reading.
The standard place for learning physics is the course of Feynman [2], [3], [4]. If you already know a bit of physics, you can go as well with the lovely books of Griffiths [5], [6], [7]. And if you know a bit more, good books are those of Weinberg [8], [9], [10]. In the hope that this helps, and I will not tell you of course what these books contain. Expect however lots of complex numbers, all beautiful, and used majestically.
2c. Further graphs
Motivated by the above, let us try now to diagonalize the adjacency matrices of other graphs, more complicated than the simplex. But, what graphs to start with?
We already faced this kind of question in chapter 1, when discussing random walks and other basic questions, and we will proceed in a similar way here, by using our intuition. Passed the simplices, the “simplest” graphs are most likely the unions of simplices:
But the diagonalization question for such graphs is quickly settled, by using our results for the simplex from the previous section, and the following general fact:
Given a disjoint union of connected graphs [math]X=X_1\cup\ldots\cup X_k[/math], the corresponding adjacency matrix is block-diagonal,
This is indeed something trivial and self-explanatory, coming from definitions, the idea being that when labeling the vertices of [math]X[/math] by starting with those of [math]X_1[/math], then with those of [math]X_2[/math], and so on up to those of [math]X_k[/math], the adjacency matrix [math]d[/math] takes the block-diagonal form in the statement. Thus, we are led to the above conclusions.
Summarizing, we know how to diagonalize the adjacency matrix of [math]X=X_1\cup\ldots\cup X_k[/math], provided that we know how to diagonalize the adjacency matrix of each [math]X_i[/math]. As an illustration here, for the graph pictured before Theorem 2.16, the diagonalization is as follows, with [math]F_N[/math] being as usual the Fourier matrix, and with [math]J_N=diag(N-1,-1,\ldots,-1)[/math] being the diagonal [math]N\times N[/math] matrix found in Theorem 2.14:
Back to the general case, as in Theorem 2.16, of particular interest is the case where all the components [math]X_i[/math] appear as copies of the same graph [math]Y[/math]. And with the picture here, when [math]Y[/math] is a simplex, being something which looks quite nice, as follows:
In this case we can further build on Theorem 2.16, and say a bit more about the diagonalization of the adjacency matrix, the final result being as follows:
Given a disjoint union of type [math]X=kY[/math], with [math]Y[/math] being connected, the corresponding adjacency matrix can be written in the following way,
This follows from Theorem 2.16, with the only issue being at the level of the labeling of vertices, and of the tensor product notations. In order to explain this, let us take as example the graph pictured before the statement, and label it as follows:
Observe that this is in agreement with what we did in the proof of Theorem 2.16, with the vertices of the first square coming first, then with the vertices of the second square coming second, and with the vertices of the third square coming third, according to:
Thus, we can use the conclusions of Theorem 2.16, as such. But since our adjacency matrix has now double indices, as matrix indices, we can switch to tensor product notations, according to the following standard rule, from tensor product calculus:
Thus, we are led to the conclusions in the statement.
As before with Theorem 2.16, nothing better at this point than an illustration for all this. For the graph pictured before the statement, and in the proof too, we obtain:
Moving forward, now that we understood the disjoint union operation, time to study the other basic operation for graphs, which is the complementation operation. However, things here are more tricky, with the general result on the subject being as follows:
The adjacency matrix of a graph [math]X[/math] and of its complement [math]X^c[/math] are related by the following formula, with [math]\mathbb I[/math] being the all-one matrix,
This is something trivial, and which of course does not look very good, with our point coming from the following implication, which is immediate:
Thus, we are led to the conclusion in the statement.
In practice now, the next step would be to look at the complements of various particular graphs, such as the disjoint unions [math]X=X_1\cup\ldots\cup X_k[/math] from Theorem 2.16. But this does not bring any simplification, as you can easily verify yourself. Moreover, when further particularizing, and looking at the complements of the graphs [math]X=kY[/math] from Theorem 2.17, there is no simplification either. However, when particularizing even more, and taking [math]Y=K_n[/math], we are led to something nice, which is good to know, namely:
Given a disjoint union of simplices [math]X=kK_n[/math], the adjacency matrix of the complement [math]X^c[/math] can be written as
By using the various labeling and tensor product conventions from Theorem 2.17 and its proof, the adjacency matrix of [math](kK_n)^c[/math] is given by:
Thus, we are led to the various conclusions in the statement.
As before with other results of the same type, let us work out an illustration. Consider the following graph, that we already met before, in the context of Theorem 2.17:
The adjacency matrix of the complementary graph diagonalizes then as follows:
And with this being actually not the end of the story, for this particular graph, because it is possible to prove that we have [math]F_{12}=F_4\otimes F_3[/math], with this coming from [math]F_{mn}=F_m\otimes F_n[/math], for [math]m,n[/math] prime to each other. But more on such phenomena later in this book.
Moving forward, now that we have some good understanding of the various operations for the graphs, time to get into the real thing, namely study of the graphs which are nice, simple and “indecomposable”. Although we don't have yet a definition, for what indecomposable should mean, as a main example here, we certainly have the circle:
So, in what follows we will study this circle graph, and its complement too. Let us first compute the eingenvalues and eigenvectors. This is easily done, as follows:
For the adjacency matrix [math]d[/math] of the circle graph, we have
In what regards the first assertion, this follows from:
As for the second assertion, this comes from this, because with [math]k=0,1,\ldots,N-1[/math] the eigenvectors that we found, which are the rows of [math]F_N[/math], are linearly independent.
We can now formulate a final result about the circle graph, as follows:
The adjacency matrix of the circle graph diagonalizes as
In terms of the vector [math]\xi=(w^p)[/math], and as usual with indices [math]p=0,1,\ldots,N-1[/math], the eigenvector formula that we found in Proposition 2.20 reads:
By putting all these formulae together, with [math]k=0,1,\ldots,N-1[/math], we obtain:
But, since we have [math]F_N=(\xi^k)[/math], as a square matrix, this formula reads:
Now by multipliying to the right by [math]F_N^{-1}=F_N^*/N[/math], we obtain the result.
The above result is quite interesting, and as a first obvious conclusion, coming from this and from our previous results regarding the simplex, we have: \begin{conclusion} The adjacency matrices of both the simplex and the circle are diagonalized by the Fourier matrix [math]F_N[/math]. \end{conclusion} We will be back to this phenomenon later in this book, with a systematic discussion of the graphs diagonalized by the Fourier matrix [math]F_N[/math], and of related topics.
In the meantime, as a corollary of this, remember Proposition 2.18, regarding the complementary graphs [math]X^c[/math], with the quite technical assumptions that we found there, seemingly bound for the trash can? Well, as good news, Conclusion 2.22 is exactly what we need, guaranteeing that Proposition 2.18 applies indeed, to the circle graph. So, as our next theorem, based on the above study for the circle, we have:
The adjacency matrix of the complement of the circle graph diagonalizes as
Consider the circle graph [math]X[/math], that we know well from the above:
Consider as well its complement [math]X^c[/math], which is a quite crowded and scary graph, of valence [math]N-3[/math], with the picture at [math]N=8[/math] being as follows:
However, at the linear algebra level things are quite simple, because by combining Theorem 2.14 and Theorem 2.21, as indicated in Proposition 2.18, we obtain:
Thus, we are led to the formula in the statement.
Very nice all this, so done with the circle and its complement, and on our to-do list, job for later to further explore Conclusion 2.22, and related Fourier analysis topics.
2d. The segment
As a continuation of the above study, let us discuss now the segment graph. This is a graph that we already met in chapter 1, in relation with random walk questions, and with our conclusions there being quite muddy, basically amounting in saying that this graph is something quite tough to fight with, knowing more mathematics than we do.
We will investigate here this graph, from the diagonalization point of view. As a first observation, since diagonalization solves the random walk question, as explained in chapter 1, we cannot expect much from our diagonalization study, done as we do, with bare hands. So, modesty, let us just start working on this, and we'll see what we get.
Some numerics first. These do not look very good, the result being as follows:
The eigenvalues of the segment graph are as follows:
- At [math]N=2[/math] we have [math]-1,1[/math].
- At [math]N=3[/math] we have [math]0,\pm\sqrt{2}[/math].
- At [math]N=4[/math] we have [math]\pm\frac{\sqrt{5}\pm1}{2}[/math].
- At [math]N=5[/math] we have [math]0,\pm1,\pm\sqrt{3}[/math].
- At [math]N=6[/math] we have the solutions of [math]x^6-5x^4+6x^2-1=0[/math].
These are some straightforward linear algebra computations, with some tricks being needed only at [math]N=4[/math], the details being as follows:
(1) At [math]N=2[/math] the adjacency matrix and its eigenvalues are as follows:
(2) At [math]N=3[/math] the adjacency matrix and its eigenvalues are as follows:
(3) At [math]N=4[/math] the adjacency matrix and its characteristic polynomial are as follows:
Now by solving the degree 2 equation, and fine-tuning the answer, we obtain:
(4) At [math]N=5[/math] the adjacency matrix and its eigenvalues are as follows:
(5) At [math]N=6[/math] the adjacency matrix and its eigenvalues are as follows:
Thus, we are led to the formulae in the statement.
All the above does not look very good. However, as a matter of having the dirty job fully done, with mathematical pride, let us look as well into eigenvectors. At [math]N=2[/math] things are quickly settled, with the diagonalization of the adjacency matrix being as follows:
At [math]N=3[/math] the diagonalization formula becomes more complicated, as follows:
At [math]N=4[/math] now, in view of the eigenvalue formula that we found, [math]x^4-3x^2+1=0[/math], we must proceed with care. The equation for the eigenvectors [math]dv=xv[/math] is as follows:
In other words, the equations to be satisfied are as follows:
With the choice [math]a=1[/math], the solutions of these equations are as follows:
In order to compute now the [math]c[/math] components of the eigenvectors, we can use the formula [math]x=(\pm1\pm\sqrt{5})/2[/math] from Proposition 2.24. Indeed, this formula gives:
Thus, almost done, and we deduce that the passage matrix is as follows:
To be more precise, according to our equations above, the first row must consist of [math]a=1[/math] entries. Then the second row must consist of [math]b=x[/math] entries, with [math]x=(\pm1\pm\sqrt{5})/2[/math]. Then the third row must consist of [math]c=x^2-1[/math] entries, but these are easily computed, as explained above. Finally, the fourth row must consist of [math]d=(x^2-1)/x[/math] entries, which means that the fourth row appears by dividing the third row by the second row, which is easily done too. In case you wonder why [math]d=\pm1[/math], here is another proof of this:
Very nice all this, and leaving the computation of [math]P^{-1}[/math] for you, here are a few more observations, in relation with what we found in Proposition 2.24:
(1) In the case [math]N=5[/math] the eigenvectors can be computed too, and the diagonalization finished, via the standard method, namely system of equations, and with the numerics involving powers of the eigenvalues that we found. Exercise for you.
(2) The same stays true at [math]N=6[/math], again with the eigenvector numerics involving powers of the eigenvalues, and with these eigenvalues being explicitly computable, via the Cardano formula for degree 3 equations. Have fun with this too, of course.
(3) However, all this does not look very good, and at [math]N=7[/math] and higher we will certainly run into difficult questions, and save for some interesting remarks, normally depending on the parity of [math]N[/math], we will not be able to fully solve the diagonalization problem.
So, what to do? Work some more, of course, the hard way. Proposition 2.24 and its proof look quite trivial, but if you get into the full details of the computations, and let me assign here this, as a key exercise, you will certainly notice that, when computing that determinants, you can sort of use a recurrence method. And, this leads to:
The characteristic polynomials [math]P_N[/math] of the segment graphs satisfy
Obviously, many things going on here, ranging from precise to definitional, or even informal, the idea with all this being as follows:
(1) By computing determinants, as indicated above, we are led to the recurrence formula in the statement. Here is the proof at [math]N=4[/math], the general case being similar:
(2) Regarding now the intial values, according to Proposition 2.24 these should normally be [math]P_2=x^2-1[/math] and [math]P_3=x^3-2x[/math], but we can formally add the values [math]P_0=1[/math] and [math]P_1=x[/math], as for the final statement to look better, with this being justified by:
(3) Thus, we have the recurrence formula in the statement, and with the initial values there, and an internet search, or some advanced calculus know-how, tells us that these are the well-known Chebycheff polynomials, enjoying lots of interesting properties.
Very nice all this, and as a continuation, barring as usual the wish to fully solve the diagonalization problem, which remains something difficult, we can for instance go back now to the random walk computations from chapter 1, for the segment graph, with the various choices there for the distinguished vertex, and improve our results there.
However, while this is certainly something doable and interesting, we will not do it here, and rather keep the above as a mere remark, or as an exercise for you. This is because we will be back to all this in the next chapter, with some even better tools.
General references
Banica, Teo (2024). "Graphs and their symmetries". arXiv:2406.03664 [math.CO].
References
- W. Rudin, Real and complex analysis, McGraw-Hill (1966).
- R.P. Feynman, R.B. Leighton and M. Sands, The Feynman lectures on physics I: mainly mechanics, radiation and heat, Caltech (1963).
- R.P. Feynman, R.B. Leighton and M. Sands, The Feynman lectures on physics II: mainly electromagnetism and matter, Caltech (1964).
- R.P. Feynman, R.B. Leighton and M. Sands, The Feynman lectures on physics III: quantum mechanics, Caltech (1966).
- D.J. Griffiths, Introduction to electrodynamics, Cambridge Univ. Press (2017).
- D.J. Griffiths and D.F. Schroeter, Introduction to quantum mechanics, Cambridge Univ. Press (2018).
- D.J. Griffiths, Introduction to elementary particles, Wiley (2020).
- S. Weinberg, Foundations of modern physics, Cambridge Univ. Press (2011).
- S. Weinberg, Lectures on quantum mechanics, Cambridge Univ. Press (2012).
- S. Weinberg, Lectures on astrophysics, Cambridge Univ. Press (2019).