Local estimates
12a. Norm maximizers
We discuss here some further analytic questions, regarding the complex Hadamard matrices, following [1], in analogy with the considerations from chapter 3. We will be interested in the complex analogue of the notion of almost Hadamard matrix. This looks more as a routine topic, and for a long time it was believed that there is no hurry in developing all this, since complex Hadamard matrices exist anyway at any [math]N\in\mathbb N[/math], and so there is no really need for almost Hadamard matrices, in the complex setting.
However, some work on this subject was eventually done in [1], and surprise, it turned out that, at least conjecturally, there are no almost Hadamard matrices, in the complex sense. Which is very good news, because this shows, again conjecturally, that for a matrix [math]H\in\sqrt{N}U_N[/math], the property of being complex Hadamard is “local”. Which itself is a surprising and potentially far-reaching statement, suggesting reformulating all the Hadamard matrix problematics, including the HC and CHC, in local terms.
We will explain this exciting material in this chapter. To start with, we have the following basic estimate, that we already know, from chapter 11:
Given [math]\psi:[0,\infty)\to\mathbb R[/math], the following function over [math]U_N[/math],
This follows indeed from the Jensen inequality applied to the function in the statement, exactly as in the real case, as explained in chapter 2.
Of particular interest for us are the power functions [math]\psi(x)=x^{p/2}[/math], which are concave at [math]p\in[1,2)[/math], and convex at [math]p\in(2,\infty)[/math]. These lead to the following statement:
Let [math]U\in U_N[/math], and set [math]H=\sqrt{N}U[/math].
- For [math]p\in[1,2)[/math] we have [math]||U||_p\leq N^{2/p-1/2}[/math],
- For [math]p\in(2,\infty][/math] we have [math]||U||_p\geq N^{2/p-1/2}[/math].
In both cases, the equality situation happens precisely when [math]H[/math] is Hadamard.
Consider indeed the [math]p[/math]-norm on [math]U_N[/math], which at [math]p\in[1,\infty)[/math] is given by:
By the above discussion, involving the functions [math]\psi(x)=x^{p/2}[/math], Proposition 12.1 applies and gives the results at [math]p\in[1,\infty)[/math], the precise estimates being as follows:
As for the case [math]p=\infty[/math], this follows with [math]p\to\infty[/math], or directly via Cauchy-Schwarz.
For future reference, let us record as well the particular cases [math]p=1,4,\infty[/math] of the above result, that we already met before, and which are of particular interest:
For any matrix [math]U\in U_N[/math] we have the estimates
These results follow from Theorem 12.2 at [math]p=1,4,\infty[/math], with the remark that for each of these particular exponents, we do not really need the Hölder inequality, with a basic application of the Cauchy-Schwarz inequality doing the job.
The above results suggest the following definition:
Given [math]U\in U_N[/math], the matrix [math]H=\sqrt{N}U[/math] is called:
- Almost Hadamard, if [math]U[/math] locally maximizes the [math]1[/math]-norm on [math]U_N[/math].
- [math]p[/math]-almost Hadamard, with [math]p \lt 2[/math], if [math]U[/math] locally maximizes the [math]p[/math]-norm on [math]U_N[/math].
- [math]p[/math]-almost Hadamard, with [math]p \gt 2[/math], if [math]U[/math] locally minimizes the [math]p[/math]-norm on [math]U_N[/math].
- Absolute almost Hadamard, if it is [math]p[/math]-almost Hadamard at any [math]p\neq2[/math].
We have as well real versions of these notions, with [math]U_N[/math] replaced by [math]O_N[/math].
All this might seem a bit complicated, but this is the best way of presenting things. We are mainly interested in (1), but as explained in chapter 9, the exponent [math]p=4[/math] from (3) is interesting as well, and once we have (3) we must formulate (2) as well, and finally (4) is a useful thing too, because the absolute case is sometimes easier to study. As for the “doubling” of all these notions, via the last sentence, this is necessary too, because given a function [math]F:U_N\to\mathbb R[/math], an element [math]U\in O_N[/math] can be a local extremum of the restriction [math]F_{|O_N}:O_N\to\mathbb R[/math], but not of the function [math]F[/math] itself, and we will see examples of this.
Let us first study the critical points. Things are quite tricky here, and complete results are available so far only at [math]p=1[/math]. Following [1], we first have the following result:
If [math]U\in U_N[/math] locally maximizes the [math]1[/math]-norm, then
We use the same method as in the real case, namely a rotation trick. Let us denote by [math]U_1,\ldots,U_N[/math] the rows of [math]U[/math], and let us perform a rotation of [math]U_1,U_2[/math]:
In order to compute the 1-norm, let us permute the columns of [math]U[/math], in such a way that the first two rows look as follows, with [math]X,Y,A,B[/math] having nonzero entries:
The rotated matrix will look then as follows:
Our claim is that [math]X,Y[/math] must be empty. Indeed, if [math]A[/math] and [math]B[/math] are not empty, let us fix a column index [math]k[/math] for both [math]A,B[/math], and set [math]\alpha=A_k[/math], [math]\beta=B_k[/math]. We have then:
Since [math]\alpha,\beta\neq 0[/math], the above function is differentiable at [math]t=0[/math], and we obtain:
Thus at [math]t=0[/math], we obtain the following formula:
Now since our matrix [math]U[/math] locally maximizes the 1-norm, both directional derivatives of [math]||U^t||_1[/math] must be negative in the limit [math]t\to 0[/math]. On the other hand, if we denote by [math]C[/math] the contribution coming from the right, which might be zero in the case where [math]A[/math] and [math]B[/math] are empty, i.e. the sum over [math]k[/math] of the above quantities, we have:
As for the derivative at left, this is given by the following formula:
We therefore obtain the following inequalities, where [math]C[/math] is as above:
Consider now the matrix obtained from [math]U[/math] by interchanging [math]U_1,U_2[/math]. Since this matrix must be as well a local maximizer of the 1-norm, and since the above formula shows that [math]C[/math] changes its sign when interchanging [math]U_1,U_2[/math], we obtain:
The four inequalities that we have give altogether the following conclusion:
Now from [math]||X||_1+||Y||_1=0[/math] we obtain that both the vectors [math]X,Y[/math] must be empty, as claimed. As a conclusion, up to a permutation of the columns, the first two rows of our matrix [math]U[/math] must be of the following form, with [math]A,B[/math] having only nonzero entries:
By permuting the rows of [math]U[/math], the same must hold for any two rows [math]U_i,U_j[/math]. Now since [math]U[/math] cannot have a zero column, we conclude that [math]U[/math] cannot have zero entries, as claimed.
Let us compute now the critical points. Following [1], we have:
Let [math]\varphi:[0,\infty)\to\mathbb R[/math] be a differentiable function. A unitary matrix with nonzero entries [math]U\in U_N^*[/math] is a critical point of the quantity
Again, this follows like in the real case, by performing modifications where needed. We regard [math]U_N[/math] as a real algebraic manifold, with coordinates [math]U_{ij},\bar{U}_{ij}[/math]. This manifold consists by definition of the zeroes of the following polynomials:
A given matrix [math]U\in U_N[/math] is then a critical point of [math]F[/math] precisely when [math]dF\in span(dA_{ij})[/math]. Regarding the space [math]span(dA_{ij})[/math], this consists of the following quantities:
In order to compute [math]dF[/math], observe first that, with [math]S_{ij}=sgn(U_{ij})[/math], we have:
In terms of [math]W_{ij}=sgn(U_{ij})\varphi'(|U_{ij}|)[/math], as in the statement, we obtain:
We conclude that [math]U\in U_N[/math] is a critical point of [math]F[/math] if and only if there exists a matrix [math]M\in M_N(\mathbb C)[/math] such that the following two conditions are satisfied:
But this means [math]WU^*=UW^*[/math], and so that [math]WU^*[/math] must be self-adjoint, as claimed.
12b. Balanced matrices
In order to process the above result, we proceed exactly as in chapter 3, by adding some complex conjugates where needed. We can use the following notion:
Given [math]U\in U_N[/math], we consider its “color decomposition”
- Semi-balanced, if [math]U_rU^*[/math] and [math]U^*U_r[/math], with [math]r \gt 0[/math], are all self-adjoint.
- Balanced, if [math]U_rU_s^*[/math] and [math]U_r^*U_s[/math], with [math]r,s \gt 0[/math], are all self-adjoint.
These conditions are quite natural, because for a unitary matrix [math]U\in U_N[/math], the relations [math]UU^*=U^*U=1[/math] translate as follows, in terms of the color decomposition:
Thus, our balancing conditions express the fact that the various components of the above sums all self-adjoint. Now back to our critical point questions, we have:
For a matrix [math]U\in U_N^*[/math], the following are equivalent:
- [math]U[/math] is a critical point of [math]F(U)=\sum_{ij}\varphi(|U_{ij}|)[/math], for any [math]\varphi:[0,\infty)\to\mathbb R[/math].
- [math]U[/math] is a critical point of all the [math]p[/math]-norms, with [math]p\in[1,\infty)[/math].
- [math]U[/math] is semi-balanced, in the above sense.
We use Theorem 12.6. The matrix constructed there is given by:
We conclude that we have the following formula for this matrix:
Now when [math]\varphi:[0,\infty)\to\mathbb R[/math] varies, as a differentiable function, or as a power function [math]\varphi(x)=x^p[/math] with [math]p\in[1,\infty)[/math], the individual components must be self-adjoint, as desired.
In practice now, most of the known examples of semi-balanced matrices are actually balanced. We have the following collection of simple facts, regarding such matrices:
The class of balanced matrices is as follows:
- It contains the matrices [math]U=H/\sqrt{N}[/math], with [math]H\in M_N(\mathbb C)[/math] Hadamard.
- It is stable under transposition, complex conjugation, and taking adjoints.
- It is stable under taking tensor products.
- It is stable under the Hadamard equivalence relation.
- It contains the matrix [math]V_N=\frac{1}{N}(2\mathbb I_N-N1_N)[/math], where [math]\mathbb I_N[/math] is the all-[math]1[/math] matrix.
All these results are elementary, the proof being as follows:
(1) Here [math]U\in U_N[/math] follows from the Hadamard condition, and since there is only one color component, namely [math]U_{1/\sqrt{N}}=H[/math], the balancing condition is satisfied as well.
(2) Assuming that [math]U=\sum_{r \gt 0}rU_r[/math] is a color decomposition of a given matrix [math]U\in U_N[/math], the following are color decompositions too, and this gives the assertions:
(3) Assuming that [math]U=\sum_{r \gt 0}rU_r[/math] and [math]V=\sum_{s \gt 0}sV_s[/math] are the color decompositions of two given unitary matrices [math]U,V[/math], we have the following formula:
Thus the color components of [math]W=U\otimes V[/math] are the following matrices:
It follows that if [math]U,V[/math] are both balanced, then so is [math]W=U\otimes V[/math].
(4) We recall that the Hadamard equivalence consists in permuting rows and columns, and switching signs on rows and columns. Since all these operations correspond to certain conjugations at the level of the matrices [math]U_rU_s^*,U_r^*U_s[/math], we obtain the result.
(5) The matrix in the statement, which goes back to [2], is as follows:
Observe that this matrix is indeed unitary, its rows being of norm one, and pairwise orthogonal. The color components of this matrix are:
It follows that this matrix is balanced as well, as claimed.
Let us look now more in detail at [math]V_N[/math], and at the matrices having similar properties. Following [2], let us call [math](a,b,c)[/math] pattern any matrix [math]M\in M_N(0,1)[/math], with [math]N=a+2b+c[/math], such that any two rows look as follows, up to a permutation of the columns:
As explained in [2], there are many interesting examples of [math](a,b,c)[/math] patterns, coming from the balanced incomplete block designs (BIBD), and all these examples can produce two-entry unitary matrices, by replacing the [math]0,1[/math] entries with suitable numbers [math]x,y[/math]. Now back to the matrix [math]V_N[/math] from Proposition 12.9 (5), observe that this matrix comes from a [math](0,1,N-2)[/math] pattern, in the above sense. And also, independently of this, this matrix has the remarkable property of being at the same time circulant and self-adjoint. We have in fact the following result, generalizing Proposition 12.9 (5):
The following matrices are balanced:
- The orthogonal matrices coming from [math](a,b,c)[/math] patterns.
- The unitary matrices which are circulant and self-adjoint.
These observations basically go back to [2], the proofs being as follows:
(1) If we denote by [math]P,Q\in M_N(0,1)[/math] the matrices describing the positions of the [math]0,1[/math] entries inside the pattern, then we have the following formulae:
Since all these matrices are symmetric, [math]U[/math] is balanced, as claimed.
(2) Assume that [math]U\in U_N[/math] is circulant, [math]U_{ij}=\gamma_{j-i}[/math], and in addition self-adjoint, which means [math]\bar{\gamma}_i=\gamma_{-i}[/math]. Consider the following sets, which must satisfy [math]D_r=-D_r[/math]:
In terms of these sets, we have the following formula:
With [math]k=i+j-m[/math] we obtain, by using [math]D_r=-D_r[/math], and then [math]\bar{\gamma}_i=\gamma_{-i}[/math]:
Now by interchanging [math]i\leftrightarrow j[/math], and with [math]m\to k[/math], this formula becomes:
We recognize here the complex conjugate of [math](U_rU_s^*)_{ij}[/math], as previously computed above, and we therefore deduce that [math]U_rU_s^*[/math] is self-adjoint. The proof for [math]U_r^*U_s[/math] is similar.
12c. Hessian computations
Let us compute now derivatives. As in Theorem 12.6, it is convenient to do the computations in a more general framework, where we have a function as follows:
In order to study the local extrema of these quantities, consider the following function:
Here [math]U\in U_N[/math] is a unitary matrix, and [math]A\in M_N(\mathbb C)[/math] is assumed to be anti-hermitian, [math]A^*=-A[/math], as for having [math]e^A\in U_N[/math]. Let us first compute the derivative of [math]f[/math]. We have:
We have the following formula,
The matrices [math]U,e^{tA}[/math] being both unitary, we have:
We can now differentiate our function [math]f[/math], and by using once again the unitarity of the matrices [math]U,e^{tA}[/math], along with the formula [math]A^*=-A[/math], we obtain:
But this gives the formula in the statement, and we are done.
Before computing the second derivative, let us evaluate [math]f'(0)[/math]. We have:
We have the following formula,
We use the formula in Proposition 12.11. At [math]t=0[/math], we obtain:
Consider now the color decomposition of [math]U[/math]. We have the following formulae:
Now by getting back to the above formula of [math]f'(0)[/math], we obtain:
Our claim now is that we have the following formula:
Indeed, in the case [math]|U_{ij}|\neq r[/math] this formula reads [math]\overline{U}_{ij}\cdot 0=r\cdot 0[/math], which is true, and in the case [math]|U_{ij}|=r[/math] this formula reads [math]r\bar{S}_{ij}\cdot 1=r\cdot\bar{S}_{ij}[/math], which is once again true. Thus:
But this gives the formula in the statement, and we are done.
Let us compute now the second derivative. The result here is as follows:
We have the following formula,
We use the formula in Proposition 12.11, namely:
Since the real part on the right, or rather its double, appears as the derivative of the quantity [math]|(Ue^{tA})_{ij}|^2[/math], when differentiating a second time, we obtain:
In order to compute now the missing derivative, observe that we have:
Summing up, we have obtained the following formula:
But at [math]t=0[/math] this gives the formula in the statement, and we are done.
By using the function [math]\psi(x)=\sqrt{x}[/math], corresponding to [math]F(U)=||U||_1[/math], we obtain:
Let [math]U \in U_N^*[/math]. For the function [math]F(U)=||U||_1[/math] we have the formula
We use the formula in Proposition 12.13, with the following data:
We obtain the following formula:
But this gives the formula in the statement, and we are done.
We are therefore led to the following result, regarding the 1-norm:
A matrix [math]U\in U_N^*[/math] locally maximizes the one-norm on [math]U_N[/math] precisely when [math]S^*U[/math] is self-adjoint, where [math]S_{ij}={\rm sgn}(U_{ij})[/math], and when
According to Theorem 12.6 and Proposition 12.14, the local maximizer condition requires [math]X=S^*U[/math] to be self-adjoint, and the following inequality to be satisfied:
Now observe that since both [math]X[/math] and [math]A^2[/math] are self-adjoint, we have:
Thus we can remove the real part, and we obtain the inequality in the statement.
As a comment here, the above computations can be of course interpreted by using more advanced differential geometric language. The unitary group [math]U_N[/math] is a Lie group, and its tangent space at [math]U\in U_N[/math] is isomorphic to the corresponding Lie algebra, which consists of the anti-hermitian matrices [math]A\in M_N(\mathbb C)[/math]. With this picture in hand, our formulae for [math]f'(0)[/math] translate into the fact that the gradient of the 1-norm is given by:
Regarding now the second derivative, [math]f''(0)[/math], our computations here provide us with a formula for the Hessian of the 1-norm. Indeed, with the change of variables [math]A=iB[/math] on the tangent space, the Hessian [math]H[/math] of the 1-norm is given by the following formula, where [math]\Phi(U,iA)[/math] is the quantity appearing in Theorem 12.15:
Getting back to Theorem 12.15 as stated, the story is of course not over there. In order to further improve this result, we will need the following standard fact:
For a self-adjoint [math]X\in M_N(\mathbb C)[/math], the following are equivalent:
- [math]Tr(XA^2)\leq0[/math], for any anti-hermitian matrix [math]A\in M_N(\mathbb C)[/math].
- [math]Tr(XB^2)\geq0[/math], for any hermitian matrix [math]B\in M_N(\mathbb C)[/math].
- [math]Tr(XC)\geq0[/math], for any positive matrix [math]C\in M_N(\mathbb C)[/math].
- [math]X\geq0[/math].
These equivalences are well-known, the proof being as follows:
[math](1)\implies(2)[/math] follows by taking [math]B=iA[/math].
[math](2)\implies(3)[/math] follows by taking [math]C=B^2[/math].
[math](3)\implies(4)[/math] follows by diagonalizing [math]X[/math], and then taking [math]C[/math] to be diagonal.
[math](4)\implies(1)[/math] is clear as well, because with [math]Y=\sqrt{X}[/math] we have:
Thus, the above four conditions are indeed equivalent.
Following [1], we can now formulate a final result on the subject, as follows:
Given [math]U\in U_N[/math], set [math]S_{ij}={\rm sgn}(U_{ij})[/math], and let:
is positive, for any hermitian matrix [math]B\in M_N(\mathbb C)[/math].
This follows from Theorem 12.15, by setting [math]A=iB[/math], and by using Proposition 12.16, which shows that we must have indeed [math]X\geq0[/math].
12d. The conjecture
In relation with the above, quite surprisingly, the basic real almost Hadamard matrix [math]K_N[/math] is not an almost Hadamard matrix in the complex sense. That is, while [math]K_N/\sqrt{N}[/math] locally maximizes the 1-norm on [math]O_N[/math], it does not do so over [math]U_N[/math]. Moreover, as we will see in a moment, the same happens for the other basic real almost Hadamard matrices discussed in chapter 3, such as the circulant ones, and the 2-entry ones studied there. Thus, the situation in the complex case is drastically different from the one in the real case, and we are led in this way to the following remarkable statement: \begin{conjecture}[Almost Hadamard conjecture (AHC)] Any local maximizer of the [math]1[/math]-norm on [math]U_N[/math],
must be a global maximizer, i.e. must be a rescaled Hadamard matrix. \end{conjecture}
In other words, our conjecture is that, in the complex setting, almost Hadamard implies Hadamard. This would be something very useful, because we would have here a new approach to the complex Hadamard matrices, which is analytic and local. Which new approach, importantly, could potentially shed some new light on all the Hadamard matrix problems, be them real or complex, including the HC and CHC.
In order to explain all this, and the evidence that we have for the above conjecture, let us study more in detail the quantity [math]\Phi(U,B)[/math] from Theorem 12.17, namely:
As a first observation here, we have the following result:
With [math]S_{ij}=sgn(U_{ij})[/math] and [math]X=S^*U[/math] as above, we have
The matrices [math]X,B,D[/math] being all self-adjoint, we have:
Thus when computing [math]\Phi(U,B+D)[/math], the trace term decomposes as follows:
Regarding now the second term, in order to compute it, observe that with the notation [math]D=diag(\lambda_1,\ldots,\lambda_N)[/math], with [math]\lambda_i\in\mathbb R[/math], we have the following formula:
Thus the second term decomposes as follows:
Now observe that the middle term in this expression is given by:
As for the term on the right in the above expression, this is given by:
Thus when doing the substraction we obtain [math]\Phi(U,B+D)=\Phi(U,B)[/math], as claimed.
Observe that with [math]B=0[/math] we obtain [math]\Phi(U,D)=0[/math], for any [math]D\in M_N(\mathbb R)[/math] diagonal, so the inequality is Theorem 12.17 is an equality, when [math]B[/math] is diagonal. Getting now to the real thing, we have the following result, providing the first piece of evidence for the AHC:
Consider the matrix [math]U=(2\mathbb I_N-N1_N)/N[/math]. Assuming that a matrix [math]B\in M_N(\mathbb R)[/math] is symmetric and satisfies [math]UB=\lambda B[/math], we have
- For [math]B=\mathbb I_N[/math] we have the formula
[[math]] \Phi(U,B)=\frac{N^2(N-1)(N-4)}{2(N-2)} [[/math]]and this quantity is negative at [math]N=3[/math].
- For [math]B\in M_N(\mathbb R)[/math] nonzero, symmetric, with [math]B\,\mathbb I_N=0[/math], [math]diag(B)=0[/math] we have
[[math]] \Phi(U,B)=\left(2-\frac{N}{2}\right)Tr(B^2) [[/math]]and this quantity is negative at [math]N\geq5[/math].
With [math]U\in O_N[/math], [math]B\in M_N(\mathbb R)[/math], the formula in Theorem 12.17 reads:
Asusming now [math]U=\frac{1}{N}(2\mathbb I_N-N1_N)[/math] and [math]UB=\lambda B[/math], this formula becomes:
Now observe that in our case, we have the following formula:
Thus the trace term is given by the following formula:
Regarding now the sum on the right, this can be computed as follows:
We obtain the following formula, which gives the one in the statement:
We can now prove our various results, as follows:
(1) For [math]B=\mathbb I_N[/math] we have [math]\lambda=1[/math], and we obtain, as claimed:
(2) For [math]B\in M_N(\mathbb R)[/math] nonzero, symmetric, and satisfying [math]B\,\mathbb I_N=0[/math] and [math]diag(B)=0[/math], we have [math]\lambda=-1[/math], and we obtain, as claimed:
It remains to prove that matrices [math]B[/math] as in the statement exist, at any [math]N\geq5[/math]. As a first remark, such matrices cannot exist at [math]N=2,3[/math]. At [math]N=4[/math], however, we have solutions, which are as follows, with [math]x+y+z=0[/math], not all zero:
At [math]N\geq5[/math] now, we can simply use this matrix, completed with [math]0[/math] entries, and we are led to the conclusion in the statement.
Let us go back now to the inequality in Theorem 12.17. When [math]U[/math] is a rescaled complex Hadamard matrix we have of course equality, and in addition, the following happens:
For a rescaled complex Hadamard matrix, a stronger version of the inequality in Theorem 12.17, namely [math]\Phi(U,B)\geq0[/math] with
holds, with the real part replaced by the absolute value.
Indeed, for a rescaled Hadamard matrix [math]U=H/\sqrt{N}[/math] we have:
Thus [math]X=\sqrt{N}1_N[/math]. We therefore obtain the following estimate:
But this proves our claim, and we are done.
In relation with the Tadej-\.Zyczkowski notion of defect [3], we have:
For a rescaled complex Hadamard matrix, the space
Since a self-adjoint matrix [math]B\in M_N(\mathbb C)[/math] belongs to [math]E_U[/math] precisely when the only inequality in the proof of Proposition 12.21 is saturated, we have:
The condition on the right tells us that the matrix [math]A=(UB)_{ij}\bar{U}_{ij}[/math] must be real. Now since the construction [math]B\to A[/math] is injective, we obtain an isomorphism, as follows:
Our claim is that the space on the right is [math]D_U[/math]. Indeed, let us pick [math]A\in M_N(\mathbb R)[/math]. The condition [math]A_{ij}=(UB)_{ij}\bar{U}_{ij}[/math] is then equivalent to:
Thus in terms of the matrix [math]C_{ij}=U_{ij}A_{ij}[/math] we have [math](UB)_{ij}=NC_{ij}[/math], and so:
Thus we have [math]B=NU^*C[/math], and we can now perform the study of the self-adjointness condition [math]B=B^*[/math], as follows:
Thus we have reached to the condition defining [math]D_U[/math], and we are done.
Regarding now the known verifications of the AHC, these basically concern the natural “candidates” coming from Theorem 12.10, as well as some straightforward complex generalizations of these candidates. All this is quite technical, and generally speaking, we refer here to [1]. As a first illustration, however, which is of theoretical importance, in the circulant orthogonal case, we have the following result, from [1]:
If [math]U\in O_N[/math] is circulant, [math]U_{ij}=\gamma_{j-i}[/math], we have
where the symbol [math]\mathbb E[/math] stands as usual for “average”.
We have [math]U\mathbb I_N=u\mathbb I_N[/math], which gives the following formula:
Similarly, once again from [math]U\mathbb I_N=u\mathbb I_N[/math], we obtain the following formula:
By substracting, we obtain the formula in the statement, which gives the result.
Here is another exclusion criterion, also from [1], which is useful as well:
If [math]U\in U_N[/math] is circulant, [math]U_{ij}=\gamma_{i-j}[/math], and self-adjoint, we have
Since [math]U[/math] is circulant and self-adjoint, we have [math]U=Fdiag(\beta)F^*[/math], for some vector [math]\beta\in\{\pm 1\}^N[/math]. The first term in the expression of [math]\Phi(U,U)[/math] reads:
For the second term in the formula of [math]\Phi[/math], we have the following formula:
We therefore obtain the following formula:
But this leads to the conclusion in the statement.
Still following [1], here is now a more advanced result, also in the circulant self-adjoint case, making this time use of a random derivative method:
If [math]U\in U_N[/math] is circulant, [math]U_{ij}=\gamma_{j-i}[/math], and self-adjoint, we have
Since [math]B[/math] is circulant, this matrix is Fourier-diagonal. That is, we can diagonalize it with the help of the normalized Fourier matrix [math]F=F_N/\sqrt{N}[/math], as follows:
The requirement that [math]B[/math] is unitary and self-adjoint amounts then to [math]\alpha_i=\pm 1[/math]. The expectation is taken in the probability space where the random variables [math]\alpha_i[/math] are i.i.d., with symmetric Bernoulli distributions [math](\delta_{-1}+\delta_1)/2[/math]. In particular, we have:
By using [math]B^2=1_N[/math], the first term in the expression of [math]\Phi(U,B)[/math] reads:
For the second term in the formula of [math]\Phi[/math], observe first that we have:
We have the following computation, by using the formula [math]\mathbb E[\alpha_i\alpha_j]=\delta_{ij}[/math]:
We therefore obtain the following formula, for the above quantity:
Similarly, we have the following formula, for the last term:
Since in both the cases [math]i=j[/math] and [math]i=j+N/2[/math], when [math]N[/math] is even, we have [math]S_{ij}\in\{\pm 1\}[/math], the above two formulae are all that we need, and we obtain the following formula:
Now by summing over [math]i,j[/math], and taking into account as well the first term in the expression of [math]\Phi(U,B)[/math], computed above, we obtain the formula in the statement.
In the orthogonal case now, we have a similar result, also from [1], as follows:
If [math]U\in O_N[/math] is circulant, [math]U_{ij}=\gamma_{j-i}[/math], and symmetric, we have
As before, in the proof of Theorem 12.25, the expectation is taken with respect to the distribution of the eigenvalues [math]\alpha_0,\ldots,\alpha_{N-1}=\pm 1[/math] of the matrix [math]B[/math], which are now, in the present real case, subject to the following extra condition:
The first term in the expression of [math]\Phi(U,B)[/math] is then equal to [math]N \sum_i |\gamma_i|[/math]. For the second term in [math]\Phi[/math], we need the following covariance term, in the present real case:
Since all quantities are real in this case, we have the following formula:
We have then the following formula:
On the other hand, we have as well the following formula:
Now by putting everything together, gives the formula in the statement.
As an illustration for the above methods, we can now go back to the matrices in Theorem 12.20, and find a better proof for the fact that these matrices are not complex AHM. Indeed, we have the following result, which basically solves the problem:
With [math]U=\frac{1}{N}(2\mathbb I_N-N1_N)[/math] we have the formula
This follows indeed from the general formula in Theorem 12.26.
We can therefore recover Theorem 12.20, modulo a bit of extra work still needed at [math]N=5[/math]. Regarding the case [math]N=5[/math], here the above expectation vanishes, but by using Proposition 12.23 or Proposition 12.24, we conclude that the vanishing of the expectation must come from both positive and negative contributions, and we are done.
In fact, the above results can be used for excluding all the explicit examples of circulant AHM found in [2]. All these verifications suggest the following conjecture:
\begin{conjecture} For any [math]U\in O_N[/math] which is circulant and symmetric we have
where [math]B[/math] varies over the space of orthogonal circulant symmetric matrices. In addition, a similar result should hold in the unitary, circulant and self-adjoint case. \end{conjecture} This looks like a subtle Fourier analysis question. In fact, the main idea that emerges from the computations in [1], including the block design ones, is that of using a random derivative, pointing towards a suitable homogeneous space coset. However, no one really knows how to do that. And so we will have it as an exercise for you, reader.
General references
Banica, Teo (2024). "Invitation to Hadamard matrices". arXiv:1910.06911 [math.CO].
References
- 1.00 1.01 1.02 1.03 1.04 1.05 1.06 1.07 1.08 1.09 1.10 T. Banica and I. Nechita, Almost Hadamard matrices with complex entries, Adv. Oper. Theory 3 (2018), 149--189.
- 2.0 2.1 2.2 2.3 2.4 T. Banica, I. Nechita and K. \.Zyczkowski, Almost Hadamard matrices: general theory and examples, Open Syst. Inf. Dyn. 19 (2012), 1--26.
- W. Tadej and K. \.Zyczkowski, Defect of a unitary matrix, Linear Algebra Appl. 429 (2008), 447--481.