8a. Free probability

[math] \newcommand{\mathds}{\mathbb}[/math]

This article was automatically generated from a tex file and may contain conversion errors. If permitted, you may login and edit this article to improve the conversion.

We have seen in chapters 5-6 that the easiness property of [math]O_N,U_N[/math] and [math]O_N^+,U_N^+[/math] leads to a number of interesting probabilistic consequences, notably in what concerns the computation of the law of the main character [math]\chi[/math], in the [math]N\to\infty[/math] limit. Our purpose here will be two-fold. On one hand, we would like to have similar results for the various quantum groups introduced in chapter 7. And on the other hand, we would like to upgrade our results about characters [math]\chi[/math] into something more advanced.


In order to do this, we are in need of more probability knowledge. We have certainly met some classical and free probability in chapters 5-6, but the computations and results there were a bit ad-hoc, adapted to what we wanted to prove about [math]O_N,U_N[/math] and [math]O_N^+,U_N^+[/math]. And this kind of ad-hoc point of view will not do it, for what we want to do here.


In short, time for a crash course on probability. We will be following the book of Voiculescu, Dykema, Nica [1], and we will be quite brief, because at the core of classical and free probability are the Gaussian laws, real and complex, and the semicircular and circular laws, that we already know about, from chapters 5-6. Thus, our job will be basically that of putting what we know in a more conceptual framework.


Let us first talk about classical probability. The starting point here is:

Definition

Let [math]X[/math] be a probability space.

  • The real functions [math]f\in L^\infty(X)[/math] are called random variables.
  • The moments of such a variable are the numbers [math]M_k(f)=\mathbb E(f^k)[/math].
  • The law of such a variable is the measure given by [math]M_k(f)=\int_\mathbb Rx^kd\mu_f(x)[/math].

Here the fact that [math]\mu_f[/math] exists indeed is not trivial. By linearity, we would like to have a real probability measure making hold the following formula, for any [math]P\in\mathbb R[X][/math]:

[[math]] \mathbb E(P(f))=\int_\mathbb RP(x)d\mu_f(x) [[/math]]


By using a continuity argument, it is enough to have this for the characteristic functions [math]\chi_I[/math] of the measurable sets [math]I\subset\mathbb R[/math]. Thus, we would like to have [math]\mu_f[/math] such that:

[[math]] \mathbb P(f\in I)=\mu_f(I) [[/math]]


But this latter formula can serve as a definition for [math]\mu_f[/math], so we are done. Next in line, we need to talk about independence. Once again with the idea of doing things a bit abstractly, and most adapted to what we want to do here, the definition is as follows:

Definition

Two variables [math]f,g\in L^\infty(X)[/math] are called independent when

[[math]] \mathbb E(f^kg^l)=\mathbb E(f^k)\cdot\mathbb E(g^l) [[/math]]
happens, for any [math]k,l\in\mathbb N[/math].

Again, this definition hides some non-trivial things. Indeed, by linearity, we would like to have a formula as follows, valid for any polynomials [math]P,Q\in\mathbb R[X][/math]:

[[math]] \mathbb E(P(f)Q(g))=\mathbb E(P(f))\cdot\mathbb E(Q(g)) [[/math]]


By continuity, it is enough to have this for characteristic functions of type [math]\chi_I,\chi_J[/math], with [math]I,J\subset\mathbb R[/math]. Thus, we are led to the usual definition of independence, namely:

[[math]] \mathbb P(f\in I,g\in J)=\mathbb P(f\in I)\cdot\mathbb P(g\in J) [[/math]]


Here is now our first result, providing tools for the study of the independence:

Theorem

Assume that [math]f,g\in L^\infty(X)[/math] are independent.

  • We have [math]\mu_{f+g}=\mu_f*\mu_g[/math], where [math]*[/math] is the convolution of measures.
  • We have [math]F_{f+g}=F_fF_g[/math], where [math]F_f(x)=\mathbb E(e^{ixf})[/math] is the Fourier transform.


Show Proof

This is something very standard, the idea being as follows:


(1) We have the following computation, using the independence of [math]f,g[/math]:

[[math]] \begin{eqnarray*} M_k(f+g) &=&\mathbb E((f+g)^k)\\ &=&\sum_l\binom{k}{l}\mathbb E(f^lg^{k-l})\\ &=&\sum_l\binom{k}{l}M_l(f)M_{k-l}(g) \end{eqnarray*} [[/math]]


On the other hand, by using the Fubini theorem, we have as well:

[[math]] \begin{eqnarray*} \int_\mathbb Rx^kd(\mu_f*\mu_g)(x) &=&\int_{\mathbb R\times\mathbb R}(x+y)^kd\mu_f(x)d\mu_g(y)\\ &=&\sum_l\binom{k}{l}\int_\mathbb Rx^kd\mu_f(x)\int_\mathbb Ry^ld\mu_g(y)\\ &=&\sum_l\binom{k}{l}M_l(f)M_{k-l}(g) \end{eqnarray*} [[/math]]


Thus the measures [math]\mu_{f+g}[/math] and [math]\mu_f*\mu_g[/math] have the same moments, and so coincide.


(2) We have indeed the following computation, using (1) and Fubini:

[[math]] \begin{eqnarray*} F_{f+g}(x) &=&\int_\mathbb Re^{ixy}d\mu_{f+g}(y)\\ &=&\int_{\mathbb R\times\mathbb R}e^{ix(y+z)}d\mu_f(y)d\mu_g(z)\\ &=&\int_\mathbb Re^{ixy}d\mu_f(y)\int_\mathbb Re^{ixz}d\mu_g(z)\\ &=&F_f(x)F_g(x) \end{eqnarray*} [[/math]]


Thus, we are led to the conclusion in the statement.

Let us discuss now the normal distributions. We have here:

Definition

The normal law of parameter [math]t \gt 0[/math] is the following measure:

[[math]] g_t=\frac{1}{\sqrt{2\pi t}}\,e^{-x^2/2t}dx [[/math]]
This is also called Gaussian distribution, with “g” standing for Gauss.

As a first remark, the above law has indeed mass 1, as it should. This follows indeed from the Gauss formula, which gives, with [math]x=y/\sqrt{2t}[/math]:

[[math]] \int_\mathbb R e^{-y^2/2t}dy=\sqrt{2\pi t} [[/math]]


Generally speaking, the normal laws appear as bit everywhere, in real life. The reasons behind this phenomenon come from the Central Limit Theorem (CLT), that we will explain in a moment, after developing the needed general theory. We first have:

Proposition

We have the following formula, for any [math]t \gt 0[/math]:

[[math]] F_{g_t}(x)=e^{-tx^2/2} [[/math]]
In particular, the normal laws satisfy [math]g_s*g_t=g_{s+t}[/math], for any [math]s,t \gt 0[/math].


Show Proof

The Fourier transform formula can be established as follows:

[[math]] \begin{eqnarray*} F_{g_t}(x) &=&\frac{1}{\sqrt{2\pi t}}\int_\mathbb Re^{-y^2/2t+ixy}dy\\ &=&\frac{1}{\sqrt{2\pi t}}\int_\mathbb Re^{-(y/\sqrt{2t}-\sqrt{t/2}ix)^2-tx^2/2}dy\\ &=&\frac{1}{\sqrt{\pi}}\int_\mathbb Re^{-z^2-tx^2/2}dz \end{eqnarray*} [[/math]]


As for the last assertion, this follows from the linearization result from Theorem 8.3 (2) above, because [math]\log F_{g_t}[/math] is linear in [math]t[/math].

We are now ready to state and prove the CLT, as follows:

Theorem (CLT)

Given random variables [math]f_1,f_2,f_3,\ldots\in L^\infty(X)[/math] which are i.i.d., centered, and with variance [math]t \gt 0[/math], we have, with [math]n\to\infty[/math], in moments,

[[math]] \frac{1}{\sqrt{n}}\sum_{i=1}^nf_i\sim g_t [[/math]]
where [math]g_t[/math] is the Gaussian law of parameter [math]t[/math].


Show Proof

We have the following formula for [math]F_f(x)=\mathbb E(e^{ixf})[/math], in terms of moments:

[[math]] F_f(x)=\sum_{k=0}^\infty\frac{i^kM_k(f)}{k!}\,x^k [[/math]]


Thus, the Fourier transform of the variable in the statement is:

[[math]] \begin{eqnarray*} F(x) &=&\left[F_f\left(\frac{x}{\sqrt{n}}\right)\right]^n\\ &=&\left[1-\frac{tx^2}{2n}+O(n^{-2})\right]^n\\ &\simeq&e^{-tx^2/2} \end{eqnarray*} [[/math]]


But this latter function being the Fourier transform of [math]g_t[/math], we obtain the result.

Let us record as well the complex version of the CLT. This is as follows:

Theorem (Complex CLT)

Given variables [math]f_1,f_2,f_3,\ldots\in L^\infty(X)[/math] whose real and imaginary parts are i.i.d., centered, and with variance [math]t \gt 0[/math], we have, with [math]n\to\infty[/math],

[[math]] \frac{1}{\sqrt{n}}\sum_{i=1}^nf_i\sim G_t [[/math]]
where [math]G_t[/math] is the complex Gaussian law of parameter [math]t[/math], appearing as the law of [math]\frac{1}{\sqrt{2}}(a+ib)[/math], where [math]a,b[/math] are real and independent, each following the law [math]g_t[/math].


Show Proof

This is clear from Theorem 8.6, by taking real and imaginary parts.

In the noncommutative setting now, the starting definition is as follows:

Definition

Let [math]A[/math] be a [math]C^*[/math]-algebra, given with a trace [math]tr[/math].

  • The elements [math]a\in A[/math] are called random variables.
  • The moments of such a variable are the numbers [math]M_k(a)=tr(a^k)[/math].
  • The law of such a variable is the functional [math]\mu:P\to tr(P(a))[/math].

Here [math]k=\circ\bullet\bullet\circ\ldots[/math] is as usual a colored integer, and the powers [math]a^k[/math] are defined by multiplicativity and the usual formulae, namely:

[[math]] a^\emptyset=1\quad,\quad a^\circ=a\quad,\quad a^\bullet=a^* [[/math]]


As for the polynomial [math]P[/math], this is a noncommuting [math]*[/math]-polynomial in one variable:

[[math]] P\in\mathbb C \lt X,X^* \gt [[/math]]


Observe that the law is uniquely determined by the moments, because:

[[math]] P(X)=\sum_k\lambda_kX^k\implies\mu(P)=\sum_k\lambda_kM_k(a) [[/math]]


Generally speaking, the above definition is something quite abstract, but there is no other way of doing things, at least at this level of generality. We have indeed:

Theorem

Given a [math]C^*[/math]-algebra with a faithful trace [math](A,tr)[/math], any normal variable,

[[math]] aa^*=a^*a [[/math]]
has a usual law, namely a complex probability measure [math]\mu\in\mathcal P(\mathbb C)[/math] satisfying:

[[math]] tr(a^k)=\int_\mathbb Cz^kd\mu(z) [[/math]]
This law is unique, and is supported by the spectrum [math]\sigma(a)\subset\mathbb C[/math]. In the non-normal case, [math]aa^*\neq a^*a[/math], such a usual law does not exist.


Show Proof

This is something that we know from chapter 6, coming from the Gelfand theorem, which gives [math] \lt a \gt =C(\sigma(a))[/math], and the Riesz theorem. As for the last assertion, we know this too from chapter 6, coming via [math]tr(aa^*aa^*) \gt tr(aaa^*a^*)[/math] for [math]aa^*\neq a^*a[/math].

Let us discuss now the independence, and its noncommutative versions, in the above setting. As a starting point here, we have the following notion:

Definition

Two subalgebras [math]B,C\subset A[/math] are called independent when the following condition is satisfied, for any [math]b\in B[/math] and [math]c\in C[/math]:

[[math]] tr(bc)=tr(b)tr(c) [[/math]]
Equivalently, the following condition must be satisfied, for any [math]b\in B[/math] and [math]c\in C[/math]:

[[math]] tr(b)=tr(c)=0\implies tr(bc)=0 [[/math]]
Also, two variables [math]b,c\in A[/math] are called independent when the algebras that they generate,

[[math]] B= \lt b \gt \quad,\quad C= \lt c \gt [[/math]]
are independent inside [math]A[/math], in the above sense.

Observe that the above two conditions are indeed equivalent. In one sense this is clear, and in the other sense, with [math]a'=a-tr(a)[/math], this follows from:

[[math]] \begin{eqnarray*} tr(bc) &=&tr[(b'+tr(b))(c'+tr(c))]\\ &=&tr(b'c')+t(b')tr(c)+tr(b)tr(c')+tr(b)tr(c)\\ &=&tr(b'c')+tr(b)tr(c)\\ &=&tr(b)tr(c) \end{eqnarray*} [[/math]]


The other remark is that the above notion generalizes indeed the usual notion of independence, from the classical case, the result here being as follows:

Theorem

Given two compact measured spaces [math]Y,Z[/math], the algebras

[[math]] C(Y)\subset C(Y\times Z)\quad,\quad C(Z)\subset C(Y\times Z) [[/math]]
are independent in the above sense, and a converse of this fact holds too.


Show Proof

We have two assertions here, the idea being as follows:


(1) First of all, given two arbitrary compact spaces [math]Y,Z[/math], we have embeddings of algebras as in the statement, defined by the following formulae:

[[math]] f\to[(y,z)\to f(y)]\quad,\quad g\to[(y,z)\to g(z)] [[/math]]


In the measured space case now, the Fubini theorems tells us that:

[[math]] \int_{Y\times Z}f(y)g(z)=\int_Yf(y)\int_Zg(z) [[/math]]


Thus, the algebras [math]C(Y),C(Z)[/math] are independent in the sense of Definition 8.3.


(2) Conversely now, assume that [math]B,C\subset A[/math] are independent, with [math]A[/math] being commutative. Let us write our algebras as follows, with [math]X,Y,Z[/math] being certain compact spaces:

[[math]] A=C(X)\quad,\quad B=C(Y)\quad,\quad C=C(Z) [[/math]]

In this picture, the inclusions [math]B,C\subset A[/math] must come from quotient maps, as follows:

[[math]] p:Z\to X\quad,\quad q:Z\to Y [[/math]]


Regarding now the independence condition from Definition 8.3, in the above picture, this tells us that the folowing equality must happen:

[[math]] \int_Xf(p(x))g(q(x))=\int_Xf(p(x))\int_Xg(q(x)) [[/math]]


Thus we are in a Fubini type situation, and we obtain from this [math]Y\times Z\subset X[/math]. Thus, the independence of [math]B,C\subset A[/math] appears as in (1) above.

It is possible to develop some theory here, but this is ultimately not very interesting. As a much more interesting notion now, we have Voiculescu's freeness [1]:

Definition

Two subalgebras [math]B,C\subset A[/math] are called free when the following condition is satisfied, for any [math]b_i\in B[/math] and [math]c_i\in C[/math]:

[[math]] tr(b_i)=tr(c_i)=0\implies tr(b_1c_1b_2c_2\ldots)=0 [[/math]]
Also, two variables [math]b,c\in A[/math] are called free when the algebras that they generate,

[[math]] B= \lt b \gt \quad,\quad C= \lt c \gt [[/math]]
are free inside [math]A[/math], in the above sense.

As a first observation, of theoretical nature, there is actually a certain lack of symmetry between Definition 8.10 and Definition 8.12, because in contrast to the former, the latter does not include an explicit formula for the quantities of the following type:

[[math]] tr(b_1c_1b_2c_2\ldots) [[/math]]


However, this is not an issue, and is simply due to the fact that the formula in the free case is something more complicated, the result being as follows:

Proposition

Assuming that [math]B,C\subset A[/math] are free, the restriction of [math]tr[/math] to [math] \lt B,C \gt [/math] can be computed in terms of the restrictions of [math]tr[/math] to [math]B,C[/math]. To be more precise,

[[math]] tr(b_1c_1b_2c_2\ldots)=P\Big(\{tr(b_{i_1}b_{i_2}\ldots)\}_i,\{tr(c_{j_1}c_{j_2}\ldots)\}_j\Big) [[/math]]
where [math]P[/math] is certain polynomial in several variables, depending on the length of the word [math]b_1c_1b_2c_2\ldots[/math], and having as variables the traces of products of type

[[math]] b_{i_1}b_{i_2}\ldots\quad,\quad c_{j_1}c_{j_2}\ldots [[/math]]
with the indices being chosen increasing, [math]i_1 \lt i_2 \lt \ldots[/math] and [math]j_1 \lt j_2 \lt \ldots[/math]


Show Proof

This is something that we know from chapter 6, which is based on a computation which is similar to that made after Definition 8.10.

Let us discuss now some models for independence and freeness. We first have:

Theorem

Given two algebras [math](B,tr)[/math] and [math](C,tr)[/math], the following hold:

  • [math]B,C[/math] are independent inside their tensor product [math]B\otimes C[/math], endowed with its canonical tensor product trace, given on basic tensors by [math]tr(b\otimes c)=tr(b)tr(c)[/math].
  • [math]B,C[/math] are free inside their free product [math]B*C[/math], endowed with its canonical free product trace, given by the formulae in Proposition 8.13.


Show Proof

Both the assertions are clear from definitions, as follows:


(1) This is clear with either of the definitions of the independence, from Definition 8.10 above, because we have by construction of the trace:

[[math]] tr(bc) =tr[(b\otimes1)(1\otimes c)] =tr(b\otimes c) =tr(b)tr(c) [[/math]]


(2) This is clear from definitions, the only point being that of showing that the notion of freeness, or the recurrence formulae in Proposition 8.13, can be used in order to construct a canonical free product trace, on the free product of the two algebras involved:

[[math]] tr:B*C\to\mathbb C [[/math]]


But this can be checked for instance by using a GNS construction. Indeed, consider the GNS constructions for the algebras [math](B,tr)[/math] and [math](C,tr)[/math]:

[[math]] B\to B(l^2(B))\quad,\quad C\to B(l^2(C)) [[/math]]


By taking the free product of these representations, we obtain a representation as follows, with the [math]*[/math] symbol on the right being a free product of pointed Hilbert spaces:

[[math]] B*C\to B(l^2(B)*l^2(C)) [[/math]]


Now by composing with the linear form [math]T\to \lt T\xi,\xi \gt [/math], where [math]\xi=1_B=1_C[/math] is the common distinguished vector of [math]l^2(B)[/math] and [math]l^2(C)[/math], we obtain a linear form, as follows:

[[math]] tr:B*C\to\mathbb C [[/math]]


It is routine then to check that [math]tr[/math] is indeed a trace, and this is the “canonical free product trace” from the statement. Then, an elementary computation shows that [math]B,C[/math] are indeed free inside [math]B*C[/math], with respect to this trace, and this finishes the proof.

As a concrete application of the above results, still following [1], we have:

Theorem

We have a free convolution operation [math]\boxplus[/math] for the distributions

[[math]] \mu:\mathbb C \lt X,X^* \gt \to\mathbb C [[/math]]
which is well-defined by the following formula, with [math]b,c[/math] taken to be free:

[[math]] \mu_b\boxplus\mu_c=\mu_{b+c} [[/math]]
This restricts to an operation, still denoted [math]\boxplus[/math], on the real probability measures.


Show Proof

We have several verifications to be performed here, as follows:


(1) We first have to check that given two variables [math]b,c[/math] which live respectively in certain [math]C^*[/math]-algebras [math]B,C[/math], we can recover them inside some [math]C^*[/math]-algebra [math]A[/math], with exactly the same distributions [math]\mu_b,\mu_c[/math], as to be able to sum them and then talk about [math]\mu_{b+c}[/math]. But this comes from Theorem 8.14, because we can set [math]A=B*C[/math], as explained there.


(2) The other verification which is needed is that of the fact that if [math]b,c[/math] are free, then the distribution [math]\mu_{b+c}[/math] depends only on the distributions [math]\mu_b,\mu_c[/math]. But for this purpose, we can use the general formula from Proposition 8.13, namely:

[[math]] tr(b_1c_1b_2c_2\ldots)=P\Big(\{tr(b_{i_1}b_{i_2}\ldots)\}_i,\{tr(c_{j_1}c_{j_2}\ldots)\}_j\Big) [[/math]]


Here [math]P[/math] is certain polynomial, depending on the length of [math]b_1c_1b_2c_2\ldots[/math], having as variables the traces of products [math]b_{i_1}b_{i_2}\ldots[/math] and [math]c_{j_1}c_{j_2}\ldots[/math], with [math]i_1 \lt i_2 \lt \ldots[/math] and [math]j_1 \lt j_2 \lt \ldots[/math] Now by plugging in arbitrary powers of [math]b,c[/math] as variables [math]b_i,c_j[/math], we obtain a family of formulae of the following type, with [math]Q[/math] being certain polynomials:

[[math]] tr(b^{k_1}c^{l_1}b^{k_2}c^{l_2}\ldots)=P\Big(\{tr(b^k)\}_k,\{tr(c^l)\}_l\Big) [[/math]]


Thus the moments of [math]b+c[/math] depend only on the moments of [math]b,c[/math], with of course colored exponents in all this, according to our moment conventions, and this gives the result.


(3) Finally, in what regards the last assertion, regarding the real measures, this is clear from the fact that if [math]b,c[/math] are self-adjoint, then so is their sum [math]b+c[/math].

We would like now to have a linearization result for [math]\boxplus[/math], in the spirit of the previous result for [math]*[/math]. We will do this slowly, in several steps. As a first observation, both independence and freeness are nicely modelled inside group algebras, as follows:

Theorem

We have the following results, valid for group algebras:

  • [math]C^*(\Gamma),C^*(\Lambda)[/math] are independent inside [math]C^*(\Gamma\times\Lambda)[/math].
  • [math]C^*(\Gamma),C^*(\Lambda)[/math] are free inside [math]C^*(\Gamma*\Lambda)[/math].


Show Proof

In order to prove these results, we have two possible methods:


(1) We can use here the general results in Theorem 8.14 above, along with the following two isomorphisms, which are both standard:

[[math]] C^*(\Gamma\times\Lambda)=C^*(\Lambda)\otimes C^*(\Gamma) [[/math]]

[[math]] C^*(\Gamma*\Lambda)=C^*(\Lambda)*C^*(\Gamma) [[/math]]



(2) We can prove this directly as well, by using the fact that each group algebra is spanned by the corresponding group elements. Indeed, it is enough to check the independence and freeness formulae on group elements, which is in turn trivial.

Regarding now the linearization problem for [math]\boxplus[/math], the situation here is quite tricky. We need good models for the pairs of free random variables [math](b,c)[/math], and the problem is that the models that we have will basically lead us into the combinatorics from Proposition 8.13 and its proof, that cannot be solved with bare hands, and that we want to avoid.


The idea will be that of temporarily lifting the self-adjointness assumption on our variables [math]b,c[/math], and looking instead for arbitrary random variables [math]\beta,\gamma[/math], not necessarily self-adjoint, modelling in integer moments our given laws [math]\mu,\nu\in\mathcal P(\mathbb R)[/math], as follows:

[[math]] tr(\beta^k)=M_k(\mu),\ \forall k\in\mathbb N [[/math]]

[[math]] tr(\gamma^k)=M_k(\nu),\ \forall k\in\mathbb N [[/math]]


To be more precise, assuming that [math]\beta,\gamma[/math] are indeed not self-adjoint, the above formulae are not the general formulae for [math]\beta,\gamma[/math], simply because these latter formulae involve colored integers [math]k=\circ\bullet\bullet\circ\ldots[/math] as exponents. Thus, in the context of the above formulae, [math]\mu,\nu[/math] are not the distributions of [math]\beta,\gamma[/math], but just some “pieces” of these distributions.


Now with this tricky idea in mind, due to Voiculescu [1], the solution to our law modelling problem comes in a quite straightforward way, involving the good old Hilbert space [math]H=l^2(\mathbb N)[/math] and the good old shift operator [math]S\in B(H)[/math], as follows:

Theorem

Consider the shift operator on the space [math]H=l^2(\mathbb N)[/math], given by:

[[math]] S(e_i)=e_{i+1} [[/math]]
The variables of the following type, with [math]f\in\mathbb C[X][/math] being a polynomial,

[[math]] S^*+f(S) [[/math]]
model then in moments, up to finite order, all the distributions [math]\mu:\mathbb C[X]\to\mathbb C[/math].


Show Proof

The adjoint of the shift is given by the following formula:

[[math]] S^*(e_i)=\begin{cases} e_{i-1}&(i \gt 0)\\ 0&(i=0) \end{cases} [[/math]]


Consider now a variable as in the statement, namely:

[[math]] T=S^*+a_0+a_1S+a_2S^2+\ldots+a_nS^n [[/math]]


We have then [math]tr(T)=a_0[/math], then [math]tr(T^2)[/math] will involve [math]a_1[/math], then [math]tr(T^3)[/math] will involve [math]a_2[/math], and so on. Thus, we are led to a certain recurrence, that we will not attempt to solve now, with bare hands, but which definitely gives the conclusion in the statement.

Before getting further, let us point out the following fundamental fact:

Proposition

In the context of the above correspondence, the variable

[[math]] T=S+S^* [[/math]]
follows the Wigner semicircle law on [math][-2,2][/math].


Show Proof

This is something that we know from chapter 6, the idea being that the combinatorics of [math](S+S^*)^k[/math] leads us into paths on [math]\mathbb N[/math], and to the Catalan numbers.

Getting back now to our linearization program for [math]\boxplus[/math], the next step is that of taking a free product of the model found in Theorem 8.17 with itself. We have here:

Proposition

We can define the algebra of creation operators

[[math]] S_x:v\to x\otimes v [[/math]]
on the free Fock space associated to a real Hilbert space [math]H[/math], given by

[[math]] F(H)=\mathbb C\Omega\oplus H\oplus H^{\otimes2}\oplus\ldots [[/math]]
and at the level of examples, we have:

  • With [math]H=\mathbb C[/math] we recover the shift algebra [math]A= \lt S \gt [/math] on [math]H=l^2(\mathbb N)[/math].
  • With [math]H=\mathbb C^2[/math], we obtain the algebra [math]A= \lt S_1,S_2 \gt [/math] on [math]H=l^2(\mathbb N*\mathbb N)[/math].


Show Proof

We can talk indeed about the algebra [math]A(H)[/math] of creation operators on the free Fock space [math]F(H)[/math] associated to a real Hilbert space [math]H[/math], with the remark that, in terms of the abstract semigroup notions from chapter 6 above, we have:

[[math]] A(\mathbb C^k)=C^*(\mathbb N^{*k})\quad,\quad F(\mathbb C^k)=l^2(\mathbb N^{*k}) [[/math]]


Thus, we are led to the conclusions in the statement.

With the above notions in hand, we have the following key freeness result:

Proposition

Given a real Hilbert space [math]H[/math], and two orthogonal vectors [math]x\perp y[/math], the corresponding creation operators [math]S_x[/math] and [math]S_y[/math] are free with respect to

[[math]] tr(T)= \lt T\Omega,\Omega \gt [[/math]]
called trace associated to the vacuum vector.


Show Proof

This is something that we know from chapter 6, coming from the formula [math]S_x^*S_y= \lt x,y \gt id[/math], valid for any two vectors [math]x,y\in H[/math], which itself is elementary.

With this technology in hand, let us go back to our linearization program for [math]\boxplus[/math]. We have the following key result, further building on Proposition 8.20:

Theorem

Given two polynomials [math]f,g\in\mathbb C[X][/math], consider the variables

[[math]] R^*+f(R)\quad,\quad S^*+g(S) [[/math]]
where [math]R,S[/math] are two creation operators, or shifts, associated to a pair of orthogonal norm [math]1[/math] vectors. These variables are then free, and their sum has the same law as

[[math]] T^*+(f+g)(T) [[/math]]
with [math]T[/math] being the usual shift on [math]l^2(\mathbb N)[/math].


Show Proof

Again, this is something that we know from chapter 6, the idea being that this comes from Proposition 8.20, by using a [math]45^\circ[/math] rotation trick.

We can now solve the linearization problem. Following Voiculescu [1], we have:

Theorem

Given a real probability measure [math]\mu[/math], define its [math]R[/math]-transform as follows:

[[math]] G_\mu(\xi)=\int_\mathbb R\frac{d\mu(t)}{\xi-t}\implies G_\mu\left(R_\mu(\xi)+\frac{1}{\xi}\right)=\xi [[/math]]
The free convolution operation is then linearized by this [math]R[/math]-transform.


Show Proof

This can be done by using the above results, in several steps, as follows:


(1) According to Theorem 8.21, the operation [math]\mu\to f[/math] from Theorem 8.10 linearizes the free convolution operation [math]\boxplus[/math]. We are therefore left with a computation inside [math]C^*(\mathbb N)[/math]. To be more precise, consider a variable as in Theorem 8.21:

[[math]] X=S^*+f(X) [[/math]]


In order to establish the result, we must prove that the [math]R[/math]-transform of [math]X[/math], constructed according to the procedure in the statement, is the function [math]f[/math] itself.


(2) In order to do so, fix [math]|z| \lt 1[/math] in the complex plane, and let us set:

[[math]] w_z=\delta_0+\sum_{k=1}^\infty z_k\delta_k [[/math]]


The shift and its adjoint act then as follows, on this vector:

[[math]] Sw_z=z^{-1}(w_z-\delta_0)\quad,\quad S^*w_z=zw_z [[/math]]


It follows that the adjoint of our operator [math]X[/math] acts as follows on this vector:

[[math]] \begin{eqnarray*} X^*w_z &=&(S+f(S^*))w_z\\ &=&z^{-1}(w_z-\delta_0)+f(z)w_z\\ &=&(z^{-1}+f(z))w_z-z^{-1}\delta_0 \end{eqnarray*} [[/math]]


Now observe that this formula can be written as follows:

[[math]] z^{-1}\delta_0=(z^{-1}+f(z)-X^*)w_z [[/math]]


The point now is that when [math]|z|[/math] is small, the operator appearing on the right is invertible. Thus, we can rewrite this formula as follows:

[[math]] (z^{-1}+f(z)-X^*)^{-1}\delta_0=zw_z [[/math]]


Now by applying the trace, we are led to the following formula:

[[math]] \begin{eqnarray*} tr\left[(z^{-1}+f(z)-X^*)^{-1}\right] &=&\left \lt (z^{-1}+f(z)-X^*)^{-1}\delta_0,\delta_0\right \gt \\ &=& \lt zw_z,\delta_0 \gt \\ &=&z \end{eqnarray*} [[/math]]


(3) Let us apply now the complex function procedure in the statement to the real probability measure [math]\mu[/math] modelled by [math]X[/math]. The Cauchy transform [math]G_\mu[/math] is given by:

[[math]] \begin{eqnarray*} G_\mu(\xi) &=&tr((\xi-X)^{-1})\\ &=&\overline{tr\big((\bar{\xi}-X^*)^{-1}\big)}\\ &=&tr((\xi-X^*)^{-1}) \end{eqnarray*} [[/math]]


Now observe that, with the choice [math]\xi=z^{-1}+f(z)[/math] for our complex variable, the trace formula found in (2) above tells us precisely that we have:

[[math]] G_\mu\big(z^{-1}+f(z)\big)=z [[/math]]


Thus, we have [math]R_\mu(z)=f(z)[/math], which finishes the proof, as explained in step (1).

With the above linearization technology in hand, we can now establish the following free analogue of the CLT, also due to Voiculescu [1]:

Theorem (Free CLT)

Given self-adjoint variables [math]x_1,x_2,x_3,\ldots[/math] which are f.i.d., centered, with variance [math]t \gt 0[/math], we have, with [math]n\to\infty[/math], in moments,

[[math]] \frac{1}{\sqrt{n}}\sum_{i=1}^nx_i\sim\gamma_t [[/math]]
\ where [math]\gamma_t[/math] is the Wigner semicircle law of parameter [math]t[/math], having density:

[[math]] \gamma_t=\frac{1}{2\pi t}\sqrt{4t^2-x^2}dx [[/math]]


Show Proof

We follow the same idea as in the proof of the CLT:


(1) At [math]t=1[/math], the [math]R[/math]-transform of the variable in the statement on the left can be computed by using the linearization property from Theorem 8.22, and is given by:

[[math]] R(\xi) =nR_x\left(\frac{\xi}{\sqrt{n}}\right) \simeq\xi [[/math]]


(2) Regarding now the right term, also at [math]t=1[/math], our claim is that the [math]R[/math]-transform of the Wigner semicircle law [math]\gamma_1[/math] is given by the following formula:

[[math]] R_{\gamma_1}(\xi)=\xi [[/math]]


But this follows via some calculus, or directly from the following formula, coming from Proposition 8.18, and from the technical details of the [math]R[/math]-transform:

[[math]] S+S^*\sim\gamma_1 [[/math]]


Thus, the laws in the statement have the same [math]R[/math]-transforms, and so they are equal.


(4) Summarizing, we have proved the free CLT at [math]t=1[/math]. The passage to the general case, [math]t \gt 0[/math], is routine, by some standard dilation computations.

Similarly, in the complex case, we have the following result, also from [1]:

Theorem (Free complex CLT)

Given variables [math]x_1,x_2,x_3,\ldots,[/math] whose real and imaginary parts are f.i.d., centered, and with variance [math]t \gt 0[/math], we have, with [math]n\to\infty[/math],

[[math]] \frac{1}{\sqrt{n}}\sum_{i=1}^nx_i\sim\Gamma_t [[/math]]
where [math]\Gamma_t[/math] is the Voiculescu circular law of parameter [math]t[/math], appearing as the law of [math]\frac{1}{\sqrt{2}}(a+ib)[/math], where [math]a,b[/math] are self-adjoint and free, each following the law [math]\gamma_t[/math].


Show Proof

This is clear from Theorem 8.23, by taking real and imaginary parts.

There are of course many other things that can be said about [math]g_t,\gamma_t,G_t,\Gamma_t[/math], but for the moment, this is all we need. We will be back later to these laws, with more details.

General references

Banica, Teo (2024). "Introduction to quantum groups". arXiv:1909.08152 [math.CO].

References

  1. 1.0 1.1 1.2 1.3 1.4 1.5 1.6 D.V. Voiculescu, K.J. Dykema and A. Nica, Free random variables, AMS (1992).