Free Probability Notes
Free Probability Notes
Preliminary version
These notes were used in a topics course Free probability and combinatorics taught at Texas A&M Uni-
versity in the fall of 2012. The course was based on Nica and Speicher’s textbook [NS06], however there
were sufficient differences to warrant separate notes. I tried to minimize the list of references, but many
more are included in the body of the text. Thanks to the students in the course, and to March Boedihardjo,
for numerous comments and corrections; further corrections or comments are welcome. Eventually I will
probably replace hand-drawn figures by typeset ones.
Contents
3 Free independence. 19
3.1 Independence. Free independence. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.2 Free products. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
1
6 Lattice paths and Fock space models. 54
6.1 Dyck paths and the full Fock space. . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
6.2 Łukasiewicz paths and Voiculescu’s operator model. . . . . . . . . . . . . . . . . . . . 62
6.3 Motzkin paths and Schürmann’s operator model. Freely infinitely divisible distributions. 63
6.4 Motzkin paths and orthogonal polynomials. . . . . . . . . . . . . . . . . . . . . . . . . 69
6.5 q-deformed free probability. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
9 Belinschi-Nica evolution. 85
2
Chapter 1
Then E (usually called the expectation) is a state, in particular positive and unital. Thus (A, E) is a
(commutative) probability space. Note: f = f ∗ means f real-valued.
3
A related construction is \
A = L∞− (X, P ) = Lp (X, P ),
p≥1
the space of complex-valued random variables all of whose moments are finite.
Example 1.4. Let X be a compact topological space (e.g. X = [0, 1]), and µ a Borel probability measure
on X. Then for Z
A = C(X), ϕ [f ] = f dµ,
X
(A, ϕ) is again a commutative probability space.
Example 1.5. Let A = C[x] (polynomials in x with complex coefficients), x∗ = x. Then any state ϕ on
C[x] gives a commutative probability space. Does such a state always come from a measure?
A = Chx1 , x2 , . . . , xd i = Chxi
Then any state ϕ on Chxi gives a non-commutative probability space. These do not come from measures.
One example: for z = (z1 , z2 , . . . , zd ) ∈ Rd ,
δz (f (x1 , x2 , . . . , xd )) = f (z1 , z2 , . . . , zd )
the normalized trace of A. Note that indeed, tr(A∗ ) = tr(A) and tr(A∗ A) ≥ 0.
Definition 1.8. Let ϕ be a state on A.
ϕ [ab] = ϕ [ba] .
4
b. ϕ is faithful if ϕ [a∗ a] = 0 only if a = 0.
These are random matrices = matrix-valued random variables = matrices with random entries. Take
Z
ϕ [A] = (tr ⊗ E)(A) = tr(A) dP.
X
Example 1.10. Let H be a Hilbert space, and A a ∗-subalgebra of B(H), the algebra of bounded linear
operators on H. a∗ is the adjoint operator to a. If ξ ∈ H is a unit vector, then
ϕ [a] = haξ, ξi
in other words X X
(f g)(z) = f (x)g(y) = f (x)g(x−1 z),
z=xy x∈Γ
f ∗ (x) = f (x−1 ).
5
Check that indeed, (f g)∗ = g ∗ f ∗ .
So C[Γ] is a unital ∗-algebra, with the unit δe , where for x ∈ Γ
(
1, y = x,
δx (y) =
0, y 6= x
Exercise 1.12. Prove that τ is a faithful, tracial state, called the von Neumann trace.
are the moments of a. In particular m0 [ϕa ] = 1 and m1 [ϕa ] = ϕ [a] is the mean of a.
b. More generally, let a1 , a2 , . . . , ad ∈ (A, ϕ) be symmetric. Their joint distribution is the state
The numbers
ϕ au(1) au(2) . . . au(n) : n ≥ 0, 1 ≤ u(i) ≤ d
are the joint moments of a1 , . . . , ad .
Denote by D(d) the space of all joint distributions of d-tuples of symmetric random variables, which
is the space of all states on Chx1 , . . . , xd i.
6
c. We say that
(N ) (N )
(a1 , . . . , ad ) → (a1 , . . . , ad )
in moments (or, for d > 1, in distribution) if for each p,
as N → ∞.
Remark 1.14. Each µ ∈ D(d) can be realized as a distribution of a d-tuple of random variables, namely
(x1 , x2 , . . . , xd ) ⊂ (Chxi, µ).
Definition 1.15. a1 , a2 , . . . , ad ∈ (A, ϕ) are singleton independent if
ϕ au(1) au(2) . . . au(n) = 0
whenever all ai are centered (that is, ϕ [ai ] = 0) and some index in ~u appears only once.
Proposition 1.16 (Weak law of large numbers). Suppose {an : n ∈ N} ⊂ (A, ϕ) are singleton indepen-
dent, identically distributed (that is, all ϕai are the same), and uniformly bounded, in the sense that for a
fixed C and all ~u,
ϕ au(1) au(2) . . . au(n) ≤ C n .
Denote
1
sn =
(a1 + a2 + . . . + an ).
n
Then sn → ϕ [a1 ] in moments (sn converges to the mean, a scalar).
7
n(n − 1)(n − 2) . . . (n − r + 1) ≤ nr ≤ nk/2
because of the singleton condition. Each term in the sum bounded by
Thus
1
ϕ xkn ≤ k B(k)nk/2 C k → 0
n
as n → ∞ for any k > 0.
8
Chapter 2
ka∗ ak = kak2 .
A C ∗ -probability space is a pair (A, ϕ), where A is a C ∗ -algebra and ϕ is a state on it continuous in the
norm topology. (It follows from Corollary 2.13(c) that continuity is actually automatic).
Example 2.2. If X is a compact Hausdorff space, the algebra of continuous functions on it C(X) is a C ∗ -
algebra (with the uniform norm). States on C(X) come from integration with respect to Borel probability
measures.
Theorem 2.3 (Gelfand-Naimark theorem). Any unital, commutative C ∗ -algebra is of the form in the
preceding example.
Example 2.4. If A is a ∗-subalgebra of B(H) closed in the norm topology, then A is a C ∗ -algebra. If
ξ ∈ H is a unit vector, then ϕ = h·ξ, ξi is a state on A.
Theorem 2.5. Any abstract C ∗ -algebra is a concrete C ∗ -algebra, that is, it has a representation as in the
preceding example.
Definition 2.6. A ∗-subalgebra A ⊂ B(H) is a von Neumann algebra (or a W ∗ -algebra) if A is closed in
the weak operator topology. A W ∗ -probability space is a pair (A, ϕ), where A is a W ∗ -algebra and ϕ is
a normal state (continuous in the ultraweak operator topology).
9
Example 2.7. Let X be a compact Hausdorff space, and µ be a Borel probability measure on X. Then
(C(X), µ) is a C ∗ -probability space. But C(X) ⊂ B(L2 (X, µ)) is not WOT-closed. However, L∞ (X, µ)
is so closed, and therefore is a von Neumann algebra. In fact, the WOT on L∞ (X, µ) is the weak-∗
topology on L∞ = (L1 )∗ . Normal states on L∞ (X, µ) come from integration with respect to f dµ,
f ∈ L1 (X, µ).
a. kak2 = r(a∗ a). Thus norm (topology) is determined by the spectrum (algebra).
Definition 2.10. In an (algebraic) n.c. probability space A, say a ≥ 0 if ∃b1 , . . . , bk such that
k
X
a= b∗i bi .
i=1
Remark 2.11. Note that in the algebra C[x1 , x2 , . . . , xd ] of multivariate polynomials in commuting vari-
ables, we can have p(x) ≥ 0 for all x without being able to write p as a sum of squares.
Theorem 2.12 (Spectral theorem and continuous functional calculus, bounded symmetric case). Let a =
a∗ ∈ B(H). Then the C ∗ -algebra generated by a
(isometric C ∗ -isomorphism). Moreover, this defines a map f 7→ f (a) ∈ B(H), so that f (a) is a well-
defined operator for any continuous f .
10
c. For any a ∈ A, |ϕ [a]| ≤ kak.
d. Assume ϕ is faithful. Then for any a ∈ A, kak = limn→∞ (ϕ [(a∗ a)n ])1/2n . Thus the norm can be
computed by using moments.
Proof. For (a), using the identification in the spectral theorem, in C(σ(a)), a corresponds to the identity
function (f (x) = x), kak = kf ku , and kak − a corresponds to kf ku − f , which is positive.
For (b), using part (a) we can write kak − a = c∗ c, from which
If a is symmetric, (c) follows from (b) (with b = 1). In general, Cauchy-Schwartz inequality below
implies p p
|ϕ [a]| ≤ ϕ [a∗ a] ϕ [1∗ 1] ≤ ka∗ ak = kak .
Finally, applying the spectral theorem to a symmetric element a∗ a, the last statement follows from that
fact that for a finite measure µ, limn→∞ kf k2n,µ = kf k∞,µ and, if µ has full support, kf k∞,µ = kf ku .
λ(a)b = ab.
Note that
hλ(a∗ )b, ciϕ = ϕ [c∗ a∗ b] = ϕ [(ac)∗ b] = hb, λ(a)ciϕ ,
so with respect to this inner produce, the adjoint of the operator λ(a) is λ(a∗ ). We conclude that
{λ(a) : a ∈ A} is a ∗-representation of A on V . Moreover,
11
in other words |ϕ [b∗ a]|2 ≤ ϕ [a∗ a] ϕ [b∗ b]. Let
n o
N = a ∈ A : ϕ [a∗ a] = kak2ϕ = 0 .
which is a Hilbert space. Denote H (or perhaps V ) by L2 (A, ϕ). We have a natural map a 7→ â = a + N
of A → Â ⊂ L2 (A, ϕ) with dense range.
Moreover, for any a ∈ A and b ∈ N
λ(a)b̂ = ab + aN = ab.
b
Thus have a (not necessarily faithful) ∗-representation of A on the Hilbert space H, by a priori unbounded
operators with a common dense domain  = V /N .
Corollary 2.15. Each µ ∈ D(d) can be realized as the joint distribution of (possibly unbounded) opera-
tors on a Hilbert space with respect to a vector state.
Proof. µ is the joint distribution of the operators λ(x1 ), . . . , λ(xd ) in the GNS representation of
(Chx1 , . . . , xd i, µ)
Finally,
ϕ [a] = λ(a)1̂, 1̂ ϕ
.
12
Corollary 2.17. If (A, ϕ) is a C ∗ -probability space and ϕ is faithful, may assume A ⊂ B(H), ϕ = h·ξ, ξi.
In addition, ξ is cyclic, that is
k·k
Aξ = H.
w
Moreover, in this case (A , h·ξ, ξi) is a W ∗ -probability space.
How is the preceding statement related to Theorem 2.5? Does it give a complete proof for it?
The following result is inspired by Proposition 1.2 in [PV10].
Proposition 2.18. Let µ be a state on Chx1 , . . . , xd i such that for a fixed C and all ~u,
Then µ can be realized as a joint distribution of a d-tuple of bounded operators on a Hilbert space.
Proof. The construction of the Hilbert space H is as in the GNS construction I. We need to show that
the representation is by bounded operators. It suffices to show that each λ(xi ) is bounded. Define the
“non-commutative ball of radius C” to be the space of formal power series
( )
X X
BC hxi = α~u x~u : α~u ∈ C, |α~u | C |~u| < ∞ .
~
u ~
u
13
2.3 Other topics.
Remark 2.20 (Group algebras).
( )
X
L2 (C[Γ], τ ) = L2 (Γ) = f : Γ → C, |f (x)|2 < ∞ .
x∈Γ
(there is also a full C ∗ -algebra C ∗ (Γ)). The group von Neumann algebra is
weak
L(Γ) = W ∗ (Γ) = C[Γ] ⊂ B(L2 (Γ)).
The vector state τ = h·δe , δe i is the extension of the von Neumann trace, which is still faithful and tracial
on Cr∗ (Γ) and L(Γ).
F2 6' F3
Remark 2.22 (Distributions and moments). Let a be a symmetric, bounded operator. C ∗ (a) ' C(σ(a)),
σ(a) compact. So (continuous) states on C ∗ (a) correspond to Borel probability measures on σ(a). In
14
particular, if a is a symmetric, bounded operator in (A, ϕ), there exists a Borel probability measure on R
such that for any f ∈ C(σ(a)), Z
ϕ [f (a)] = f dµa .
µa is the distribution of a (with respect to ϕ). Note that µa is supported on σ(a), so in particular (for a
bounded operator) compactly supported. Note also that for a polynomial p,
Z
ϕa [p(x)] = p(x) dµa (x).
By the Weierstrass theorem and continuity of ϕ, ϕa determines µa . On the other hand, if ϕ is faithful on
C ∗ (a), for a = a∗
kak = sup {|z| : z ∈ supp(µa )} .
Exercise 2.23. Let A ∈ (Mn (C), tr) be a Hermitian matrix, with real eigenvalues λ1 , . . . , λn . Compute
the algebraic distribution ϕA and the analytic distribution µA of A with respect to tr.
Remark 2.24 (Generating functions).
Z
n
mn = ϕ [a ] = xn dµa (x)
15
Example 2.25. Let Γ = F1 = Z, with a single generator x. In (C[Z], τ ), have the generating element
u = δx . Note that
u∗ = δx−1 = u−1 ,
so u is a unitary. Moreover,
τ [un ] = δn=0 , n ∈ Z,
which by definition says that u is a Haar unitary.
What is the distribution of the symmetric operator u + u∗ ? Moments:
τ (u + u−1 )n =?
Number of walks with n steps, starting and ending at zero, with steps of length 1 to the right or to the left.
−1 2n+1
−1 2n
2n
τ (u + u ) = 0, τ (u + u ) = .
n
Cauchy transform of the distribution is
∞ ∞ ∞
2n −2n−1 X (2n)! −2n−1 X 2n 1 · 3 · . . . · (2n − 1) −2n−1
X
G(z) = z = z = z
n=0
n n=0
n!n! n=0
n!
∞
4n 1
X 3 2n − 1
= ... z −2n−1
n=0
n! 2 2 2
∞
X (−4)n
1 3 2n − 1 −1/2 1
= − − ... − z −2n−1 = z −1 1 − 4z −2 =√ .
n=0
n! 2 2 2 z 2−4
Unbounded operators.
Remark 2.27 (Commutative setting). Let (X, Σ, P ) be a probability space.
A = L∞ (X, µ).
f ∈ Ae ⇔ ∀g ∈ Cb (C, C), g ◦ f ∈ A.
16
Remark 2.28. On a Hilbert space H, an unbounded operator a is defined only on a (dense) subspace
D(a). Since may have D(a) ∩ D(b) = {0}, cannot in general define a + b.
For a von Neumann algebra A, an unbounded, self-adjoint operator a is affiliated to A, if f (a) ∈ A for
all bounded continuous f . A general operator T is affiliated to A, T ∈ A, e if in its polar decomposition
T = ua, the partial isometry u ∈ A and the positive operator a is affiliated to A.
If (A, ϕ) is a W ∗ -probability space with ϕ a faithful, tracial state (so that A is a finite von Neumann
algebra), then Ae form an algebra.
Moment problem.
If a is an unbounded, self-adjoint operator, there exists a probability measure µa on R, not necessarily
compactly supported, such that Z
ϕ [f (a)] = f dµa
R
R
for f ∈ Cb (R). Moments ϕ [an ] = R xn dµa (x) need not be finite, so ϕa may be undefined. Also, can
have non-uniqueness in the moment problem: µ 6= ν with
Z Z
p(x) dµ = p(x) dν ∀p ∈ C[x].
R
Proposition 2.30. If all µak , µa compactly supported (in particular, if ak , a are bounded), then
ak → a in moments ⇔ ak → a in distribution.
More generally, if ak → a in moments and µa is determined by its moments, then ak → a is distribution.
Tensor products.
Remark 2.31 (Vector spaces). Let V1 , . . . , Vn be vector spaces. Their algebraic tensor product is
( k )
X (i) (i)
V1 ⊗ V2 ⊗ . . . ⊗ Vn = a1 ⊗ . . . ⊗ an(i) , k ≥ 0, aj ∈ Vj /linearity relations
i=1
17
Remark 2.32 (Algebras). Let A1 , . . . , An be algebras. Their algebraic tensor product A1 ⊗ . . . ⊗ An has
the vector space structure from the preceding remark and algebra structure
(a1 ⊗ . . . ⊗ an )(b1 ⊗ . . . ⊗ bn ) = a1 b1 ⊗ a2 b2 ⊗ . . . ⊗ an bn .
via
ai 7→ 1 ⊗ . . . ⊗ ai ⊗ . . . ⊗ 1.
Note that the images of Ai , Aj commute:
Also have tensor products of Hilbert spaces, C ∗ , von Neumann algebras. Require the taking of closures,
which in some cases are not unique.
Remark 2.33 (Probability spaces). Let (A1 , ϕ1 ), . . . , (An , ϕn ) be n.c. probability spaces. The tensor
product state
O n
ϕ= ϕi
i=1
Nn
on i=1 Ai is defined via the linear extension of
18
Chapter 3
Free independence.
ϕ [a1 a2 a1 ] = 0,
19
but
ϕ [a1 a2 a1 a2 ]
is not determined.
Want an independence-type rule for joint distributions of non-commuting variables.
Definition 3.4 (Voiculescu). Let (A, ϕ) be an n.c. probability space.
a. Subalgebras A1 , A2 , . . . , Ak ⊂ (A, ϕ) are freely independent (or free) with respect to ϕ if whenever
b. Elements a1 , a2 , . . . , ak are freely independent if the ∗-subalgebras they generate are freely inde-
pendent.
Proposition 3.5. Let Fn be the free group with generators x1 , x2 , . . . , xn , and consider the n.c. prob-
ability space (C[Fn ], τ ). Then with respect to τ , the Haar unitaries λ(x1 ), λ(x2 ), . . . , λ(xn ) are freely
independent.
whenever
τ Pj (λ(x), λ(x−1 )) = 0
(3.1)
and
u(1) 6= u(2) 6= u(3) 6= . . . 6= u(n). (3.2)
Note first that X (j)
Pj (λ(x), λ(x−1 )) = αk λ(xk ),
k∈Z
(j)
and so equation (3.1) implies that each α0 = 0. Moreover,
n
Y X
(1) (n) k(1) k(n)
Pj λ(xu(j) ), λ(x−1
u(j) ) = αk(1) . . . αk(n) λ(xu(1) . . . xu(n) ),
j=1 ~k
is reduced, has no cancellations, and in particular never equals e. It follows that τ applied to any of the
terms in the sum is zero.
20
Remark 3.6. Pairwise free does not imply free. For example, in (C[F2 ], τ ), let
Then the elements in each pair {a, b}, {a, c}, {b, c} are free, but the triple {a, b, c} is not free.
Remark 3.7. If A1 , . . . , An ⊂ (A, ϕ) ⊂ B(H) are star-subalgebras which are free, then
C ∗ (A1 ), . . . , C ∗ (An ) are free, and W ∗ (A1 ), . . . , W ∗ (An ) are free.
Claim 3.8. Suppose know ϕa1 , . . . , ϕan and a1 , . . . , an are free. Then ϕa1 ,...,an is determined.
Then
ϕ [abab] = ϕ [(a◦ + ϕ [a])(b◦ + ϕ [b])(a◦ + ϕ [a])(b◦ + ϕ [b])] .
Using freeness and linearity, this reduces to
Alg∗ (a1 , a2 , . . . , an )
∞
[ [
b1 b2 . . . bk : bi ∈ Alg∗ au(i) , ϕ [bi ] = 0 .
= C ⊕ Span
k=1 u(1)6=u(2)6=...6=u(k)
Exercise 3.10. Exercise 5.25 in [NS06]. In this exercise we prove that free independence behaves well
under successive
n decompositions
o and thus is associative. Let {Ai : i ∈ I} be unital ∗-subalgebras of
(j)
(A, ϕ), and Bi : j ∈ J(i) be unital ∗-subalgebras of Ai . Then we have the following.
n o
(j)
a. If {Ai : i ∈ I} are freely independent in (A, ϕ) and for each i ∈ I, Bi : j ∈ J(i) are freely
n o
(j)
independent in (Ai , ϕ|Ai ), then all Bi : i ∈ I, j ∈ J(i) are freely independent in (A, ϕ).
21
n o
(j)
b. If all Bi : i ∈ I, j ∈ J(i) are freely independent in (A, ϕ) and if, for each i ∈ I, Ai is the
n o
(j)
∗-algebra generated by Bi : j ∈ J(i) , then {Ai : i ∈ I} are freely independent in (A, ϕ).
Exercise 3.11. Let (A, τ ) be a tracial n.c. probability space, a1 , a2 , . . . , an free, and u a unitary. Then
{u∗ ai u, 1 ≤ i ≤ n} are free. Is the conclusion still true if τ is not assumed to be a trace?
Exercise 3.12. Exercise 5.24 in [NS06]. Let (A, ϕ) be a n.c. probability space. Consider a unital ∗-
subalgebra B ⊂ A and a Haar unitary u ∈ A freely independent from B. Show that then also B and u∗ Bu
are free.
Groups.
Γ1 , Γ2 , . . . , Γn groups with units ei ∈ Γi .
Any word x1 x2 . . . xk with all xi ∈ ni=1 Γi can be reduced by identifying
S
x1 x2 . . . xi−1 ej xi . . . xk = x1 x2 . . . xi−1 xi . . . xk
and
x1 x2 . . . xi−1 yzxi . . . xk = x1 x2 . . . xi−1 (yz)xi . . . xk
if y, z ∈ Γi for the same i. In this way any word can be reduced to a unique (!) reduced word: unit e or
x1 x2 . . . xk , xi ∈ Γu(i) \ eu(i) , u(1) 6= u(2) . . . 6= u(k).
and product
(x1 x2 . . . xk )(y1 y2 . . . yl ) = reduced version of (x1 . . . xk y1 . . . yl ).
22
Reduced free product of n.c. probability spaces.
Let (A1 , ϕ1 ), . . . , (An , ϕn ) be n.c. probability spaces. Denote
A◦i = {a ∈ Ai : ϕi [a] = 0} ,
W∅ = C = C1, and for u(1) 6= u(2) . . . 6= u(k), define the vector space
Let ∞ ∞
M M M M
A= W~u = C1 ⊕ W~u
k=0 |~
u|=k k=1 u(1)6=u(2)...6=u(k)
u(1)6=u(2)...6=u(k)
Next, let ai ∈ A◦u(i) , bi ∈ A◦v(i) , where ~u and ~v are alternating. If u(1) 6= v(1),
(ak . . . a2 a1 )(b1 b2 . . . bl ) = ak . . . a2 a1 b1 b2 . . . bl .
In general,
where j ≤ min(k, l) is the first index (if it exists) such that u(j) 6= v(j).
On A, define the free product state
ϕ = ∗ni=1 ϕi
by: ϕ [1] = 1 and on an alternating word in centered elements, ϕ [a1 a2 . . . ak ] = 0. We will prove later ϕ
is positive; other properties are clear.
Remark 3.13. This is a reduced free product, since we identify the units of all Ai with the unit 1 ∈ A.
Thus this is a product with amalgamation over C. There is also a full free product construction.
Remark 3.14. If we identify
23
Remark 3.15.
∗ni=1 C[Γi ] = C [∗ni=1 Γi ] .
Proposition 3.16. If each ϕi is tracial, so is their free product ϕ.
Corollary 3.17. Free product of commutative algebras is tracial. In particular, free product of one-
dimensional distributions is tracial.
Proof of the Proposition. Let ai ∈ A◦u(i) , bi ∈ A◦v(i) , u(1) 6= u(2) 6= . . . 6= u(k), v(1) 6= v(2) 6= . . . 6=
v(l). Suppose
u(1) = v(1), u(2) = v(2), . . . u(j) = v(j), u(j + 1) 6= v(j + 1).
Then
ϕ [ak . . . a1 b1 . . . bl ] = ϕ [ak . . . a2 (a1 b1 )◦ b2 . . . bl ] + ϕu(1) [a1 b1 ] ϕ [ak . . . a3 (a2 b2 )◦ b3 . . . bl ] + . . .
+ ϕu(1) [a1 b1 ] ϕu(2) [a2 b2 ] . . . ϕu(j−1) [aj−1 bj−1 ] ϕ [ak . . . aj bj . . . bl ]
This is zero unless j = k = l, in which case, since each ϕi is tracial,
ϕ [ak . . . a1 b1 . . . bk ] = ϕu(1) [a1 b1 ] ϕu(2) [a2 b2 ] . . . ϕu(j) [aj bj ]
= ϕu(1) [b1 a1 ] ϕu(2) [b2 a2 ] . . . ϕu(j) [bj aj ] = ϕ [b1 . . . bk ak . . . a1 ] .
This implies that ϕ has the trace property for general ai , bi (why?).
Remark 3.18 (Reduced free product of Hilbert spaces with distinguished vectors). Given Hilbert spaces
with distinguished vectors (H1 , ξ1 )ni=1 , define their reduced free product
(H, ξ) = ∗ni=1 (Hi , ξi )
as follows. Denote Hi◦ = Hi Cξi . Then
∞
M M
◦ ◦ ◦
Halg = Cξ ⊕ H~u = Hu(1) ⊗ Hu(2) ⊗ . . . ⊗ Hu(k) .
k=1 |~
u|=k
u(1)6=u(2)...6=u(k)
H is the completion of Halg with respect to the inner product for which H~u ⊥ H~v for ~u 6= ~v , and on each
H~u we use the usual tensor inner product
k
Y
hf1 ⊗ f2 ⊗ . . . ⊗ fk , g1 ⊗ g2 ⊗ . . . ⊗ gk i = hfi , gi iu(i)
i=1
◦
for fi , gi ∈ Hu(i) .
If (Hi , ξi ) = L2 (Ai , ϕ1 ), then, at least in the faithful case,
(H, ξ) = L2 (∗ni=1 (Ai , ϕi )) .
More generally, if each Ai is represented on Hi with ϕi = h·ξi , ξi i, then can represent ∗ni=1 (Ai , ϕi ) on H
so that ϕ = h·ξ, ξi. This implies in particular that ϕ is positive.
By using this representation and taking appropriate closures, can define reduced free products of C ∗ and
W ∗ -probability spaces.
24
Positivity.
How to prove positivity of a joint distribution? Represent it as a joint distribution of symmetric operators
on a Hilbert space.
How to prove positivity of the free product state? One way is to use the preceding remark. We will use a
different, more algebraic proof.
Remark 3.19. A matrix A ∈ Mn (C) is positive if and only if one of the following equivalent conditions
holds:
Proposition 3.20. A linear functional ϕ on A is positive if and only if for all n and all
a1 , a2 , . . . , an ∈ A, the (numerical) matrix [ϕ(a∗i aj )]ni,j=1 is positive.
Proof. ⇐ clear by taking n = 1. For the converse, let a1 , . . . , an ∈ A and z1 , . . . zn ∈ C. Then since ϕ is
positive, " n !∗ n #
X X X n
0≤ϕ zi ai zj aj = z i ϕ [a∗i aj ] zj .
i=1 j=1 i,j=1
Definition 3.21. Let A be a ∗-algebra. Then Mn (A) = Mn (C) ⊗ A is also a ∗-algebra, and so has a
notion of positivity. If T : A → B, define
Remark 3.22. Usually defined only for C ∗ -algebras. Even in that case, positive does not imply com-
pletely positive.
Proposition 3.23. If (A, ϕ) is an n.c. probability space, so that ϕ is positive, then each ϕn : Mn (A) →
Mn (C) is positive, so that ϕ is completely positive.
25
Proof. Let A ∈ Mn (A) be positive. By definition, that means A = N ∗
P
i=1 Bi Bi . It suffices to show that
∗ ∗
each ϕn [Bi Bi ] is positive. So without loss of generality, assume that A = B B. That is
n
X
Aij = b∗ki bkj , buv ∈ A.
k=1
Then n
X
[ϕn (A)]ni,j=1 = [ϕ(b∗ki bkj )]ni,j=1
k=1
Theorem 3.26. Let (Ai , ϕi )ni=1 be n.c. probability spaces and (A, ϕ) = ∗ni=1 (Ai , ϕi ). Then ϕ is positive.
If each ϕi is faithful, so is ϕ.
26
Then
" r r
#
X ∗ ∗ X
∗ (i) (i) (j) (j)
ϕ [ξ ξ] = ϕ ak ... a1 a1 . . . ak
i=1 j=1
r
X h i h i
(i) (j) (i) ∗ (j)
= ϕu(1) (a1 )∗ a1 . . . ϕu(k) (ak ) ak .
i,j=1
h ir
(i) ∗ (j)
Each matrix ϕu(s) (as ) as is positive, therefore so is their Schur product.
i,j=1
The proof of faithfulness is similar, see Proposition 6.14 in [NS06].
Remark 3.27 (Free convolutions). Let µ, ν be two distributions (probability measures on R). Free prod-
ucts allow us to construct two symmetric, freely independent variables a, b in some tracial C ∗ -probability
space (A, ϕ) such that µ = µa , ν = µb (why?). Know the distribution of a + b is a probability measure on
R, depends only on µ, ν and not on a, b. Thus define the additive free convolution µ ν by µ ν = µa+b .
How to compute?
Similarly, define the multiplicative free convolution µ ν = ϕab . Note that ab is not symmetric, so in
general µ ν is not positive and does not correspond to a measure on R. Also note that since ϕ is tracial,
is commutative.
Suppose that µ, ν are supported in [0, ∞). Then we may choose a, b to be positive. In that case, a1/2 ba1/2
is also positive, and since ϕ is trace,
ϕab = ϕa1/2 ba1/2 .
In this case µ ν may be identified with a probability measure on [0, ∞).
Now suppose instead that µ, ν are supported on the unit circle T = {z ∈ C : |z| = 1}. Then we may
choose a, b to be unitary. Then ab is also unitary, so in this case µ ν may be identified with a probability
measure on T.
Proposition 3.28 (Compare with Exercise 3.11). Let (A, τ ) be a n.c. probability space, a1 , a2 , . . . , an
free, and u a unitary free from them. Then {u∗ ai u, 1 ≤ i ≤ n} are free.
Proof. The key point is that even if τ is not tracial, if a and u are freely independent then
τ [u∗ au] = τ [u∗ τ [a] u] + τ [u∗ a◦ u]
= τ [a] τ [u∗ u] + τ [(u∗ )◦ a◦ u◦ ] + τ [u∗ ] τ [a◦ u◦ ] + τ [(u∗ )◦ a◦ ] τ [u] + τ [u∗ ] τ [a◦ ] τ [u]
= τ [a] .
In particular, τ [a] = 0 if and only if τ [u∗ au] = 0. The rest of the argument proceeds as in the solution to
the exercise.
By a similar argument, we can also get (how?) the following useful weakening of the hypothesis in the
definition of free independence:
27
Proposition 3.29. Let (A, ϕ) be an n.c. probability space. Subalgebras A1 , A2 , . . . , Ak ⊂ (A, ϕ) are
freely independent with respect to ϕ if and only if whenever
ϕ [a1 a2 . . . an ] = 0.
28
Chapter 4
See Lectures 9, 10 of [NS06]. For general combinatorial background, see [Sta97]. The general theory
of incidence algebras, Möbius inversion etc. goes back to the series On the foundations of combinatorial
theory I–X by Gian-Carlo Rota and various co-authors.
B1 ∪ B2 ∪ . . . ∪ Bk = S, Bi ∩ Bj = ∅ for i 6= j, Bi 6= ∅.
if i and j are in the same block of π. Denote |π| the number of blocks of π.
Usually take S = [n] = {1, 2, . . . , n}. Write P([n]) = P(n).
Partitions are partially ordered by reverse refinement: if
π = {B1 , . . . , Bk } , σ = {C1 , . . . , Cr }
then
π≤σ ⇔ ∀i ∃j : Bi ⊂ Cj .
For example, if π = {(1, 3, 5)(2)(4, 6)} and σ = {(1, 3, 5)(2, 4, 6)}, then π ≤ σ. The largest partition is
29
30
Example 4.1. P(3) has 5 elements: one largest, one smallest, and three which are not comparable to each
other.
Partitions form a lattice: for any π, σ ∈ P(S), there exists a largest partition τ such that τ ≤ π, τ ≤ σ,
called the meet and denoted π ∧ σ; and the smallest partition τ such that π ≤ τ , σ ≤ τ , called the join and
denoted π ∨ σ. Clearly
π∧σ π σ
i ∼ j ⇔ i ∼ j and i ∼ j;
π ∨ σ can be thought of as the equivalence relation on S generated by π and σ. For example,
if π = {(1, 3, 5)(2, 4, 6)} and σ = {(1, 2, 3)(4, 5, 6)}, then π ∧ σ = {(1, 3)(2)(4, 6)(5)}.
If π = {(1, 2)(3, 4)(5, 6)(7)} and σ = {(1)(2, 3)(4, 5)(6, 7)}, then π ∨ σ = 1̂7 .
Note that 0̂, 1̂ ∈ NC(S). With the same operations ∧ and ∨ as above, NC(S) is a lattice. If π, σ ∈ NC(S)
they have the same π ∧σ in P(S) and in NC(S). However, for π = {(1, 3)(2)(4)} and σ = {(1)(2, 4)(3)},
π ∨ σ = {(1, 3)(2, 4)} in P(4) but π ∨ σ = 1̂4 in NC(4).
Definition 4.3. π ∈ NC(S) ⊂ P(S) is an interval partition if its classes are intervals. See Figure 4. They
also form a lattice. Denote them by Int(S), Int(n).
For this reason, interval partitions are also called Boolean partitions.
Will study NC(n), but also P(n), Int(n), in a lot more detail.
Enumeration I.
Definition 4.5.
B(n) = |P(n)| = Bell number.
S(n, k) = |{π ∈ P(n) : |π| = k}| = Stirling number of the second kind.
No formula.
31
Lemma 4.6. The Bell numbers satisfy a recursion relation
n
X n
B(n + 1) = B(n − i), n ≥ 1,
i=0
i
Proof of the Lemma. If π ∈ P(n + 1) and the block containing 1 contains i + 1 elements, there are ni
ways to choose the remaining i elements of the block, and the partition induced by π on the complement
of this block is arbitrary. This gives the recursion. Differentiating the generating function term-by-term,
∞ ∞ n
!
X 1 X 1 X n!
F 0 (z) = B(n + 1)z n = B(n − i) z n
n=0
n! n=0
n! i=0
i!(n − i)!
∞ ∞
X 1 iX 1
= z B(j)z j = ez F (z).
i=0
i! j=0 j!
Solving this differential equation with the initial condition F (0) = 1, we get the result.
Exercise 4.8. Derive a recursion for the Stirling numbers, and use it to show that their generating function
is ∞ X n
X 1 z
F (z, w) = 1 + S(n, k)z n wk = ew(e −1) .
n=1 k=1
n!
Definition 4.9.
cn = |NC(n)| = Catalan number.
32
Their ordinary generating function is
∞ √
X
n 1− 1 − 4z
F (z) = cn z = .
n=0
2z
Consequently,
1 2n 1 2n + 1
cn = = .
n+1 n 2n + 1 n
Thus
zF (z)2 − F (z) + 1 = 0,
and √
1 − 4z 1−
F (z) =.
2z
It is easy to see that the second recursion leads to the same quadratic equation, and so holds as well.
Finally, using the binomial series,
∞ ∞ ∞
1 X 1/2 k
X (2k − 2)! k−1 X (2n)!
F (z) = − (−4z) = z = zn,
2z k=1 k k=1
(k − 1)!k! n=0
n!(n + 1)!
I = functions on L
33
and
I2 = functions on {(σ, π) : σ, π ∈ L, σ ≤ π} .
Then I2 is the incidence algebra under convolution
X
(f ∗ g)(σ, π) = f (σ, τ )g(τ, π).
σ≤τ ≤π
Möbius inversion.
Definition 4.12. The Möbius function µ ∈ I2 is defined by
ζ ∗ µ = µ ∗ ζ = δ.
34
b. Partial Möbius inversion: for any ω, π ∈ L,
X X
g(τ ) = f (σ)µ(σ, π).
ω∨τ =π ω≤σ≤π
Proof. X X
f (π) = g(σ) = g(σ)ζ(σ, π) = (g ∗ ζ)(π).
0̂≤σ≤π 0̂≤σ≤π
Therefore X
g(π) = (g ∗ ζ ∗ µ)(π) = (f ∗ µ)(π) = f (σ)µ(σ, π).
σ≤π
More generally,
X X X X X
f (σ)µ(σ, π) = g(τ )µ(σ, π) = g(τ ) ζ(ω ∨ τ, σ)µ(σ, π)
ω≤σ≤π ω≤σ≤π τ ≤σ τ ∈L (ω∨τ )≤σ≤π
X X
= g(τ )δ(ω ∨ τ, π) = g(τ ).
τ ∈L ω∨τ =π
Multiplicative functions.
f (σ, π) or even f (π) too complicated. Want multiplicative functions.
Remark 4.14. In a lattice L, denote
[σ, π] = {τ ∈ L : σ ≤ τ ≤ π} .
In P(S), Y Y
[0̂, π] ' P(Vi ) ' P(|Vi |)
Vi ∈π Vi ∈π
In NC(S), Y Y
[0̂, π] ' NC(Vi ) ' NC(|Vi |).
Vi ∈π Vi ∈π
35
Definition 4.15. Let π ∈ NC(n). Its (right) Kreweras complement K[π] ∈ NC(n) is the largest partition
in NC({1̄, 2̄, . . . , n̄}) such that
b. K is an anti-isomorphism of lattices.
c. |K[π]| + |π| = n + 1.
Proof.
a. Picture.
c. Any element of NC(n) can be included in a maximal chain, all of which look like
with |K[σk ]| = n − k + 1.
{(σ, π) : σ ≤ π, σ, π ∈ Li , i ∈ N}
36
is multiplicative if there exist fixed {fi : i ∈ N} such that whenever
n
Y
[σ, π] ' Lki i ,
i=1
then n
Y
f (σ, π) = fiki .
i=1
Conversely, given any {fi : i ∈ N}, can use this formula to define a multiplicative function f .
Remark 4.21. By general theory, the Möbius function is also always multiplicative. In fact, in general
µL×L0 = µL µL0 .
37
Chapter 5
M :A {z. . . × A} → C
| ×A×
n
by
M [a1 , a2 , . . . , an ] = ϕ [a1 a2 . . . an ] .
For each n, M is a multilinear functional. Combine these into a multiplicative function
Y
Mπ [a1 , a2 , . . . , an ] = M [ai : i ∈ V ]
V ∈π
for π ∈ NC(n).
Example 5.2.
Definition 5.3. In the context of the preceding definition, define the partitioned free cumulants
Rπ [a1 , a2 , . . . , an ]
of {a1 , . . . , an } via
M = R ∗ ζ, R = M ∗ µ.
38
That is, X
Mπ [a1 , a2 , . . . , an ] = Rσ [a1 , a2 , . . . , an ]
σ≤π
and X
Rπ [a1 , a2 , . . . , an ] = Mσ [a1 , a2 , . . . , an ]µ(σ, π).
σ≤π
and X
M [a1 , a2 , . . . , an ] = Rσ [a1 , a2 , . . . , an ],
σ∈NC(n)
where Y
Rσ [a1 , a2 , . . . , an ] = R[ai : i ∈ V ].
V ∈σ
Example 5.4.
R[a] = M [a] = ϕ [a]
is the mean of a.
M [a, b] = R[a, b] + R[a]R[b],
so
R[a, b] = ϕ [ab] − ϕ [a] ϕ [b]
is the covariance of a and b.
Notation 5.5.
R[a1 , . . . , an ] = Rϕ [a1 , . . . , an ]
is a multilinear functional on (A, ϕ). If µ ∈ D(d) is the joint distribution of (a1 , . . . , ad ), then x1 , . . . , xd ∈
Chx1 , . . . , xd i also have joint distribution µ with respect to µ. So can consider
These satisfy
X
Rπµ [xu(1) , xu(2) , . . . , xu(n) ].
ϕ au(1) au(2) . . . au(n) = µ xu(1) xu(2) . . . xu(n) =
π∈NC(n)
39
In one variable, we have the moments
Here
rn [µ] = Rµ [xn ] = Rµ [x, x, . . . , x].
Finally, we have the free cumulant generating function (combinatorial R-transform)
∞
X
µ
R (z) = rn [µ]z n
n=1
Proposition 5.6. The moment and free cumulant generating functions are related by
(recall that M has a constant term while R does not), and more generally
M (z1 , . . . , zd ) = 1 + R z1 M (z), . . . , zd M (z)
= 1 + R M (z)z1 , . . . , M (z)zd ,
where ∞ X
X
M (z1 , . . . , zd ) = 1 + M [x~u ]z~u .
n=1 |~
u|=n
40
Proof. Very similar to the enumeration argument.
X
M [xu(1) , xu(2) , . . . , xu(n) ]z~u = Rπ [x~u ]z~u
π∈NC(n)
n
X X
=
k=1 S⊂[n]
S={1=v(1),v(2),...,v(k)}
v(k+1)=n+1
k
Y X Y
R[xu(v(j)) : 1 ≤ j ≤ k] zu(v(j)) Rπj [xu(i) : i ∈ Vj ] zu(i)
j=1 πj ∈NC(Vj =[v(j)+1,v(j+1)−1]) i∈Vj
n
X X k
Y Y
= R[xu(v(j)) : 1 ≤ j ≤ k] zu(v(j)) M [xu(v(j)+1) , . . . , xu(v(j+1)−1) ] zu(i) .
k=1 S⊂[n] j=1 i∈Vj
S={1=v(1),v(2),...,v(k)}
v(k+1)=n+1
Therefore
∞ X
X
M (z) = 1 + R[xw~ ]zw(1) M (z)zw(2) M (z) . . . zw(k) M (z)
k=1 |w|=k
~
= 1 + R z1 M (z), z2 M (z), . . . , zd M (z) .
41
Exercise 5.8. For the Boolean cumulant generating function
∞ X
X
B(z) = B[xu(1) , . . . , xu(n) ]z~u ,
n=1 |~
u|=n
show that
1 − B(z) = M (z)−1
(inverse with respect to multiplication).
Enumeration II.
Proposition 5.9. On NC(n), the Möbius function is given by
µn = (−1)n−1 cn−1 ,
1 + z = 1 + F (z(1 + z)).
√
−1+ 1+4w
If w = z 2 + z, then z = 2
and
∞ √ ∞
X
n −1 + 1 + 4w X
F (w) = µn w = = (−1)n cn−1 wn
n=1
2 n=1
as before.
Exercise 5.10. Use Remark 5.7 and Exercise 5.8 to compute the Möbius functions on P(n) and Int(n).
0̂ ≤ σ1 ≤ σ2 ≤ . . . ≤ σp ≤ 1̂,
42
Proposition 5.12. The number M (n, p) of multi-chains of length p in NC(n) is the Fuss-Catalan number
1 (p + 1)n
.
n n−1
Similarly, X
M (n, p) = 1 = (ζ ∗ ζ ∗ . . . ∗ ζ )(0̂n , 1̂n ) = (ζ ∗(p+1) )n .
| {z }
0̂≤σ1 ≤σ2 ≤...≤σp ≤1̂ p+1
P∞
If Fp (z) = 1 + n=1 (ζ ∗(p+1) )n z n , then F0 (z) = 1
1−z
and Fp (z) = Fp−1 (zFp (z)). It follows by induction
that Fp satisfies z(Fp (z))p+1 = Fp (z) − 1, since
Fp (z) − 1 = Fp−1 (zFp (z)) − 1 = (zFp (z))(Fp−1 (zFp (z)))p = z(Fp (z))p+1 .
It is known that the coefficients in the expansion of the solution of this equation are the Fuss-Catalan
numbers.
Proposition 5.13. The number of non-crossing partitions in NC(n) with ki blocks of size i is the multino-
mial coefficient
n!
N (n; ki , i ∈ N) = Q P .
i∈N ki !(n + 1 − i∈N ki )!
For the proof, see Lecture 9 from [NS06], or an argument in [Spe98] using the Lagrange inversion formula.
Distributions.
Recall: Stieltjes inversion formula
1
dµ(x) = − lim =Gµ (x + iy) dx,
π y→0+
where ∞
1 X 1
G(z) = M (1/z) = mn n+1
z n=0
z
is the Cauchy transform. Define the analytic R-transform
∞
1 X
R(z) = R(z) = rn z n−1 = r1 + r2 z + r3 z 2 + . . . .
z n=1
43
Exercise 5.14. Use Proposition 5.6 to show that (as formal power series)
1 1
G + R(z) = z, i.e. R(z) = G−1 (z) − ,
z z
inverse with respect to composition.
Example 5.15. Let r1 [µt ] = t, rn [µt ] = 0 for n 6= 1. Then mn [µt ] = tn and µt = δt .
Lemma 5.16. Denote NC2 (n) the non-crossing pair partitions. Then
Proof. By considering the n choices of elements in [2n] to which 1 can be paired by a non-crossing
partitions, it is easy to see that |NC2 (2n)| satisfy the second recursion for the Catalan numbers.
Remark 5.17. A direct bijection NC(n) and NC2 (2n) is provided by the map
π 7→ K[π ∪ K[π]].
Then
1 1 + (t − 1)z
+ R(z) = ,
z z − z2
44
and so p
z+1−t− (z − (t + 1))2 − 4t
Gµt (z) = .
2z
From this one can deduce that
p
1 4t − (x − (t + 1))2
dµ(x) = + max(1 − t, 0)δ0 ,
2π 2x
√ √
the distribution being supported on the interval [( t − 1)2 , ( t + 1)2 ] and has an atom at 0 for 0 < t < 1.
These are the free Poisson distributions, also called the Marchenko-Pastur distributions, which we will
denote by πt . Of particular interest is the free Poisson distribution with parameter t = 1,
r
1 4−x
dµ(x) =
2π x
on [0, 4]. Note that
mn [π1 ] = |NC(n)| = cn = m2n [S(0, 1)].
Thus if s has semicircular distribution, s2 has free Poisson distribution. Thus “free χ2 = free Poisson”.
Remark 5.20. Note that in the previous example,
X n
X
|π|
mn [µt ] = t = tk |{π ∈ NC(n) : |π| = k}| .
π∈NC(n) k=1
Since we know the generating function of these moments, from this we may deduce that
1 n n
|{π ∈ NC(n) : |π| = k}| = N (n, k) =
n k k−1
are the Narayana numbers.
and let
A1 = a1 . . . av(1) ,
A2 = av(1)+1 . . . av(2)
45
up to
Ak = av(k−1)+1 . . . an .
Q
In other words, for Vj = [v(j − 1) + 1, v(j)] (where as usual v(0) = 0), Aj = i∈Vj ai . Clearly
a1 . . . an = A1 . . . Ak .
So
M [A1 , A2 , . . . , Ak ] = ϕ [A1 A2 . . . Ak ] = M [a1 , a2 , . . . , an ].
What about R[A1 , A2 , . . . , Ak ] in terms of Rπ [a1 , a2 , . . . , an ]?
Let
σ = {V1 , V2 , . . . , Vk } ∈ Int(n).
Theorem 5.21. X
R[A1 , A2 , . . . , Ak ] = Rπ [a1 , a2 , . . . , an ].
π∈NC(n)
π∨σ=1̂n
Here π ∨ σ = 1̂n means any two A’s are eventually connected via π. More generally,
X
Rτ [A1 , A2 , . . . , Ak ] = Rπ [a1 , a2 , . . . , an ].
π∈NC(n)
π∨σ=τ̂
Example 5.22.
R[a1 a2 , a3 ] = R{(1,2,3)} + R{(1,3)(2)} + R{(1)(2,3)} = R[a1 , a2 , a3 ] + R[a1 , a3 ]R[a2 ] + R[a1 ]R[a2 , a3 ].
Remark 5.23. Idea of proof:
X
R[A1 , A2 , . . . , Ak ] = Mπ [A1 , . . . , Ak ]µ(π, 1̂)
π∈NC(k)
For example, if σ = {(1)(2, 3, 4)(5, 6)} ∈ Int(6) and τ = {(1, 2)(3)} ∈ NC(3), then
τ̂ = {(1, 2, 3, 4)(5, 6)} ∈ NC(6).
If instead τ = {(1, 3)(2)} then τ̂ = {(1, 5, 6)(2, 3, 4)}.
Then
46
a. τ̂ ∈ NC(n).
Thus [0̂k , 1̂k ] ' [σ, 1̂n ]. In particular, µ(σ, π) = µ(σ̂, π̂), where these are two Möbius functions on
different lattices.
Denoting ω = π̂ and using partial Möbius inversion (Proposition 4.13), the expression continues as
X X
= Mω [a1 , a2 , . . . , an ]µ(ω, τ̂ ) = Rπ [a1 , a2 , . . . , an ].
ω∈NC(n) π∈NC(n)
σ≤ω≤τ̂ π∨σ=τ̂
Proof of the easy direction. Suppose all the mixed free cumulants are zero. Let u(1) 6= u(2) 6= . . . 6=
u(n), ai ∈ Ai , ϕ [ai ] = 0. Write
X
ϕ [a1 a2 . . . an ] = Rπ [a1 , a2 , . . . , an ].
π∈NC(n)
R[ai , ai+1 , . . . , aj ].
If i < j, this is a mixed free cumulant, and so is zero. If i = j, then R[ai ] = ϕ [ai ], and so also is zero. It
follows that each term in the sum is zero.
47
Lemma 5.27. If n ≥ 2 and any of a1 , a2 , . . . , an is a scalar, then
R[a1 , a2 , . . . , an ] = 0.
where we have used the induction hypothesis in the second step. On the other hand, this expression is also
equal to X
ϕ [a1 a2 . . . an−1 ] = Rπ [a1 , a2 , . . . , an−1 ].
π∈NC(n−1)
Proof of the hard direction of the theorem. Suppose A1 , . . . , Ak are freely independent and ai ∈ Au(i) .
First write
R[a1 , . . . , an ] = R[a◦1 + ϕ [a1 ] , . . . , a◦n + ϕ [an ]].
Since R is multilinear, and using the lemma, we may assume that ai = a◦i , ϕ [ai ] = 0. The rest of the
proof is by induction on n. For n = 2, mixed cumulants condition says u(1) 6= u(2), so by freeness
since the first term is zero by freeness, and the second by the induction hypothesis. Finally, if not all but
some of the u(i)’s are different, we may group them so that
u(1) = . . . = u(v(1)),
u(v(1) + 1) = . . . = u(v(2))
u(v(j − 1) + 1) = . . . = u(n),
48
with u(v(1)) 6= u(v(2)) 6= u(v(3)) 6= . . . 6= u(n). Write
A1 = a1 . . . , av(1) , . . . , Aj = av(j−1) . . . an ,
with j < n. By the induction hypothesis, R[A1 , A2 , . . . , Aj ] = 0. But by Theorem 5.21, this expression
also equals X
Rπ [a1 , a2 , . . . , an ]
π∨σ=1̂n
for a certain σ ∈ Int(n), σ 6= 1̂n . Take π 6= 1̂n . By the induction hypothesis, in any non-zero term in
the sum, all the ai ’s in the same block of π have to belong to the same subalgebra. Also by construction,
all the ai ’s in the same block of σ belong to the same subalgebra. But then the condition π ∨ σ = 1̂n
implies that all the ai ’s belong to the same subalgebra, contradicting the assumption. It follows that all
Rπ [a1 , a2 , . . . , an ] = 0 for π 6= 1̂n , and so also for π = 1̂n .
Proof.
(µ ν) xu(1) , xu(2) , . . . , xu(n) = ϕ Xu(1) + Yu(1) , . . . , Xu(n) + Yu(n) ,
where {Xi }di=1 , {Yi }di=1 ⊂ (A, ϕ) are free, µ = ϕX1 ,...,Xd , and ν = ϕY1 ,...,Yd . Thus
Note that the same property holds for the free cumulant generating function and for the R-transform.
Remark 5.29. Random variables are classically independent if and only if their mixed classical cumulants
are zero. The cumulant generating function satisfies
`µ∗ν = `µ + `ν ,
where µ∗ν is the classical convolution, which plays for independence the same role as the free convolution
plays for free independence.
49
Exercise 5.30. Let A1 , A2 , . . . , Ak ⊂ (A, ϕ) be subalgebras not containing the unit of A. They are
Boolean independent if for any u(1) 6= u(2) 6= . . . 6= u(n) and ai ∈ Au(i) ,
ϕ [a1 a2 . . . an ] = ϕ [a1 ] ϕ [a2 ] . . . ϕ [an ] .
a. Show that this definition gives a degenerate structure if the subalgebras were allowed to contain the
unit.
b. Show that Boolean independence is equivalent to the vanishing of mixed Boolean cumulants.
Part (b) implies in particular (you don’t need to prove it, the argument is exactly the same as above) that
B µ]ν = B µ + B ν . Here ] is the Boolean convolution on D(d): if a, b have distributions µ, ν, respectively,
and are Boolean independent, then the distribution of a + b is µ ] ν.
Example 5.31 (Symmetric random walk on a free group). Recall the setting from Example 2.25: let Fn
be the free group with generators x1 , . . . , xn . In (C[Fn ], τ ), have the generating elements ui = δxi , all of
which are Haar unitaries. Consider the element
1
(u1 + u−1 −1 −1
X= 1 ) + (u 2 + u 2 ) + . . . + (u n + un ) .
2n
Note that
mk [X] = probability of the return to e after k steps in a simple random walk on Fn starting at e.
So want to find the distribution of X. Know that each ui + u∗i has the arcsine distribution, and they are
freely independent. Thus
1
Gui +u∗i (z) = √
z2 − 4
and r
1 1
Rui +u∗i (z) = 4 + 2 − .
z z
So s ! r
1 1 n n2 n
RX (z) = n 4+ − = 1 + −
2n (z/2n)2 z/2n z2 z
and r
1 n2 n − 1
+ RX (z) = − 1+ .
z z2 z
Then p
−(n − 1)z + n2 z 2 − (2n − 1)
GX (z) =
z2 − 1
and p
1 (2n − 1) − n2 x2
dµX (x) = dx.
π 1 − x2
These are the Kesten-McKay distributions.
50
Exercise 5.32. Let
1 1
µ = δ−a + δa
2 2
be a symmetric Bernoulli distribution. Compute µ µ. Conclude that a free convolution of two discrete
distributions may be continuous.
Limit theorems.
Theorem 5.33 (Free central limit theorem). Let {Xn : n ∈ X} ⊂ (A, ϕ) be freely independent, identically
distributed, mean zero, variance t
ϕ [Xi ] = 0, ϕ Xi2 = t.
Then
X1 + . . . + Xn
Sn = √
n
converges in distribution to S(0, t), the semicircular distribution with mean 0 and variance t.
Equivalently, denoting by Dα the dilation operator
Z Z
f (x) d(Dα µ)(x) = f (αx) dµ(x)
R R
and µ = µXi ,
D1/√n (µ µ . . . µ) → S(0, t).
| {z }
n
Proof. Note first that by definition mn [Dα µ] = αn mn [µ], from which it follows that rn [Dα µ] = αn rn [µ].
Since S(0, t) is compactly supported, to prove convergence in distribution it suffices to prove convergence
in moments, Z
k
ϕ Sn → xk dS(0, t).
R
By the moment-cumulant formula, it is in fact enough to prove the convergence of free cumulants, that is
Indeed,
nr1 [X]
r1 [Sn ] = √ = 0,
n
nr2 [X]
r2 [Sn ] = = t,
n
and
nrk [X]
rk [Sn ] = → 0, k > 2.
nk/2
51
Remark 5.34. Can replace the “identically distributed” condition with a “uniformly bounded moments”
condition. See [NS06] for an extensive discussion.
Remark 5.36. If µX = (1 − α)δ0 + αδ1 , then mn [X] = α for all n ≥ 1 (m0 [X] = 1). Thus X 2 and X
have the same distribution. May take
X 2 = X = X ∗,
so that X is an (orthogonal) projection, ϕ [X] = α.
t
δ0 + nt δ1 . Note that mk [µn ] = nt . So by the moment-cumulant formula,
Denote µn = 1 − n
t 1
rk [µn ] = + o ,
n n
and
nrk [µn ] = t + o(1) → t
as n → ∞.
Remark 5.37 (Generalized Poisson limit theorem). Let ν ∈ D(d) be a state on Chx1 , . . . , xd i. Denote
t t
µn = 1− δ0 + ν ∈ D(d).
n n
52
We have thus proved the existence of the limit
n
t t
lim 1− δ0 + ν = πν,t ,
n→∞ n n
which is called the free compound Poisson distribution. In particular, πt = πδ1 ,t . Note that since a limit
of states is a state, we have proved that for any ν ∈ D(d) there exists πν,t ∈ D(d) such that
Rπν,t = tν.
53
Chapter 6
See Lectures 2, 7, 8, 9, 13, 21 of [NS06]. For the background on lattice paths, see [Fla80], [Vie84],
[Vie85].
Lattice paths.
Piece-wise linear paths with vertices in Z × Z.
Each step: one unit to the right, {−1, 0} ∪ N units up.
(1, n) rising or NE step.
(1, 0) flat or E step.
(1, −1) falling of SE step.
Start at (0, 0) at height 0. Never go below the x-axis.
In most cases, end on the x axis at height 0.
Combinatorics: move to the right. Fock spaces: flip, move to the left.
Paths needed for enumeration: attach to each step a weight or a label w(step) = a symbol, number,
operator, etc. Y
w(path) = w(step).
Often look at X XY
w(path) = w(step).
paths paths
Definition 6.1. Dyck paths start and end at height 0, and consist of steps: (1, 1) (rise), (1, −1) (fall).
Incomplete Dyck paths: same conditions except ending at the height ≥ 0.
54
55
56
Lemma 6.2. Dyck paths with 2n steps are in a bijective correspondence with NC2 (2n).
Proof. See Figure 6. Rising step corresponds to an opener, falling step to a closer in a partitions. Key
point: a Dyck path corresponds to a unique non-crossing partition.
Corollary 6.3. |Dyck paths(2n)| = cn , the Catalan number. Moreover, if we put on the steps the weights
w(rise) = 1, w(fall) = t, then
X
w(path) = tn cn = m2n [S(0, t)].
Dyck paths(2n)
Here Ω is called the vacuum vector. The full Fock space is the completion of Falg (H) with respect to the
norm coming from the inner product
n
Y
hf1 ⊗ f2 ⊗ . . . ⊗ fn , g1 ⊗ g2 ⊗ . . . ⊗ gk i = δn=k hfi , gi iH .
i=1
Let f ∈ HR . Define the (free) creation operator `(f ) and (free) annihilation operator `∗ (f ) on Falg (H)
by
`(f )(g1 ⊗ g2 ⊗ . . . ⊗ gn ) = f ⊗ g1 ⊗ g2 ⊗ . . . ⊗ gn ,
`(f )Ω = f ,
`∗ (f )(g1 ⊗ g2 ⊗ . . . ⊗ gn ) = hg1 , f i g2 ⊗ . . . ⊗ gn ,
`∗ (f )(g) = hg, f i Ω,
`∗ (f )Ω = 0.
57
Proof.
h`(f )(g1 ⊗ . . . ⊗ gn , h0 ⊗ h1 ⊗ . . . ⊗ hn i = hf ⊗ g1 ⊗ . . . ⊗ gn , h0 ⊗ h1 ⊗ . . . ⊗ hn i
n
Y
= hf, h0 i hgi , hi i
i=1
D E
= g1 ⊗ . . . ⊗ gn , hf, h0 ih1 ⊗ . . . ⊗ hn
= hg1 ⊗ . . . ⊗ gn , `∗ (f )(h0 ⊗ h1 ⊗ . . . ⊗ hn )i .
where the weights are as follows: w(rise) = 1; if the i’th step is a fall matched with a rise on step j, then
w(fall) = hfi , fj i, and the unmatched rises are in positions u(1) < u(2) < . . . < u(k).
Example 6.7.
X(f1 )Ω = f1 ,
X(f1 )X(f2 )Ω = X(f1 )f2 = f1 ⊗ f2 + hf1 , f2 i Ω
See Figure 6.
Note: since all fi ∈ HR , hfi , fj i = hfj , fi i.
Proof. By induction,
X
X(f0 ) w(path)fu(1) ⊗ fu(2) ⊗ . . . ⊗ fu(k)
incomplete Dyck paths(n)
X
= w(path)f0 ⊗ fu(1) ⊗ fu(2) ⊗ . . . ⊗ fu(k)
incomplete Dyck paths(n)
X
+ w(path) fu(1) , f0 fu(2) ⊗ . . . ⊗ fu(k)
incomplete Dyck paths(n)
X
= w(path)fu(1) ⊗ fu(2) ⊗ . . . ⊗ fu(k) .
incomplete Dyck paths(n+1)
58
A different argument was given in class.
a.
X X k
Y
ϕ [X(f1 )X(f2 ) . . . X(fn )] = w(path) = fu(i) , fv(i)
Dyck paths(n) π∈NC2 (n) i=1
π={(u(1),v(1))...(u(k),v(k))}
u(i)<v(i)
b.
Rϕ [X(f1 ), X(f2 ), . . . , X(fn )] = δn=2 hf1 , f2 i .
d. If f1 , f2 , . . . , fn are mutually orthogonal, then {X(f1 ), X(f2 ), . . . , X(fn )} are freely independent.
Call {X(f ) : f ∈ HR } a semicircular system, and any
{X(fi ) : i ∈ S, fi ⊥ fj for i 6= j}
Corollary 6.10. If elements in a semicircular system are uncorrelated, they are freely independent.
Warning: semicircular variables may be uncorrelated without being freely independent.
Theorem 6.11 (Multivariate free CLT). Let {ai,n : i = 1, . . . , d, n ∈ N} ⊂ (A, ϕ) be random variables
such that the d-tuples (a1,n , . . . , ad,n ) are, for n ∈ N, freely independent, identically distributed, have
mean 0 and covariance Q. Then
N N
!
1 X X
√ a1,n , . . . ad,n → (s1 , . . . , sd )
N n=1 n=1
59
Meaning of incomplete Dyck paths? Let f1 , . . . , fn ∈ HR . Define the Wick product W (f1 , f2 , . . . , fn ) ∈
B(F(H)) by W (∅) = 1 and the recursion
Proof. Use the recursion. The second statement follows from the fact (which we do not prove) that the
vector Ω is separating for the algebra, i.e. the only operator in it with AΩ = 0 is A = 0.
{Un } are Chebyshev polynomials of the second kind (with a slightly unusual normalization). Satisfy a
3-term recursion, so by general theory (Theorem 6.43) are orthogonal with respect to some distribution.
Lemma 6.13. {Un : n ∈ N} are the monic orthogonal polynomials with respect to µ = S(0, t), that is
Z
Un (x, t)Uk (x, t) dµ(x) = 0, n 6= k.
R
Proof.
Z
Un (x, t)Uk (x, t) dµ(x) = hUn (X, t)Uk (X, t)Ω, Ωi
R
(
0, k 6= n,
= hUk (X, t)Ω, Un (X, t)Ωi = f ⊗k , f ⊗n = n
t , n = k.
Corollary 6.14.
X
Xn = t#matched steps U#unmatched steps (X, t)
incomplete Dyck paths(n)
n n
= Un (X, t) + c1 tUn−2 (X, t) + c2 t2 Un−4 (X, t) + . . . .
2 4
60
Proof. Combine Proposition 6.6 with the definition of the Wick products.
= X(f )f ⊗k ⊗ g1 ⊗ . . . ⊗ gn − kf k2 f ⊗(k−1) ⊗ g1 ⊗ . . . ⊗ gn
= f ⊗(k+1) ⊗ g1 ⊗ . . . ⊗ gn .
Exercise 6.16. Let HR = Rn be a real Hilbert space, H = Cn its complexification, and
Γ0 (H) = W ∗ (X(f ) : f ∈ HR ).
Prove that
Γ0 (H) ' L(Fdim H ).
You may use without proof the analytic facts L(Z) ' L∞ (T) (proved via Fourier transforms) and, for
a = a∗ , W ∗ (a) ' L∞ (σ(a)). To be completely precise, you would also need a von Neumann algebra
version of Proposition 6.6 and Theorem 7.9 from [NS06], namely that the von Neumann algebra generated
by freely independent subalgebras is (isomorphic to) their free product.
Exercise 6.17. Exercise 8.26 in [NS06]. Fill in the details in the following argument.
For f ∈ HR , show that the operator X(f ) = `(f ) + `∗ (f ) on F(H) has, with respect to the vacuum
expectation on B(F(H)), the same distribution as the operator
f ⊕ f ⊕ ... ⊕ f f ⊕ f ⊕ ... ⊕ f ∗ f ⊕ f ⊕ ... ⊕ f
X √ =` √ +` √
N N N
on F(HN ), with respect to the vacuum expectation on B(F(HN )). Here HN = |H ⊕ .{z
. . ⊕ H}. Show that
N
the second operator is a sum of N freely independent operators, and apply the free central limit theorem
to conclude that X(f ) is a semicircular element. Hint: what is f ⊕f√⊕...⊕f
N
?
61
6.2 Łukasiewicz paths and Voiculescu’s operator model.
Lattice paths and Fock spaces II.
Definition 6.18. Łukasiewicz paths are lattice paths starting and ending at height 0, with rising steps
(1, k), k ∈ N, flat steps (1, 0), and falling steps (1, −1). We will always use weight 1 for the falling steps
and weight rk+1 for a rising step (1, k), including weight r1 for a flat step (1, 0).
Exercise 6.19. Use the idea in Figure 6 to show that Łukasiewicz paths with n steps are in a bijection
with NC(n).
Proposition 6.20 (Voiculescu’s operator model). Let µ ∈ D(1), with free cumulants {rk : k ∈ N}. Let f
be a unit vector, and ` = `(f ) a free creation operator on a full Fock space. Note that the only relation
between ` and `∗ is `∗ ` = 1. Let
∞
X
Xµ = ` + rk (`∗ )k−1 = ` + r1 + r2 `∗ + r3 (`∗ )2 + . . .
k=1
Proof.
* ∞
!n +
X X X Y
Xµn Ω, Ω = `+ rk (`∗ )k−1 Ω, Ω = w(path) = r|V | .
k=1 Łukasiewicz paths(n) π∈NC(n) V ∈π
Proposition 6.21. If f1 , . . . , fk are mutually orthogonal, then {Alg∗ (`(fi )) : 1 ≤ i ≤ k} are freely inde-
pendent with respect to the vacuum expectation.
Proof. See Proposition 7.15 in [NS06], or imitate the proof of Proposition 3.5.
Proposition 6.22. Let f1 ⊥ f2 be unit vectors, `1 = `(f1 ) and `2 = `(f2 ), so that the relation between
them is
`∗i `j = δi=j
Let ∞
X
Xµ = `1 + rk [µ](`∗1 )k−1 ,
k=1
∞
X
Xν = `2 + rk [ν](`∗2 )k−1 .
k=1
62
Then Xµ and Xν are freely independent, and Xµ + Xν has the same distribution as
∞
X
Y = l1 + (rk [µ] + rk [ν])(l1∗ )k−1 .
k=1
Can do this in the multivariate case. See [Haa97] for an alternative argument.
63
Lemma 6.26. Motzkin paths with n steps with labels as in the preceding remark are in a bijection with
NC(n). Under this bijection, the height of a step in the path equals the number of open blocks at that
location in the partition.
Proof. See Figure 6. Openers of blocks correspond to rising steps, closers to falling steps, middle ele-
ments to flat steps labeled with m, singletons to flat steps labeled with s.
Remark 6.27. Under this bijection, Dyck paths correspond to NC2 (n), Motzkin paths with no flat steps
at height 0 to non-crossing partitions with no singletons, and unlabeled Motzkin paths to NC1,2 (n), non-
crossing partitions whose blocks have at most 2 elements.
Remark 6.28. Let HR be a real Hilbert space, H its complexification, F(H) its full Fock space, and
ϕ = h·Ω, Ωi the vacuum expectation. Let
λ1 , λ2 , . . . , λd ∈ R,
ξ1 , ξ2 , . . . , ξd ∈ HR ,
and
T1 , T2 , . . . , Td ∈ L(HR )
be symmetric, possibly unbounded, linear operators defined on a common domain D ⊂ HR containing
{ξi : 1 ≤ i ≤ d} and invariant under all Ti . That is, for any ~u,
is defined. On the Fock space, we have the free creation and annihilation operators `(ξi ), `∗ (ξi ). Define
Λ(Ti )Ω = 0, and
Xi = `(ξi ) + `∗ (ξi ) + Λ(Ti ) + λi .
It is easy to check that Λ(T ) is symmetric if T is. So each Xi is symmetric. What is the joint distribution
of (X1 , . . . , Xd ) with respect to ϕ?
Lemma 6.29.
Rϕ [Xi ] = λi ,
Rϕ [Xi , Xj ] = hξi , ξj i ,
Rϕ [Xu(1) , Xu(2) , . . . , Xu(n) ] = ξu(1) , Tu(2) . . . Tu(n−1) ξu(n) .
Proof.
X
ϕ Xu(1) Xu(2) . . . , Xu(n) = Xu(1) Xu(2) . . . , Xu(n) Ω, Ω = hw(path)Ω, Ωi ,
Motzkin paths(n)
64
where the rising step in position i has weight `∗ (ξu(i) ), the falling step `(ξu(i) ), a flat “m” step weight
Λ(Tu(i) ), and a flat “s” step weight λu(i) . For example, the term corresponding to the path/partition in
Figure 6 is (moving right-to-left)
a. µ is -infinitely divisible.
b. For all t ≥ 0, all k ∈ N, and |~u| = k, t R~uµ = R~uµt for some µt ∈ D(d).
{µt : t ≥ 0} µt µs = µt+s , µ1 = µ.
Proof. µ is -infinitely divisible if and only if there exists νn with n1 R~uµ = R~uνn . Denote µ1/n = νn , and
for p/q ∈ Q+ , µp/q = νqp ∈ D(d) satisfies
µ p
R~up/q = R~uµ .
q
By continuity, can extend this to real t.
Example 6.32. Recall that rn [S(0, t)] = tδn=2 and rn [πt ] = t. It follows that {S(0, t) : t ≥ 0} and
{πt : t ≥ 0} form free convolution semigroups, and in particular are -infinitely divisible. More gen-
erally, for any ν ∈ D and any t ≥ 0, the corresponding free compound Poisson distribution µν,t is
-infinitely divisible.
so ρ is not positive.
65
Lemma 6.35. Every P ∈ C0 hx1 , . . . , xd i is of the form
d
X d
X
P = λ i xi + xi Pij (x)xj , Pij ∈ Chx1 , . . . , xd i.
i=1 i,j=1
is conditionally positive. If d = 1, the converse also holds. That is, any conditionally positive functional
on C[x] is of the form ρ [xn ] = ψ [xn−2 ] for some positive ψ and n ≥ 2.
ψ [P ∗ P ] = ρ [xP ∗ P x] ≥ 0.
Remark 6.36 (Schürmann’s construction). Let ρ on C0 hx1 , . . . , xd i be conditionally positive. GNS con-
struction for (C0 hx1 , . . . , xd i, ρ) produces a Hilbert space
ρ [xi ] = λi ,
ρ [xi xj ] = hξi , ξj i ,
ρ xu(1) xu(2) . . . xu(n) = ξu(1) , Tu(2) . . . Tu(n−1) ξu(n) .
Note also that the conditions from Remark 6.28 are satisfied. Therefore that construction give a state
µ ∈ D(d) with
66
Proposition 6.37 (Schoenberg correspondence). For distributions with finite moments, µ is -infinitely
divisible if and only if Rµ is a conditionally positive functional.
Then
d
Rµ [xu(1) xu(2) . . . xu(n) ] = Rµ [xu(1) , xu(2) , . . . , xu(n) ] =
µt xu(1) xu(2) . . . xu(n) .
dt t=0
since µ0 [P ∗ P ] = δ0 [P ∗ P ] = (P ∗ P )(0) = 0.
For the converse direction, suppose that Rµ is conditionally positive. Then tRµ is also conditionally
positive, and so by the Fock space constructions above, tRµ = Rµt for some µt ∈ D(d).
Exercise 6.38. Let (A, ψ) be an n.c. probability space, and suppose for simplicity that ψ is faithful. On
the full Fock space F(L2 (A, ψ)), for f ∈ A symmetric, define the creation and annihilation operators
`(f ), `∗ (f ) as before, and
Λ(f )(g1 ⊗ g2 ⊗ . . . ⊗ gn ) = (f g1 ) ⊗ g2 ⊗ . . . ⊗ gn ,
is symmetric. Compute
hX(f1 )X(f2 ) . . . X(fn )Ω, Ωi .
Deduce that each X(f ) has a free compound Poisson distribution, and if fi fj = 0 for i 6= j, then X(fi )
are freely independent with respect to the vacuum expectation.
Remark 6.39 (Classically infinitely divisible distributions). Let HR and the algebraic Fock space Falg (H)
be as before. On this space, define the pre-inner product
X n
Y
hf1 ⊗ . . . ⊗ fn , g1 ⊗ . . . ⊗ gk is = δn=k fi , gα(i) .
α∈Sym(n) i=1
For example
hf1 ⊗ f2 , g1 ⊗ g2 is = hf1 , g1 i hf2 , g2 i + hf1 , g2 i hf2 , g1 i .
67
This is a degenerate inner product, since kf ⊗ g − g ⊗ f ks = 0. Quotient out by tensors of norm 0, get
the symmetric Fock space Fs (H), on which f ⊗s g = g ⊗s f . On this space, define the creation operators
a+ (f )(g1 ⊗ . . . ⊗ gn ) = f ⊗ g1 ⊗ . . . ⊗ gn ,
annihilation operators
n
X
−
a (f )(g1 ⊗ . . . ⊗ gn ) = hgi , f i g1 ⊗ . . . ⊗ ĝj ⊗ . . . ⊗ gn ,
j=1
Again,
Xi = a+ (ξi ) + a− (ξi ) + Λ(T
e i ) + λi
are symmetric (with respect to the new inner product). However, for {ξi , Ti , λi } as in Remark 6.28, these
Xi ’s commute. Moreover, if µ ∈ D(d) is their joint distribution, its classical cumulants are precisely
Proposition 6.40. µ ∈ D(d) is ∗-infinitely divisible if and only if µ comes from (λi , ξi , Ti )di=1 on a sym-
metric Fock space, if and only if k µ is a symmetric conditionally positive functional.
Corollary 6.41. For d = 1, the Bercovici-Pata bijection is the bijection
where
k µ = RBP[µ] .
This is a bijection since both sets {k µ : µ ∗ −ID} and {Rµ : µ −ID} consist precisely of all condi-
tionally positive functionals. Moreover, this map is a homomorphism with respect to the convolution
operations,
BP[µ ∗ ν] = BP[µ] BP[ν],
and homeomorphism with respect to convergence in moments.
a+ (f )Ω = f, a+ (f )g = 0,
68
and annihilation operators
a− (f )Ω = 0, a− (f )g = hg, f i Ω.
Extend T ∈ L(H) to the Boolean Fock space by T Ω = 0. Finally, denote PΩ the projection onto Ω,
PΩ Ω = Ω, PΩ g = 0. Define
Xi = a+ (ξi ) + a− (ξi ) + Ti + λi PΩ .
a. Show that the joint distribution of (X1 , . . . , Xd ) with respect to the vacuum expectation is ]-
infinitely divisible.
Hint: in the free case, the argument goes as follows. Via the Schoenberg correspondence, free
infinite divisibility is equivalent to the conditional positivity of the free cumulant functional. This
in turn is equivalent to a representation on a free Fock space coming from the data (λi , ξi , Ti )di=1 :
one direction is Schürmann’s construction, the other follows from the existence of the Fock space
constructions for all re-scalings of this data by positive t.
b. Let µ ∈ D(d) be arbitrary. Using the GNS construction, we can represent the n.c. probability space
(Chx1 , . . . , xd i, µ) on a Hilbert space K with µ = h·Ω, Ωi. Let H = K CΩ,
λi = hλ(xi )Ω, Ωi ∈ C,
ξi = (λ(xi ) − λi PΩ )Ω ∈ K,
Ti = λ(xi ) − a+ (ξi ) − a− (ξi ) − λi PΩ ∈ L(K).
Show that these satisfy all the conditions in Remark 6.28 and in part (a) (on H, not on K). In
particular, show that ξi ∈ H and Ti maps H to itself. Show that on the Boolean Fock space of
H, CΩ ⊕ H ' K, the joint distribution of X1 , . . . , Xd arising from this family (λi , ξi , Ti )di=1 , with
respect to the vacuum state, is µ. Conclude that any distribution is infinitely divisible in the Boolean
sense.
|1 ⊗ 1 ⊗
{z. . . ⊗ 1} = en ,
n
e0 = Ω, we identify
F(C) = Span (en : n ≥ 0).
69
Choose the inner product so that {en : n ≥ 0} form an orthogonal basis, with ken k2 = γn γn−1 . . . γ1 ,
ke0 k = 1. Define
a+ en = en+1 ,
a− en = γn en−1 , a− e0 = 0,
a0 e n = β n e n .
X = a+ + a− + a0
a. See Figure 6.
and in general X
mn [µ] = w(path).
Motzkin paths(n)
c. Any measure µ all of whose moments are finite arises in this way. The corresponding pair of
sequences
β0 β1 β2 . . .
J(µ) =
γ1 γ2 γ3 . . .
are the Jacobi parameters of µ.
d. Let {Pn (x) : n ≥ 0} be the monic orthogonal polynomials with respect to µ, that is Pn (x) = xn +. . .
and Z
Pn (x)Pk (x) dµ(x) = 0, n 6= k
R
70
Then these polynomials satisfy the three-term recursion
Proof.
b. Figure omitted. We will show by induction that the continued fraction up to γn expands to the sum
over Motzkin paths of height at most n, plus higher order terms. Indeed, taking an extra level in the
continued fraction corresponds to replacing each γn with γn times the sum over Motzkin paths of
height at most 1, with labels βn , γn+1 . But inserting such paths between each rise at height n and
the corresponding fall is exactly how one goes from a path of height n to a path of height n + 1.
c. It suffices to show how to compute the Jacobi parameters from the moments. Indeed,
XPn (x)e0 = Xen = en+1 + βn en + γn en−1 = Pn+1 (X)e0 + βn Pn (X)e0 + γn Pn−1 (X)e0 .
Then X
mn [µt ] = tn/2 ,
Dyck paths(n)
and µt = S(0, t). The orthogonal polynomials are the Chebyshev polynomials of the second kind.
71
Example 6.45. Let βn = βt and γn = γt for all n. Thus
βt βt βt . . .
J(µt ) =
γt γt γt . . .
Then X X
mn [µt ] = (γt)#falling steps (βt)#flat steps = Rπ [µt ],
Motzkin paths(n) π∈NC1,2 (n)
where r1 [µt ] = βt, r2 [µt ] = γt, rn [µt ] = 0 for n > 2. Thus µt = S(βt, γt) = S(β, γ)t .
Then
X X
mn [µt ] = (γt)#falling steps (βt)#singleton flat steps b#middle flat steps = Rπ [µt ],
labeled Motzkin paths π∈NC(n)
where r1 [µt ] = βt, rn [µt ] = γbn−2 t. Thus µt = µt , where r1 [µ] = β, rn [µ] = γbn−2 . µ is a shifted,
scaled free Poisson distribution:
(
a + r1 [ν], n = 1,
rn [Dα ν δa ] =
αn rn [ν], n > 1.
Definition 6.47. Let β, b, c ∈ R and γ ≥ 0. The free Meixner distributions {µt } with these parameters
are determined by their Jacobi parameters
βt βt + b βt + b . . .
J(µt ) =
γt γt + c γt + c . . .
Proof. See Figure 6. Take a Motzkin path with labels βt, βt + b, γt, γt + c. Separate it into Motzkin
paths with labels βt on flat steps, b on flat steps at height at least 1, γt on falling steps, c on falling steps
at height at least 1. Form a new labeled Motzkin path by
72
This corresponds to a π ∈ NC(n). On each non-singleton class of π, have a further (b, c)-Motzkin path,
starting with a γt-rise. Thus
X Y X
mn [µt ] = (βt)|Sing(π)| (γt) w(path)
π∈NC(n) V ∈π,|V |≥2 (b,c)−Motzkin paths on |V |−2
X Y
= (βt)|Sing(π)| (γt)m|V |−2 (S(b, c))
π∈NC(n) V ∈π,|V |≥2
and
r1 [µt ] = βt, rn [µt ] = γtmn−2 (S(b, c)).
Free Meixner distributions include the semicircular, free Poisson, arcsine, Kesten-McKay, Bernoulli and
free binomial distributions.
Remark 6.49. Let β = b = 0, γ = 1, c = −1. Thus
0 0 0 ...
J(µt ) =
t t − 1 t − 1 ...
For t = 1,
0 0 0 ...
J(µt ) =
1 0 0 ...
so that
1
M µ (z) =
1 − z2
and
1 1 z 1 1 1
Gµ (z) = = 2 = − .
z 1 − z −2 z −1 2 z−1 z+1
So
1
µ = (δ−1 + δ1 )
2
is the symmetric Bernoulli distribution. Easy to see: µ∗t is defined only for t ∈ N. But µt is defined for
all t ≥ 1! Compare with the Bercovici-Pata bijection.
Exercise 6.50.
a. If
β0 β1 β2 . . .
J(µ) = ,
γ1 γ2 γ3 . . .
show that its Boolean cumulants are B1 [µ] = β0 , Bn+2 [µ] = γ1 mn [ν], where
β1 β2 . . .
J(ν) =
γ2 γ3 . . .
The map µ 7→ ν is the coefficient stripping.
b. Conclude that
]t β0 t β1 β2 . . .
J(µ ) = ,
γ1 t γ2 γ3 . . .
73
6.5 q-deformed free probability.
See [BS91], [Ans01].
Definition 6.51. Let q be a number or a symbol. The q-integers are
1 − qn
[n]q = 1 + q + q 2 + . . . + q n−1 =
1−q
The notion goes back to Euler (≈ 1750), Rogers (1894), Ramanujan (1913).
Example 6.52. [n]1 = n, (
0, n = 0,
[n]0 =
1, n ≥ 1.
Consider Dyck paths with weights 1 on rising steps, and weight [n]q = 1 + q + . . . + q n−1 on a falling
step at height n.
From general theory, if X
mn [µ(x|q)] = w(path),
Dyck paths
then µ(x|q) is the orthogonality measure for the polynomials {Pn (x|q)} which are defined via the recur-
sion
xPn (x|q) = Pn+1 (x|q) + [n]q Pn−1 (x|q).
These are the continuous (Rogers) q-Hermite polynomials. For q = 1, they are the Hermite polynomials.
For q = 0, they are the Chebyshev polynomials of the 2nd kind.
Have explicit formulas for µ(x|q), Pn (x|q). µ(x|q) is the q-Gaussian distribution.
Equivalently, X
mn [µ(x|q)] = w(path),
labeled Dyck paths
where each falling step at height n may have one of the labels 1, q, . . . , q n−1 . Using this idea, on Falg (H),
define a+ (f ) = `(f ) but
n
X
−
a (f )(g1 ⊗ . . . ⊗ gn ) = q i−1 hfi , gi g1 ⊗ . . . ⊗ ĝi ⊗ . . . ⊗ gn .
i=1
Note that on the full Fock space F(H), these are not adjoints.
Definition 6.53. On the algebraic full Fock space, define the q-inner product
X n
Y
inv(α)
hf1 ⊗ . . . ⊗ fn , g1 ⊗ . . . ⊗ gk iq = δn=k q fi , gα(i) ,
α∈Sym(n) i=1
74
Theorem 6.54 (Bożejko, Speicher 1991; Zagier 1992). The inner product h·, ·iq is positive definite for
−1 < q < 1. It is positive semi-definite for q = −1, 1.
k·kq
Proposition 6.55. Let Fq (H) = Falg (H) . On this space, (a+ (f ))∗ = a− (f ), and so
X(f ) = a+ (f ) + a− (f )
is symmetric.
Proposition 6.56 (Wick formula). Let f be a unit vector. The distribution of X(f ) is the q-Gaussian
distribution. Its moments are
X X
X(f )2n Ω, Ω = w(path) = q cr(π) ,
labeled Dyck paths π∈P 2 (2n)
where for a general π ∈ P(n), cr (π) is the number of reduced crossings of the partition π,
n o
π π
cr (π) = (i < j < k < l) : i ∼ k, j ∼ l, both consecutive in their blocks.
Proof. See Figure 6. To show: a bijection between P 2 (2n) and labeled Dyck paths with 2n steps. Openers
correspond to rising steps, closers to falling steps. Note that for each i, the number of open blocks before
i is equal to the height of the path on step i. Thus if on the i’th step, we close the j’th open block, that
step is labeled by q j−1 .
Corollary 6.58. X
q inv(α) = [n]q ! = [n]q [n − 1]q . . . [1]q .
α∈Sym(n)
Proof. For kf k = 1,
X
q inv(α) = f ⊗n , f ⊗n q
= (a− (f ))n f ⊗n , Ω q
.
α∈Sym(n)
(0) (1)
Note: rn = rn , rn = kn .
Theorem 6.60. If ρ is conditionally positive, there exists a positive measure µ such that
(q)
rn+2 [µ] = ρ [xn ] .
Denote all measures arising in this way by IDq .
75
Corollary 6.61. If define µ q ν by
Γq (H) = W ∗ (X(f ) : f ∈ HR ),
where X(f ) are now operators defined in Proposition 6.55. It is a spectacular result of (Guionnet,
Shlyakhtenko 2012), based on previous work (Dabrowski 2010), that Γq (H) ' L(Fdim H ) for sufficiently
small q (depending on dim H).
76
Chapter 7
This chapter was omitted in the course, so what follows is only an outline. See [GSS92].
Definition 7.1. In (A, ϕ), a family of random variables {X(t) : t ≥ 0} is a free Lévy process if
are free.
b. It has stationary increments: for s < t,
µX(t)−X(s) = µX(t−s) .
Assuming X(0) ∼ δ0 and denoting µX(t) = µt , it follows that {µt : t ≥ 0} form a -semigroup.
c. The semigroup of distributions is continuous at t = 0.
Remark 7.2. Can construct such processes as injective limits of free products. More explicitly, start with
a -semigroup {µt : t ≥ 0}. Let ρ = Rµ , Rµt = tρ. Let H be the Hilbert space and (λ, ξ, T ) the triple
from Remark 6.28. Let K = H ⊗ L2 [0, ∞). For I ⊂ [0, ∞) a finite half-open interval, denote
λI = λ |I| , ξI = ξ ⊗ 1I , TI (ζ ⊗ f ) = (T ζ) ⊗ (1I f ).
More generally, can construct {X(f ) : f ∈ L2 [0, ∞) ∩ L∞ [0, ∞)}. For the free Brownian motion, only
need f ∈ L2 [0, ∞).
77
Markov property.
See [Bia98].
For free Lévy processes, in particular for the free Brownian motion, have free stochastic calculus etc.
Here we only show that such a process has the Markov property.
Definition 7.3. Let (A, ϕ) be a C ∗ -probability space, B ⊂ A a C ∗ -subalgebra. The map ϕ [·|B] : A → B
is a conditional expectation if it is self-adjoint and for any a ∈ A and b1 , b2 ∈ B,
It does not always exist. If ϕ is a trace, can choose a conditional expectation so that moreover ϕ [ϕ [a|B]] =
ϕ [a]. Then
ϕ [ϕ [a|B] b] = ϕ [ϕ [ab|B]] = ϕ [ab] .
Thus ϕ [a|B] ∈ B and hb, ϕ [a|B]iϕ = hb, aiϕ . For ϕ faithful, ϕ-preserving conditional expectation is
unique. Again, it may not exist.
Theorem 7.5.
b. For t ≥ s,
1 ∗ 1
ϕ C (X(r) : r ≤ s) = .
1 − X(t)z + Rµt (z) 1 − X(s)z + Rµs (z)
1
That is, t 7→ 1−X(t)z+Rµt (z)
is a martingale. Need to be more precise on what type of object z is.
There are deep conceptual reasons for the Markov property of a process with freely independent incre-
ments discovered by Voiculescu.
78
defined by X
∂xi (xu(1) xu(2) . . . xu(n) ) = xu(1) . . . xu(j−1) ⊗ xu(j+1) . . . xu(n)
u(j)=i
and so
p(x) − p(y)
∂x p(x) = .
x−y
Definition 7.7. In an n.c. probability space (A, ϕ), define the free Appell polynomials to be multilinear
maps A(X1 , X2 , . . . , Xn ) from A × A × . . . × A to A such that
a. ∞
X 1
An (x)z n = ,
n=0
1 − xz + Rµ (z)
and a similar result holds in the multivariate case.
b.
n
X
X0 A(X1 , X2 , . . . , Xn ) = A(X0 , X1 , . . . , Xn ) + R[X0 , X1 , . . . , Xj ]A(Xj+1 , . . . , Xn ).
j=1
c. If A1 , . . . , Ad ⊂ A are free,
X1,1 , X1,2 , . . . , X1,u(1) ∈ Av(1) , ..., Xk,1 , Xk,2 , . . . , Xk,u(k) ∈ Av(k)
for v(1) 6= v(2) 6= . . ., then
79
An (x)z n , then µ[F ] = 1 and
P
Proof. For (a), if F (x, z) = n
X
∂x F (x, z) = Ai ⊗ An−1−i z n = zF ⊗ F.
n,i
1
These conditions uniquely determine F (x, z) = 1−xz+R µ (z) . The proof of part (b) uses a similar technique.
For part (c), we show that the factored polynomials also satisfy the initial conditions and the recursion
from part (b).
Proposition 7.9. In a tracial C ∗ -probability space (A, ϕ), if B ⊂ A, Y ∈ B, and X is free from B, then
ϕ A(X + Y, X + Y, . . . , X + Y ) |B = A(Y, Y, . . . , Y ).
| {z }
n
Using the preceding theorem, freeness, and the fact that the free Appell polynomials are centered, we
continue the expression as
X
= ϕ [A(X, . . . , X)A(Y, . . . , Y )A(X, . . . , X) . . . A(X or Y )b] = ϕ [A(Y, Y, . . . , Y )b] .
Proof of Theorem 7.5. We are interested in ϕ [P (X(t))|C ∗ (X(r), r ≤ s)]. Since they span the space, it
suffices to show this for P = An . In that case,
ϕ [An (X(t))|C ∗ (X(r), r ≤ s)] = ϕ [An (X(s) + (X(t) − X(s)))|C ∗ (X(r), r ≤ s)] = An (X(s))
80
Chapter 8
See Lectures 12, 14 of [NS06]. In the preceding chapters, we tried to emphasize the similarity and parallels
between classical and free probability, through the use of all/non-crossing partitions, symmetric/full Fock
space, and infinitely divisible distributions. By contrast, the phenomena described in this and the following
chapters are special to free probability.
Remark 8.1. Recall: if a, b are free, µab = µa µb , the free multiplicative convolution. Note that if
{a1 , . . . , an } and {b1 , . . . , bn } are free,
X
ϕ [a1 b1 a2 b2 . . . an bn ] = Rπ [a1 , b1 , a2 , b2 , . . . , an , bn ]
π∈NC(2n)
X X
= Rσ [a1 , a2 , . . . , an ] Rτ [b1 , b2 , . . . , bn ]
σ∈NC(n) τ ∈NC(n)
σ∪τ ∈NC(2n)
X X
= Rσ [a1 , a2 , . . . , an ] Rτ [b1 , b2 , . . . , bn ]
σ∈NC(n) τ ∈NC(n)
τ ≤K[σ]
X
= Rσ [a1 , a2 , . . . , an ]MK[σ] [b1 , b2 , . . . , bn ].
σ∈NC(n)
P
In particular, mn [µab ] = σ∈NC(n) Rσ [a]MK[σ] [b].
81
Lemma 8.2. Let a1 , . . . , ak be self-adjoint, orthogonal random variables in a n.c. probability space
(A, τ ), in the sense that ai aj = 0 for i 6= j. Let s be a semicircular variable free from {ai : 1 ≤ i ≤ k}.
Then
{sa1 s, sa2 s, . . . , sak s}
are freely independent and have free compound Poisson distributions.
See Example 12.19 in [NS06] for the proof in the non-tracial case.
It follows that
R[sau(1) s, sau(2) s, . . . , sau(n) s] = M [au(1) , au(2) , . . . , au(n) ] = τ au(1) au(2) . . . au(n) = 0
unless all u(1) = u(2) = . . . = u(n). Thus {sa1 s, sa2 s, . . . , sak s} are free. Moreover,
rn [sai s] = R[sai s, sai s, . . . , sai s] = M [ai , ai , . . . , ai ] = mn [ai ],
so sai s has a free compound Poisson distribution.
R1
Example 8.3. Let A = C ∗ (s) ∗ L∞ [0, 1], with the free product state τ = ϕ ∗ 0
· dx. Then the map
t 7→ s1[0,t) s
is a free Poisson process for t ∈ [0, 1]. Indeed, for I ∩ J = ∅, 1I 1J = 0, so the process has freely
independent increments, and
rk [s1I s] = mk [1I ] = |I|
so s1I s has distribution π|I| .
Free compression.
Proposition 8.4. Let a1 , . . . , ad be in a n.c. probability space (A, τ ). Choose p freely independent from
them, where p is a projection, p = p2 = p∗ . Denote α = τ [p]. Define a new (compressed) n.c. probability
space
(A,e τe) = (pAp, 1 τ |pAp ).
α
Then
1
Rτe[pau(1) p, pau(2) p, . . . , pau(n) p] = Rτ [α au(1) , α au(2) , . . . , α au(n) ].
α
82
See Theorem 14.10 in [NS06] for the proof in the non-tracial case.
X 1 |π|
= Rπτ [α au(1) , α au(2) , . . . , α au(n) ].
α
π∈NC(n)
Corollary 8.5. For arbitrary µ ∈ D(d) and arbitrary t ≥ 1, the expression µt is well defined, in the
sense that there exists µt ∈ D(d) with Rµt = tRµ .
1
Proof. Choose a1 , . . . , ad and p as in the proposition so that τa1 ,...,ad = Dt µ and τ [p] = t
. Then in
(pAp, τe), (pa1 p, pa2 p, . . . , pad p) has distribution µt .
Corollary 8.6. The Bercovici-Pata bijection ID∗ → ID cannot be extended to general measures.
Proof. Such an extension T would in particular have the property that T −1 [T [µ]t ] = µ∗t . However,
taking µ to be the Bernoulli distribution, the right-hand-side is defined only for integer t, while the left-
hand-side is defined for any t ≥ 1.
Exercise 8.7. Let p, q ∈ (A, τ ) be freely independent projections in a (for convenience) tracial n.c.
probability space, with τ [p] = α, τ [q] = β. Note that the distributions of pq, pqp, and qpq are all the
same. Use the results above, and either the R-transforms or an extension of Remark 6.49 to compute
the distribution of pq. The computation of the moment generating function or the Cauchy transform is
sufficient, you do not need to find the actual measure. See Example 13.5 for the answer.
Outline of the argument: for p as above and any a free from it, in the notation of Proposition 8.4
µτpap
e
= (µταa )(1/α) .
83
and
1
Gpap (z) = (1 − α) + αGµ(1/α) (z).
z αa
Finally,
1
(1/α) (z) =
Rµαa Rαa (z) = Ra (αz).
α
Using these results, we compute Gq , then Rq , then Rµ(1/α)
αq
, then Gµ(1/α)
αq
, and finally Gpqp .
See Lecture 14 of [NS06] for related results about compression by matrix units, very useful in applications
of free probability to the study of operator algebras.
84
Chapter 9
Belinschi-Nica evolution.
See [BN09].
Remark 9.1. Recall that
while
µ ] -infinitely divisible ⇔ B µ = ρ is conditionally positive,
and any µ is ]-infinitely divisible. So we can define a map
B : D(d) → ID
while X
M [a1 , a2 , . . . , an ] = Bπ [a1 , a2 , . . . , an ]
π∈Int(n)
85
86
Proof. Figure omitted. Group the blocks of π ∈ NC(n) according to the smallest σ ∈ Int(n) such that
π ≤ σ. Then the restriction of π to each block of σ connects the minimum and the maximum of that
block.
Example 9.3. See Figure 9.
B[a1 , a2 , a3 , a4 ] = R[a1 , a2 , a3 , a4 ] + R[a1 , a2 , a4 ]R[a3 ] + R[a1 , a3 , a4 ]R[a2 ]
+ R[a1 , a4 ]R[a2 , a3 ] + R[a1 , a4 ]R[a2 ]R[a3 ].
Corollary 9.4.
X X
B B[µ] [xu(1) , xu(2) , . . . , xu(n) ] = RπB[µ] [xu(1) , xu(2) , . . . , xu(n) ] = Bπµ [xu(1) , xu(2) , . . . , xu(n) ]
π∈NC(n)
g π∈NC(n)
g
a. For π ∈ NC(n),
{σ ∈ NC(n) : σ π} ' {σ ∈ NC(n) : π0 ≤ σ ≤ π} = [π0 , π] ⊂ NC(n).
b. For σ ∈ NC(n),
{π ∈ NC(n) : σ π} ' {S ⊂ σ : Out(σ) ⊂ S} .
The isomorphism is implemented by the map
π = (V1 , V2 , . . . , Vk ) 7→ {o(σ|V1 ), o(σ|V2 ), . . . , o(σ|Vk )} .
In particular, for σ ∈ NC(n),
f {π ∈ NC(n) : σ π} ' {S ⊂ σ : o(σ) ∈ S}.
Proof. See Figure 9 for an illustration. For part (a), let π = (V1 , V2 , . . . , Vk ) and
π0 = {(min(V1 ), max(V1 )), (min(V2 ), max(V2 )), . . . , singletons} .
π0 ∈ NC1,2 (n). Then π0 π and σ π if and only if π0 ≤ σ ≤ π.
For part (b), we need to show that the map is a bijection. We construct the inverse map as follows. Let
Out(σ) ⊂ S ⊂ σ. For V ∈ σ, choose U ∈ S with the largest depth such that V ≤ U (see the illustration).
Since S contains the outer blocks, U exists. In π, adjoin V to U . Get π ∈ NC(n), σ ≤ π, σ π, each
block of π contains a unique element of S.
87
Corollary 9.9. X
t|π|−|Out(σ)| s|σ|−|π| = (t + s)|Inn(σ)| .
π∈NC(n)
σπ
In particular, if σ ∈ NC(n),
f then
X
t|π|−1 s|σ|−|π| = (t + s)|σ|−1 .
π∈NC(n)
σπ
Proof.
X X
t|π|−|Out(σ)| s|σ|−|π| = t|S|−|Out(σ)| s|σ|−|S|
π∈NC(n) Out(σ)⊂S⊂σ
σπ
|Inn(σ)|
X |Inn(σ)| i |σ|−i
= ts = (t + s)|Inn(σ)| .
i=0
i
Bt transformation.
Definition 9.10. For t ∈ R, define the linear map (Belinschi-Nica remarkable transformation)
Bt : D(d) → Chx1 , . . . , xd i∗
by X
B Bt [µ] [xu(1) , xu(2) , . . . , xu(n) ] = t|π|−1 Bπµ [xu(1) , xu(2) , . . . , xu(n) ].
π∈NC(n)
g
88
Let
σ = σ1 ∪ σ2 ∪ . . . ∪ σ|π| .
Then σ ≤ π, σ π, and σ ∈ NC(n).
f So we can continue the expression above as
X X
= (1 + t) t|π|−1 Rσµ
σ∈NC(n)
g σπ
X
= (1 + t) (1 + t)|σ|−1 Rσµ (by Corollary 9.9)
σ∈NC(n)
g
X
= (1 + t)|σ| Rσµ
σ∈NC(n)
g
(1+t)
X
= Rσµ (by definition of )
σ∈NC(n)
g
(1+t)
= Bµ (by Corollary 9.2).
We conclude that
Bt [µ]](1+t) = µ(1+t) .
Proposition 9.14. {Bt : t ∈ R} form a group,
Bt ◦ Bs = Bt+s ,
with the identity element B0 and B−1
t = B−t . In particular, {Bt : t ≥ 0} form a semigroup.
so there is a commutation relation between the Boolean and free convolution powers.
89
Exercise 9.16. Prove that we could also have defined the same Bt in terms of the free cumulants via
X
RBt [µ] [xu(1) , xu(2) , . . . , xu(n) ] = t|π|−1 Rπµ [xu(1) , xu(2) , . . . , xu(n) ].
π∈NC(n)
g
(Combine Remark 9.1, Definition 9.10, Remark 9.11, and Proposition 9.14.)
Φ transformation.
Definition 9.17. Define the linear map
Φ : D(d) → Chx1 , . . . , xd i∗
by
d
X
Φ[µ]
B (z1 , z2 , . . . , zd ) = zi M µ (z1 , z2 , . . . , zd )zi .
i=1
Φ[µ]
Equivalently, B [xi ] = 0, and
90
In other words,
0, β0 β1 β2 . . .
J(Φ[µ]) =
1, γ1 γ2 γ3 . . .
It follows that in this case, the coefficient stripping operation of Exercise 6.50 is the left inverse of Φ.
Exercise 9.19. Prove, from the definition of Φ, that for µ ∈ D(d), the functional B Φ[µ] is conditionally
positive. Conclude that Φ preserves positivity.
Evolution equation.
Theorem 9.20. For any ρ ∈ D(d) and any t ≥ 0,
Bt [Φ[ρ]] = Φ[ρ γt ],
where γt = S(0, t) for d = 1, and in general is the distribution of a free semicircular system with
covariance tI.
RU = R[xu(j) : j ∈ U ].
X
B Bt [Φ[ρ]] = t|π|−1 BπΦ[ρ] [xu(1) , . . . , xu(n) ] (by definition of Bt )
π∈NC(n)
g
X Y
= t|π|−1 δu(min(Vi ))=u(max(Vi )) MVρi \{min(Vi ),max(Vi )} (by definition of Φ)
π∈NC(n)
g Vi ∈π
X Y γ1
= t|π|−1 R{min(Vi ),max(Vi )}
MVρi \{min(Vi ),max(Vi )}
π∈NC(n)
g Vi ∈π
X Y X γ1
Y
= t|π|−1 Ro(σ i)
RUρ
π∈NC(n)
g Vi ∈π σi ∈NC(V
g i) U ∈Inn(σi )
(
γ1 δu(min(Vi )=u(max(Vi )) , |o(σi )| = 2,
(by the moment-cumulant formula and because Ro(σ i)
= )
0, otherwise
91
X Y X γt
Y
= t−1 Ro(σ i)
RUρ (because Rγt = tRγ1 )
π∈NC(n)
g Vi ∈π σi ∈NC(V
g i) U ∈Inn(σi )
X X Y γt
Y
= t−1 Ro(σ| V ) RUρ (by definition of )
i
σ∈NC(n)
g σπ Vi ∈π U ∈Inn(σ|Vi )
X X Y Y
= t−1 RUγt RUρ (by Proposition 9.8)
σ∈NC(n)
g o(σ)∈S⊂σ U ∈S U 6∈S
X γt
Y γ ρ
= t−1 Ro(σ) t
RU + RU
σ∈NC(n)
g U ∈Inn(σ)
X γt
Y
= t−1 Ro(σ) RUργt (by definition of )
σ∈NC(n)
g U ∈Inn(σ)
X γ1
Y
= Ro(σ) RUργt
σ∈NC(n)
g U ∈Inn(σ)
ργt
= δu(1)=u(n) M [xu(2) , . . . , xu(n−1) ] = B Φ[ργt ] .
Thus ρ is the generator of the -semigroup {µt : t ≥ 0}. This calculation shows that any free convolution
semigroup (with finite variance), once stripped, gives a semicircular evolution.
92
Chapter 10
Definition 10.1. For a ∈ (A, ϕ), its ∗-distribution is the joint distribution of (a, a∗ ), a state on Chx, x∗ i.
ϕ[un ] = δn=0 , n ∈ Z.
Write
u~ε = uε(1) uε(2) . . . uε(n) ,
so that each term is either u∗ or u. Then the ∗-distribution is of u is
(
~ε ~ε |{i:ε(i)=·}|−|{i:ε(i)=∗}| 1, |{i : ε(i) = ·}| = |{i : ε(i) = ∗}| ,
µ x =ϕ u =ϕ u =
0, otherwise.
More pictorially, we can identify ~ε with a lattice path, with ∗ corresponding to a rising step (1, 1), and · to
a falling step (1, −1). See Figure 10. Note that the paths start at height zero, but are not necessarily Dyck
paths. In this language, (
~ε 1, path ends at height 0,
µ x =
0, otherwise.
What are their free cumulants? See Lecture 15 of [NS06].
93
94
Example 10.4. Let ` be a free creation operator, `∗ ` = 1. The ∗-distribution of ` is
( (
~ε ~ε 1, can reduce the word `~ε to 1, 1, ~ε corresponds to a Dyck path,
µ` x = ϕ ` = =
0, otherwise. 0, otherwise.
See Figure 10. It follows that the joint free cumulants of `, `∗ are R[`∗ , `] = 1, with the rest all equal to
zero.
Definition 10.5. An operator c is a standard circular element if c = √1 (s1 + is2 ), where s1 , s2 are free
2
standard semicircular.
Example 10.6. Let c be a standard circular element. What is its ∗-distribution? Clearly its free cumulants
are
1
R[c, c∗ ] = R[c∗ , c] = ϕ [(s1 + is2 )(s1 − is2 )] = 1,
2
with the rest all equal to zero. What are the ∗-moments? Using Notation 10.2,
Proof. To show:
ϕ [(c∗ c)n ] = s2n ,
where s is standard semicircular. Indeed, using the non-crossing property (see Figure 10),
Remark 10.8. In fact (Oravecz 2001, Kemp, Speicher 2007, Schumacher, Yan 2012),
Exercise 10.9. Let c1 , . . . , cp are freely independent circular elements. Use the idea from Proposition 10.7
and the preceding remark to compute
n
ϕ c∗p . . . c∗2 c∗1 c1 c2 . . . cp
.
95
Example 10.10. Let f1 ⊥ f2 be unit vectors and X(f1 ), X(f2 ) corresponding q-Gaussian elements from
Proposition 6.55.
1
Y = √ (X(f1 ) + iX(f2 ))
2
is a q-circular element. Using the q-Wick formula (Proposition 6.56),
X
ϕ Y ~ε = q cr(π) .
π∈P 2 (~
ε)
96
Chapter 11
the algebra of random variables all of whose moments are finite. On this algebra we have the expectation
functional Z Z
E[f ] = f dP = x dµf (x),
Ω R
where µf is the distribution of f . For each N , we have the algebra of N × N random matrices
We can identify A ∈ MN (L∞− (Ω, P )) with a N 2 -tuple of random variables, so it has a N 2 -dimensional
distribution. But also: for
N
1 1 X
tr[A] = Tr[A] = Aii
N N i=1
the normalized trace (note: itself a random variable)
(MN (C) ⊗ L∞− (Ω, P ), E ⊗ tr) = (MN (L∞− (Ω, P )), E ◦ tr)
is an n.c. probability space, so A also has a 1-dimensional distribution as an element in this space.
97
98
99
Definition 11.2 (Gaussian random variables). X has the Gaussian (or normal) distribution N (0, t) with
mean zero and variance t if
1 −x2 /2t
dµX (x) = √ e dx.
2πt
(X1 , . . . , Xd ) are jointly Gaussian with mean 0 and (positive definite) covariance matrix Q = (Qij )di,j=1 if
d
!
1 X
dµX1 ,X2 ,...,Xd (x1 , . . . , xd ) = exp − xi (Q−1 )ij xj /2 dx1 . . . dxd .
Z i,j=1
In particular, X1 , . . . , Xd are independent Gaussian if all Qij = 0 for i 6= j (in the jointly Gaussian case,
uncorrelated implies independent).
Exercise 11.4. Show that the (classical) cumulant generating function of a jointly Gaussian distribution
is
d
1X
`(z1 , . . . , zd ) = zi Qij zj .
2 i,j=1
Hint: denote x = (x1 , . . . , xd )t , and the same for y, z. Re-write the jointly Gaussian density as
1 −xt Q−1 x/2
dµ(x) = e dx.
Z
Recall from Remark 5.7 that Z
µ t
` (z) = log ex z dµ(x)
R
(why?). Now use the change of variables x = y + Qz.
where P 2 (n) are the pair partitions. Equivalently, their classical cumulants are
(
0, n 6= 2,
k[Xu(1) , . . . , Xu(n) ] =
E[Xu(1) Xu(2) ], n = 2.
100
Remark 11.6 (Complex Gaussians). A complex-valued random variable Y is complex Gaussian if <Y ,
=Y are independent (centered) Gaussian. We will only consider <Y ∼ N (0, a), =Y ∼ N (0, a), with the
same variance. Note that
E[Y ] = 0
and
E[Y 2 ] = E[(<Y )2 − (=Y )2 + 2i(<Y )(=Y )] = a − a = 0.
However,
E[Y Ȳ ] = E[(<Y )2 + (=Y )2 ] = 2a.
Moreover, if Y1 , . . . , Yd are independent complex Gaussian (i.e. their real and imaginary parts are inde-
pendent), by multi-linearity of classical cumulants all of their joint cumulants are zero except k[Yi , Yi∗ ] =
k[Yi∗ , Yi ] = 2ai .
Corollary 11.7 (Wick formula). Using Notation 10.2, for independent complex Gaussian Y1 , . . . , Yd ,
h i X Y h ε(i) ε(j) i
ε(1) ε(n)
E Yu(1) . . . Yu(n) = E Yu(i) Yu(j) .
π∈P 2 (~
ε) (i,j)∈π
Note that
N
1 X
2
E ◦ tr[X ] = E[Xij Xji ] = 1.
N i,j=1
Since
N
Y 2
Y 2 +(=x )2 )/(1/N )
P 2 2 ]/2
e−xii /(2/N ) e−((<xij ) ij
= e− i,j |xij | /(2/N )
= e−N Tr[X ,
i=1 i<j
101
2
the RN distribution of a GUE matrix is
N
1 −N Tr[X 2 ]/2 Y Y 1 2
= e dxii d<xij d=xij = e−N Tr[X ]/2 dX.
Z i=1 i<j
Z
Exercise 11.11.
a. Show that a d-dimensional Gaussian distribution dµ(x) with the identity covariance matrix (so that
its components are independent and have variance 1) is rotationally invariant, in the sense that
dµ(x) = dµ(Ox)
for any orthogonal d × d matrix O. (Maxwell’s theorem asserts that up to rescaling, these are the
only rotationally invariant distributions with independent components).
b. Show that a GUE matrix is unitarily invariant, in the sense that for any N × N unitary matrix U ,
if X is an N × N GUE matrix, so is U XU ∗ . (You can use the definition, or Remark 11.10). See
Remark 11.48.
Example 11.12. We would like to compute
1 hX i
E ◦ tr[X 4 ] = E Xi(1)i(2) Xi(2)i(3) Xi(3)i(4) Xi(4)i(1) .
N
Apply the Wick formula:
(
1
, if i(j) = i(k + 1), i(j + 1) = i(k),
E Xi(j)i(j+1) Xi(k)i(k+1) = N
0, otherwise.
Draw a ribbon (fat) 4-star (with a marked edge). See Figure 11. Note that these pictures are dual to those
used in topology. The rule above says we match the two edges with the correct orientation. Thus
4 1 X
E ◦ tr[X ] = E Xi(1)i(2) Xi(2)i(3) E Xi(3)i(4) Xi(4)i(1) δi(1)=i(3)
N
i(1),i(2),i(3),i(4)
+ E Xi(1)i(2) Xi(4)i(1) E Xi(2)i(3) Xi(3)i(4) δi(2)=i(4)
+ E Xi(1)i(2) Xi(3)i(4) E Xi(2)i(3) Xi(4)i(1) δi(1)=i(4) δi(2)=i(3) δi(2)=i(1)
1 1 1
= 2
(N 3 + N 3 + N ) = 1 + 1 + 2 .
NN N
What about a general moment E ◦ tr[X 2n ]? It corresponds to a ribbon 2n-star (with a marked edge). See
Figure 11.
In the diagram, identify edges pairwise, in an orientation-preserving way, get a closed surface with a graph
drawn on it. # of vertices = 1. # of edges = n. # of faces (loops) = # free parameters. So for each
π ∈ P 2 (2n), get a term
1 # faces 1
N = N #F −#E−1 .
N Nn
102
Remark 11.13. The Euler characteristic of a connected surface is
χ = #F − #E + #V = 2 − 2g,
Definition 11.15. A map is a connected (ribbon) graph embedded in an (orientable) compact surface in
which all edges are matched in an orientation-preserving way, and each face (loop) is homeomorphic to a
disk. The genus of a map is the genus of a surface into which it is embedded.
E ◦ tr[X 2n ] → cn = ϕ s2n ,
103
Remark 11.19 (Wigner’s theorem). Much stronger statements hold. First, in fact
Second, the assumption that X is GUE can be weakened to assuming it is a Wigner matrix: a self-adjoint
matrix for which
N
1 X 2
E [Xij ] = 0, lim N E |X ij | −1 =0
N →∞ N 2
i,j=1
Third, recall from Exercise 2.23 that for a self-adjoint matrix, its distribution is N1 N
P
i=1 δλi , where {λi }
are its (real) eigenvalues. Similarly, we may identify the distribution of a random matrix as
1
µ̂N = (δλ1 + . . . + δλN ) ,
N
the random spectral distribution of X, which is a random measure. Then in fact, for a Wigner matrix, as
N → ∞,
1√
µ̂N → 4 − x2 dx,
2π
the Wigner semicircle law.
The results go back to the work of (Wigner 1958) and (Dyson 1962) on modeling a heavy nucleus, and are
a (weak) example of universality, which is more properly exhibited in the behavior of eigenvalue spacings
rather than the eigenvalues themselves.
be standard semicircular, respectively, standard circular, all free among themselves. Let X ∈ MN (A)
have entries
1 1 1
Xii = √ si , Xij = √ cij , Xji = √ c∗ij for i < j.
N N N
Prove that for any N , the distribution of XN ∈ (MN (A), ϕ ◦ tr) is (exactly, not asymptotically) standard
semicircular. (Adapt Corollary 11.7 and the argument in Example 11.12 to this setting.)
104
Remark 11.21. We know that
n
X
2n 1
E ◦ tr[X ] = |{maps with 1 vertex, n edges, of genus g}| .
g=0
N 2g
What about
" k #
Y
E N Tr(X n(1) )N Tr(X n(2) ) . . . N Tr(X n(k) ) = E N Tr(X n(i) )
i=1
X
= Nk
E Xi(1)i(2) . . . Xi(n(1))i(1) Xi(n(1)+1)i(n(1)+2) . . . Xi(n(1)+n(2))i(n(1)+1) . . . ?
k stars, of types n(1), . . . , n(k), pair off all edges in orientation-preserving ways. See Figure 11 for a
single term from the expansion of E [N Tr(X 4 )N T r(X 3 )N T r(X 3 )]. Get k vertices, with the number of
faces again equal to the number of free parameters. Apply the Wick theorem, each term in the expression
above is of the form
1
N k #E N #F = N #V −#E+#F .
N
In general, if a map has c components,
c
X
2c − gi = 2c − 2g = #V − #E + #F.
i=1
Thus " #
k ∞ X
∞
Y
n(i)
X 1
E N Tr(X ) = |Gg,c (n(i), 1 ≤ i ≤ k)| ,
i=1 g=0 c=1
N 2g−2c
where
Gg,c (n(i), 1 ≤ i ≤ k) = {union of c maps built out of stars of type n(1), . . . , n(k), of total genus g}
What is the largest power of N ? Unclear.
Remark 11.22. Next,
k
! ∞ k
!n
X X 1 X
E exp ti N Tr[X i ] =E ti N Tr[X i ]
i=1 n=0
n! i=1
∞ Yk
X 1 X n n(i)
=E ti (N Tr[X i ])n(i)
n=0
n! n(1), n(2), . . . , n(k) i=1
n(1)+...+n(k)=n
∞ k n(i)
X X Y ti
=E (N Tr[X i ])n(i)
n=0 n(1)+...+n(k)=n i=1
n(i)!
∞ X
∞ ∞ k n(i)
X 1 X Y ti
= |Gg,c ((i, n(i)), 1 ≤ i ≤ k)| .
g=0 c=1
N 2g−2c i=1
n(i)!
n(1),...,n(k)=0
105
Remark 11.23 (Connected maps). A general combinatorial principle:
X X
Fdiagram = Fπ .
diagrams π∈{partitions of the components of the diagram}
This is the relation between moments and classical cumulants, and we know from Remark 5.7 that these
are also related via
∞ ∞
!
X 1 n
X 1 n
mn z = exp kn z .
n=0
n! n=1
n!
where
a. One can look at maps where every edge has a color. These arise by gluing stars where only edges
of the same color are glued. Their generating functions are matrix integrals in several independent
GUE matrices, since in the Wick formula, only matrices of the same color can be paired.
b. The expansions above are purely formal. Proofs that these are actual asymptotic expansions are
very hard, see (Ercolani, McLaughlin 2003), (Guionnet, Maurel-Segala 2006, 2007). Their results
apply to N12 log E exp (N Tr[V (X)]), where V is a convex potential.
c. Physicists use the term “planar sector of quantum field theory” or “planar approximation” for the
leading term in the expansion.
106
where " #!
n
1 X
exp N Tr −X 2 /2 + ti X i dX
Z i=1
where C is a constant independent of y. Note that differentiating this expression we get equa-
tion 11.2. Hint: Consider deformations µ + εν for a signed measure ν.
b. Show that for the quadratic potential V (x) = x2 /2, the semicircular distribution is a solution of
equationR11.2. Hint: use the real part of the Cauchy transform. Note that care is required in even
defining dµ(x)
y−x
for y real.
Example 11.27. Let X be GUE, and D(1) , . . . , D(n) (for each N ) fixed (non-random) matrices.
E ◦ tr XD(1) . . . XD(n) =?
Consider
E ◦ tr XD(1) XD(2) XD(3) XD(4) .
See Figure 11. According to the Wick formula, this expression is again a sum of three terms. The first
one is
1
E Xi(1)i(2) Di(2)i(2) Xi(2)i(1) Di(1)i(3) Xi(3)i(4) Di(4)i(4) Xi(4)i(3) Di(3)i(1)
N
1 1 X X X
= 2
D i(2)i(2) D i(4)i(4) Di(1)i(3) Di(3)i(1) = tr[D(1) ] tr[D(3) ] tr[D(2) D(4) ].
NN
i(2) i(4) i(1),i(3)
107
Similarly, the second term is tr[D(2) ] tr[D(4) ] tr[D(1) D(3) ]. In the third term,
1 1
E Xi(1)i(2) Di(2)i(3) Xi(3)i(4) Di(4)i(2) Xi(2)i(1) Di(1)i(4) Xi(4)i(3) Di(3)i(1) = 2 tr D(1) D(4) D(3) D(2) .
N N
Note the order!
P : P(n) ,→ Sym(n), π 7→ Pπ ,
Remark 11.29. Recall: a transposition is a 2-cycle (i, j). Sym(n) is generated by transpositions. In fact,
108
Proof. See Figure 11. The result follows from the observation that
γ = (1, 2, . . . , n) ∈ Sym(n)
a. d(α, β) = |α−1 β| is a metric on Sym(n). (In fact, it is a distance in a certain Cayley graph of
Sym(n).)
b. |α| = n − #cyc(α).
c. |αβ| = |βα|.
Proof. For part (a), d(α, α) = |e| = 0, since α−1 ν = (α−1 β)(β −1 ν),
α−1 ν ≤ α−1 β + β −1 ν ,
and
α−1 β = σ1 σ2 . . . σk ⇔ β −1 α = σk . . . σ2 σ1 .
For part (b), by Remark 11.29,
X
n − #cyc(α) = (|V | − 1) ≥ |α| .
V a cycle in α
#cyc(σ1 . . . σk ) ≥ n − k,
and so #cyc(α) ≥ n − |α|. The result follows. Part (c) is equivalent to proving that in general, |α−1 βα| =
|β|, which follows from part (b), since conjugation preserves the cycle structure.
109
Theorem 11.33. (Biane 1997)
|α| + α−1 β + β −1 γ = n − 1.
Equivalently, α and β lie on the same geodesic from e to γ, with α closer to e. Then the map
NC(n) → SNC (n) is a lattice isomorphism.
Then γ = σ1 σ2 . . . σn−1 . Recall that e has n cycles, γ has 1 cycle, and each multiplication by σj changes
the number of cycles by 1. It follows that
#cyc(σ1 . . . σj ) = n − j.
α = σ1 . . . σj , β = σ1 . . . σj . . . σk , γ = σ1 . . . σj . . . σk . . . σn−1 .
K[π] ↔ Pπ−1 γ.
Sketch of proof. See Figure 11. The proof is by induction. For α = σ1 = (i, j), i < j,
On the induction step, multiplying ασj connects two cycles of α, while σj α−1 γ cuts a cycle into two.
110
Remark 11.35. Biane’s embedding above allows us to define noncrossing partitions for other Coxeter
groups, for example noncrossing partitions of type B.
For example,
tr(152)(34) [X1 , X2 , X3 , X4 , X5 ] = tr [X1 X5 X2 ] tr [X3 X4 ] .
Note: this makes sense since tr is cyclically symmetric.
Theorem 11.37. Let X be a GUE matrix, and D(1) , . . . , D(2n) be fixed matrices.
a. X 1
E ◦ tr[X 2n ] = .
N n−#cyc(πγ)+1
π∈P 2 (2n)
b. X 1
E ◦ tr XD(1) XD(2) . . . XD(2n) = trπγ D(1) , D(2) , . . . , D(2n) .
N n−#cyc(πγ)+1
π∈P 2 (2n)
Proof. For (a), using Example 11.12, we only need to show that
We are computing
E Xi(1)i(2) Xi(2)i(3) . . . Xi(2n)i(1) .
The term in the Wick sum corresponding to π pairs together
Xi(j)i(j+1) , Xi(π(j))i(π(j)+1) .
In other words,
i(j) = i(π(j) + 1) = i(γπ(j)), i(j + 1) = i(γ(j)) = i(π(j)).
Thus indeed,
#free parameters = #cyc(γπ) = #cyc(πγ).
For part (b), we write
X
E ◦ tr XD(1) XD(2) . . . D(2n) =
E Xi(1)u(1) Du(1)i(2) Xi(2)u(2) Du(2)i(3) . . . Xi(2n)u(2n) Du(2n)i(1) .
111
From the Wick formula i(j) = u(π(j)), u(j) = i(π(j)). So get
X 1 X (1) (2) (2n)
Du(1)i(2) Du(2)i(3) . . . Du(2n)i(1)
N n+1
π∈P 2 (2n) i(j)=u(π(j))
X 1 X (1) (2) (2n)
= Du(1)u(π(2)) Du(2)u(π(3)) . . . Du(2n)u(π(1))
N n+1
π∈P 2 (2n) u(j)
X 1 X (1) (πγ(1)) (πγ)2 (1)
= Du(1)u(πγ(1)) Du(πγ(1))u(πγπγ(1)) Du((πγ)2 (1))u((πγ)3 (1)) . . .
N n+1
π∈P 2 (2n) u(j)
X 1
Trπγ D(1) , D(2) , . . . , D(2n)
=
N n+1
π∈P 2 (2n)
X 1
trπγ D(1) , D(2) , . . . , D(2n) .
=
N n−#cyc(πγ)+1
π∈P 2 (2n)
Remark 11.38. Note that the leading term in the expansion corresponds to g = 0, but also to π ∈
NC2 (2n). Equivalently,
a. We say that
(N ) (N )
(a1 , . . . , ak ) → (a1 , . . . , ak ) ⊂ (A, ϕ)
in distribution if for each ~u,
h i
(N ) (N ) (N )
ϕN au(1) au(2) . . . au(n) → ϕ au(1) au(2) . . . au(n)
as N → ∞. Equivalently,
h i
ϕN (N ) (N ) (N )
au(1) , au(2) , . . . , au(n) → Rϕ au(1) , au(2) , . . . , au(n) .
R
(N ) (N ) (N )
In this case we say that (a1 , a2 , . . . , ak ) have the asymptotic distribution
112
(N ) (N ) (N )
b. a1 , a2 , . . . , ak are asymptotically free if (a1 , . . . , ak ) ⊂ (A, ϕ) are free. Equivalently,
h i
(N ) (N ) (N ) N →∞
RϕN au(1) , au(2) , . . . , au(n) −→ 0
Remark 11.40. Let X1 , . . . , Xk be entry-wise independent GUE. Say π ∈ P(n) is consistent with
if
π
i∼j ⇒ u(i) = u(j).
See Figure11. Denote such π by P ~u (n). Then as in Theorem 11.37,
X 1
E ◦ tr Xu(1) Xu(2) . . . Xu(2n) = .
N n−#cyc(πγ)+1
π∈P 2,~u (2n)
and
X 1
E ◦ tr Xu(1) D(1) Xu(2) D(2) . . . Xu(2n) D(2n) = trπγ D(1) , D(2) , . . . , D(2n) .
N n−#cyc(πγ)+1
π∈P 2,~u (2n)
(X1 , . . . , Xp ) → (s1 , . . . , sp ),
(X1 , . . . , Xp , D1 , . . . , Dq ) → (s1 , . . . , sp , d1 , . . . , dq ),
Proof. Taking the limit N → ∞ in the preceding remark and using Remark 11.38,
X
E ◦ tr Xu(1) Xu(2) . . . Xu(n) → 1,
π∈NC2,~u (n)
113
which implies that for all π ∈ NC(n)
(
1, π ∈ NC2 (n) ~u-consistent,
Rπ Xu(1) , Xu(2) , . . . , Xu(n) → = Rπ su(1) , su(2) , . . . , su(n) .
0, otherwise.
Similarly, taking D(i) = fi (D1 , . . . , Dq ) in the preceding remark,
lim E ◦ tr Xu(1) f1 (D1 , . . . , Dq )Xu(2) f2 (D1 , . . . , Dq ) . . . Xu(n) fn (D1 , . . . , Dq )
N →∞
X
= lim trπγ [f1 (D1 , . . . , Dq ), f2 (D1 , . . . , Dq ), . . . , fn (D1 , . . . , Dq )]
N →∞
π∈NC2,~u (n)
X
= ϕπγ [f1 (d1 , . . . , dq ), f2 (d1 , . . . , dq ), . . . , fn (d1 , . . . , dq )]
π∈NC2,~u (n)
X ϕ
= MK[π] [f1 (d1 , . . . , dq ), f2 (d1 , . . . , dq ), . . . , fn (d1 , . . . , dq )]
π∈NC2,~u (n)
X ϕ
Rπϕ su(1) , su(2) , . . . , su(k) MK[π]
= [f1 (d1 , . . . , dq ), f2 (d1 , . . . , dq ), . . . , fn (d1 , . . . , dq )]
π∈NC(n)
= ϕ su(1) f1 (d1 , . . . , dq )su(2) f2 (d1 , . . . , dq ) . . . su(n) fn (d1 , . . . , dq ) ,
where we have used Lemma 11.34, properties of free cumulants of semicircular elements, and the results
from Remark 8.1.
Remark 11.42 (Other Gaussian ensembles).
a. GOE (Gaussian orthogonal ensemble): X real symmetric Gaussian. Counts maps on non-oriented
surfaces. The leading term still planar, since any identification of two edges with the opposite
orientation leads to non-zero genus.
b. GSE (Gaussian symplectic ensemble): X certain quaternionic matrices. These all have a common
generalization to β-ensembles, where the real case corresponds to β = 1, complex to β = 2, and
quaternionic to β = 4.
c. Ginibre ensemble: real Gaussian matrices, with no symmetry, all entries independent. Note that
this is typically not a normal matrix. It still has N complex eigenvalues, and its spectral distribution
is a measure on C. Asymptotically,
N
1 X 1
δλi → circular law 1|z|≤1 on C.
N i=1 π
Old result for Gaussian matrices; for Wigner-type matrices, proved by various authors, culminating
in the proof by (Tao, Vu 2010) under optimal conditions. In distribution,
(X, X ∗ ) → (c, c∗ ),
where c is a circular operator, from Definition 10.5. Note that a circular operator is not normal, so
its distribution cannot come from a measure; but its Brown measure is a (rescaled) circular law.
114
Exercise 11.43. The oldest random matrix model (Wishart 1928) arose in statistics, and its relatives
appear in numerous applications, from cell phone communications to medicine. Let (y1 , . . . yK )t be jointly
Gaussian, with the K × K covariance matrix Q. For j = 1, 2, . . . , N (with N ≥ K) let (Yij )K i=1 be
independent samples from this Gaussian distribution, and Y = (Yij ) a K × N random matrix, whose
columns are independent and identically distributed. Then the N × N matrix N1 Y t Y is called a Wishart
matrix. Note that the K × K matrix
N
1 t 1 X
(Y Y )ij = Yik Yjk
N N k=1
is the sample covariance matrix for this distribution, and in particular E N1 (Y Y t ) = Q, and that the
distributions of Y t Y and Y Y t are closely related, although they live in different spaces.
a. Suppose that Q is the identity matrix, so that all entries of Y are independent. Show that
1 t
Y Y = X t P X,
N
where X is a N ×N random matrix all of whose entries are independent N (0, N1 ) random variables,
and P a diagonal N × N matrix with K ones and N − K zeros on the diagonal.
b. Let now both N and K go to infinity, so that K/N → t, 0 ≤ t ≤ 1. According to the preceding
remark, (X, X t ) → (c, c∗ ). Use a variation of Theorem 11.41 and a combination of Proposition 10.7
with Lemma 8.2 to conclude that the asymptotic distribution of a Wishart matrix with identity
covariance is πt , the free Poisson distribution, which in this context is called the Marchenko-Pastur
distribution.
A similar argument for general Q shows that the limiting distribution of a general Wishart matrix is
free compound Poisson. An alternative approach of rectangular free probability (which corresponds to
taking limits of rectangular matrices directly) was developed by Benaych-Georges. Compare also with
Exercise 22.18 in [NS06].
A random N × N unitary matrix whose distribution is the Haar measure is a Haar unitary matrix. Note
that for any λ ∈ C, |λ| = 1, if U is a random unitary matrix whose distribution is the Haar measure, then
Z Z Z
n n n
U dU = (λU ) d(λU ) = λ U n dU,
UN UN UN
115
and so Z
U n dU = δn=0 , n ∈ Z.
UN
In particular, in the n.c. probability space (MN (L∞− (Ω, P )), E ◦ tr), a Haar unitary matrix is a Haar
unitary.
In physics, Haar unitary matrices are called CUE (circular unitary ensemble) matrices.
Theorem 11.45. Let U1 , . . . , Up be independent N × N Haar unitary matrices, and (D1 , . . . , Dq ) non-
random matrices converging in ∗-distribution to (d1 , . . . , dq ) ⊂ (A, ϕ). Then in ∗-distribution,
(U1 , . . . , Up , D1 , . . . , Dq ) → (u1 , . . . , up , d1 , . . . , dq ),
Remark 11.46. Recall from Exercise 3.12, that if u1 , . . . , up , d0 , d1 , . . . , dp are as above, then
d0 , u1 d1 u∗1 , . . . , up dp u∗p
are free.
Remark 11.48.
c. If X is any random matrix, and U is a Haar unitary matrix, then U ∗ XU is unitarily invariant (by
invariance of the Haar measure). Thus the corollary says “independent unitarily invariant matrices
are asymptotically free”.
116
d. In particular, asymptotic freeness for unitarily invariant matrices implies asymptotic freeness for
Gaussian matrices (Theorem 11.41). Conversely, if X is GUE, its polar decomposition is X =
U (X ∗ X)1/2 , where U ia a Haar unitary matrix. So we could use the Gaussian result to prove the
unitary result. We will use combinatorics instead.
e. GUE are the only matrices which are at once Wigner matrices and unitarily invariant.
(ii) In general, eigenvalues of A + B depend on A, B and not just on {λi }, {ρj }. They satisfy the Horn
inequalities.
(iii) If A, B are in a “general position”: so that B is fixed (without loss of generality, diagonal) and A is
random,
λ1
A=U
.. ∗
U ,
.
λN
U a Haar unitary matrix. Then if
N N
1 X 1 X
µA = δλ , µB = δρ ,
N i=1 i N i=1 i
then µA+B ≈ µA µB for large N . Thus µA+B (asymptotically) depends only on {λi }, {ρj }.
Remark 11.50. We will not prove Theorem 11.45. To prove Corollary 11.47, we need to understand
expressions of the form
E ◦ tr U A(1) U ∗ B (1) U A(2) U ∗ B (2) . . . ,
where A(i) = Au(i) , B (i) = B v(i) . What is the analog of the Wick formula?
E Ui(1)j(1) . . . Ui(n)j(n) Ūu(1)v(1) . . . Ūu(n)v(n)
X
= δi(β(1))=u(1) . . . δi(β(n))=u(n) δj(α(1))=v(1) . . . δj(α(n))=v(n) Wg(N, βα−1 ).
α,β∈Sym(n)
117
Here for β ∈ Sym(n) and N ≥ n
Wg(N, β) = E U11 . . . Unn Ū1β(1) . . . Ūnβ(n)
Example 11.52.
E U11 U12 Ū22 Ū12 = 0,
E U11 U12 Ū12 Ū11 = Wg(N, (12)) + Wg(N, e),
where for the first term α = e, β = (12), for the second α = β = (12).
Remark 11.53. There is no “useful” formula for Wg. Can write it in terms of characters. Can write it in
terms of an inverse of an explicit matrix. It is known that Wg(N, α) only depends on the conjugacy class
(cycle structure) of α, and
1 1
Wg(N, α) = µ(α) 2n−#cyc(α) + O .
N N 2n−#cyc(α)+2
118
Remark 11.54. Thus we can now compute
E ◦ tr U A(1) U ∗ B (1) U A(2) U ∗ B (2) . . . U A(n) U ∗ B (n)
1 X
Wg(N, βα−1 ) Trα A(1) , . . . , A(n) Trβ −1 γ B (1) , . . . , B (n)
=
N
α,β∈Sym(n)
1 X −1 γ)
Wg(N, βα−1 )N #cyc(α)+#cyc(β trα A(1) , . . . , A(n) trβ −1 γ B (1) , . . . , B (n) .
=
N
α,β∈Sym(n)
1
Using the Weingarten function asymptotics, the degree of N
in the leading term is
(2n − #cyc(α−1 β)) + 1 − #cyc(α) − #cyc(β −1 γ)
= 2n + 1 − (n − α−1 β ) − (n − |α|) − (n − β −1 γ )
= |α| + α−1 β + β −1 γ − (n − 1).
Recall that |α| + |α−1 β| + |β −1 γ| ≥ n − 1, with equality only if α = Pσ , β = Pπ , α ≤ β, or equivalently
σ ≤ π. Thus as N → ∞, the leading term in the expression above is
X
−1
(1) (n)
(1) (n)
1
µ(Pσ Pπ ) trσ A , . . . , A trK[π] B , . . . , B +O .
N2
σ,π∈NC(n)
σ≤π
Corollary 11.55. Let (A(1) , . . . , A(n) ), (B (1) , . . . , B (n) ) be N × N (random or not) matrices such that
(A(1) , . . . , A(n) ) → (a1 , . . . , an ) ∈ (A, ϕ), (B (1) , . . . , B (n) ) → (b1 , . . . , bn ) ∈ (A, ϕ),
with tracial state ϕ, and UN be an N × N Haar unitary matrix. Then as N → ∞,
X
E ◦ tr U A(1) U ∗ B (1) . . . U A(n) U ∗ B (n) → µ(Pσ−1 Pπ )ϕσ [a1 , . . . , an ] ϕK[π] [b1 , . . . , bn ] .
σ,π∈NC(n)
σ≤π
119
Sketch of the proof of Corollary 11.47. Say AN → a, BN → b in distribution. To prove asymptotic free
independence of UN AN UN∗ and BN , we want to show that for any fj , gj with ϕ[fj (a)] = ϕ[gj (b)] = 0,
the expression (see text)
E ◦ tr [f1 (UN AN UN∗ )g1 (BN ) . . . gn−1 (BN )fn (UN AN UN∗ )]
= E ◦ tr [UN f1 (AN )UN∗ g1 (BN ) . . . gn−1 (BN )UN fn (AN )UN∗ ]
Using Exercise 11.56 below, we conclude that each term in the sum above involves either ϕ[fi (a)] or
ϕ[gj (b)], and so equals zero.
Exercise 11.56.
a. For any σ ∈ NC(n), the partition σ ∪ K[σ] ∈ NC(2n) has exactly two outer blocks.
b. For any outer block of a non-crossing partition, there is an interval block covered by it.
120
Chapter 12
This chapter was omitted in the course, so what follows is only an outline. See [Spe98], [NSS02].
121
Remark 12.5 (Motivation).
Note: involves only the symmetric moments E[XbXb . . . bX], not general moments
E[Xb1 Xb2 . . . bn−1 X].
To get full information from generating functions, need fully matricial constructions (as studied by
Voiculescu 2004, 2010, Belinschi, Popa, Vinnikov 2010, 2011, 2012, Helton, Klep, McCullough 2009,
2010, 2011). No analog of Stieltjes inversion.
122
Definition 12.10. Subalgebras A1 , . . . , An ⊂ (A, B, E) are freely independent (with amalgamation) over
B if whenever u(1) 6= u(2) 6= . . . 6= u(k), ai ∈ Au(i) , E[ai ] = 0 (in B), then
E [a1 . . . ak ] = 0.
Remark 12.11 (Free product with amalgamation). (A, B, E) = ∗ni=1 (Ai , B, Ei ). Denote
A◦i = {a ∈ Ai : Ei [a] = 0} .
Let W∅ = B,
W~u = A◦u(1) ⊗B A◦u(2) ⊗B . . . ⊗B A◦u(k)
Then ∞
M M
A= W~u ,
k=0 |~
u|=k
u(1)6=u(2)6=...6=u(k)
Proof. One direction is clear. Assume E is a positive B-bimodule map. Let A ∈ Mn (A). Without loss of
generality A = B ∗ B, B ∈ Mn (A), i.e.
n
X
Aij = b∗ki bkj , buv ∈ A.
k=1
so ∀ b1 , . . . , bn ∈ B,
n
"
n n
#
X X X X
b∗i E[b∗ki bkj ]bj = E (bki bi )∗ (bkj bj ) ≥ 0 in B.
k=1 i=1 j=1
123
Theorem 12.14. Free product of completely positive maps is completely positive.
by X
M [b0 X1 b1 , X2 b2 , . . . , Xn bn ] = Rπ [b0 X1 b1 , X2 b2 , . . . , Xn bn ].
π∈NC(n)
Here Rπ uses the nesting structure on NC(n). In particular, this will not work for general partitions.
Rπ [b0 X1 b1 , X2 b2 , X3 b3 , X4 b4 , X5 b5 , X6 b6 , X7 b7 , X8 b8 ]
= b0 R[X1 b1 R[X2 b2 , X3 b3 R[X4 ]b4 X5 ]b5 X6 ]b6 R[X7 b7 , X8 ]b8 .
Theorem 12.16. Free independence is equivalent to the vanishing of mixed free cumulants.
is a completely positive map R2X : B → B. Thus: let η : B → B be a completely positive map. The
functional µ : Bhxi → B defined by
(
η(b1 ), n = 2,
Rµ [x, b1 x, b2 x, . . . , bn−1 x] =
0, n 6= 2
124
Proposition 12.18. µ is positive.
h·, ·i : M × M → B
such that
hb1 f b2 , gb3 i = b∗3 hf, b∗1 gi b2
(compare with Remark 12.8),
hf, gi∗ = hg, f i ,
hf, f i ≥ 0 in B.
The full Fock module F(M) is the B-bimodule
∞
M
F(M) = M⊗B k = B ⊕ M ⊕ (M ⊗B M) ⊕ (M ⊗B M ⊗B M) ⊕ . . . ,
k=0
Remark 12.20. If the inner product on M is positive, it is completely positive, which implies that the
inner product on F(M) is positive.
125
Now specialize: Let M = principal B-bimodule, M = BξB, ξ a symbol. Then
`(b1 ) = η(b1 )1 ∈ B. If we define X = `∗ + `, then with respect to the conditional expectation h·1, 1i, X
has the B-valued semicircular distribution with variance η.
Convolution powers.
Definition 12.21. µ : Bhxi → B is -infinitely divisible if for all t ≥ 0, tRµ = Rµt for some positive µt .
In that case write µt = µt . But remember: Rµt ∈ B. So can ask: if η : B → B is completely positive, is
there a µη such that
Rµη = η(Rµ )?
Theorem 12.22.
b. If µ is -infinitely divisible (i.e. µt ≥ 0), then automatically µη ≥ 0 for any c.p. η.
Proof. (a), (b) proved using a Fock module construction. For (c), Method I (A, Belinschi, Fevrier, Nica
2011) is to define the map Bη combinatorially as in Definition 9.10, prove that it is c.p., and define
Method II (Shlyakhtenko 2012) is to use an operator-valued Voiculescu’s operator model from Proposi-
tion 6.20.
126
Relation between different operator-valued cumulants.
If D ⊂ B ⊂ A, what is the relation between D and B-valued free cumulants? Between D and B-valued
freeness?
Theorem 12.23. Let 1 ∈ D ⊂ B ⊂ A, (A, B, E), (B, D, E) n.c. probability spaces. Let a1 , . . . , ak ∈ A.
Suppose that for all d1 , . . . , dn−1 ∈ D,
Then
RB au(1) d1 , au(2) d2 , . . . , au(n) = RD au(1) d1 , au(2) d2 , . . . , au(n)
Theorem 12.24. Let 1 ∈ D ⊂ B, N ⊂ A, (A, B, E), (B, D, E) n.c. probability spaces. Suppose E :
B → D is faithful, i.e.
∀b2 ∈ B, E[b1 b2 ] = 0 ⇒ b1 = 0.
Then the following are equivalent.
Remark 12.25 (Applications). Suppose X, Y ∈ (A, ϕ) are not free. How to compute µX+Y ? Sometimes
have (A, B, E), ϕ = ϕ ◦ E, and X, Y are free with respect to E. Then can compute
µBX+Y : Bhxi → B
127
Chapter 13
Proof.
−=z
Z Z
1
=Gµ (z) = = dµ(x) = dµ(x).
R z−x R |z − x|2
Remark 13.2. Gµ is invertible in a Stolz angle
z ∈ C+ : |z| > α, =z > β |<z| .
G−1
µ is defined on
z ∈ C+ : |z| < α, =z > β |<z| .
Rµ + Rν = Rµν
Exercise 13.4. Let {γt : t ≥ 0} be the free convolution semigroup of semicircular distributions; recall
that Rγt (z) = tz. Let µ be an arbitrary probability measure, and denote
What is the corresponding result for a general free convolution semigroup {νt : t ≥ 0}?
Or: Let µ, ν be probability measure on T, X, Y freely independent, unitary operators with distributions
µ, ν. Define
µ ν = µXY .
How to compute? Define
Z
xz
ψµ (z) = dµ(x) = Mµ (z) − 1,
R+ or T 1 − xz
1+z
Sµ (z) = χµ (z).
z
Then
Sµν (z) = Sµ (z)Sν (z).
But note: constructed a multivariate R-transform. No good multivariate S-transform, although there are
attempts.
129
Example 13.5. Let
µ = (1 − α)δ0 + αδ1 , ν = (1 − β)δ0 + βδ1 .
Recall: if p, q are projections, ϕ[p] = α, ϕ[q] = β, in Proposition 8.4 computed the distribution of pqp
using R-transforms. Using S-transforms:
z z 1+z
ψµ (z) = α , χµ (z) = , Sµ (z) = .
1−z α+z α+z
Thus
(1 + z)2
Sµν (z) = ,
(α + z)(β + z)
z(1 + z)
χµν (z) = ,
(α + z)(β + z)
p
1 − z(α + β) + (az − 1)(bz − 1)
ψµν (z) = ,
2(z − 1)
where p
a, b = α + β − 2αβ ± 4αβ(1 − α)(1 − β).
1 1
Then G(z) = z
1+ψ z
gives
p
1 z − (α + β) + (z − a)(z − b)
Gµν (z) = + ,
z 2(1 − z)z
and can find µ ν using Stieltjes inversion.
where ρ is a finite positive measure. This formula still holds for general µ with finite variance (Kol-
mogorov representation). In general, have a free Lévy-Khinchin representation
Z
z+x
Rµt (z) = t λ + dρ(x) .
R 1 − xz
130
It follows that n o
ID = νλ,ρ : λ ∈ R, ρ finite Borel measure .
a. µn ∗ µn ∗ . . . ∗ µn → ν∗λ,ρ weakly.
| {z }
kn
b. µn µn . . . µn → νλ,ρ weakly.
| {z }
kn
Recall however, that there does not exist a bijection T such that
T (µ ∗ ν) = T (µ) T (ν).
The theorem above has many generalizations, to ], , non-identically distributed case, etc.
An alternative description of -infinitely divisible distributions is that they are precisely those µ whose
R-transforms Rµ extend to an analytic function C+ → C+ .
Stable distributions
µ is -stable if µ µ is µ up to a shift and rescaling. For example, for µ semicircular,
µ µ = D√2 µ.
Have a complete description. Again, -stable distributions are in a bijection with ∗-stable distributions,
and their domains of attraction coincide.
Except for Gaussian/semicircular, stable distributions do not have finite moments.
Example 13.7. Define µ by Rµ (z) = α, α ∈ C. Note µ µ = D2 µ, so µ is stable.
1 1
G−1
µ (w) = + α, Gµ (z) = .
w z−α
If α is real, then µ = δα . But if α is purely imaginary, α = ia, then
1 a − =z
= =
z − ia (<z)2 + (=z − a)2
131
and
1 a
dµ(x) = dx.
π x + a2
2
These are the Cauchy distributions. They have a very special property that
µ µ = µ ∗ µ;
Remark 13.9. In this remark we give a heuristic derivation of the classical entropy formula. Let µ be a
measure with density f , and let F be its cdf, i.e. the primitive of f . Assume that µ is supported in an
interval [a, b]. Fix n, and for 0 ≤ i ≤ n let ai = F −1 (i/n). Now look at all the atomic measures
1
(δx + δx2 + . . . + δxn )
n 1
with ai−1 ≤ xi ≤ ai .
On the one hand, both the k-th moment of this measure and of µ lie
" n−1 n
#
1X k 1X k
a , a .
n i=0 i n i=1 i
Therefore the difference between them is at most (1/n)(akn − ak0 ) = (1/n)(bk − ak ). So for a fixed k, this
can be made arbitrarily small by choosing sufficiently big n.
On the other hand, the volume of the set of these atomic measures is
n n
Y Y
−1 i −1 i−1
Vn = (ai − ai−1 ) = F −F
i=1 i=1
n n
132
and so its logarithm is
n
X
−1 i −1 i−1
log F −F
i=1
n n
Now
−1 i −1 i−1 1 −1 0 i
F −F ≈ (F ) ,
n n n n
so the above expression is approximately
n
X
−1 0 i
log Vn ≈ log((F )) − n log n.
i=1
n
Γ(X1 , . . . , Xn ; m, N, ε)
= A1 , . . . , An ∈ MNsa×N (C) : τ Xu(1) . . . Xu(p) − tr Au(1) . . . Au(p) < ε for p ≤ m
(cutoff omitted for simplicity). The free entropy of (X1 , . . . , Xn ) with respect to τ is
1 n
χ(X1 , . . . Xn ) = inf inf lim sup log vol Γ(X1 , . . . , Xn ; m, N, ε) + log N .
m∈N ε>0 N →∞ N2 2
Remark 13.12. Is Γ(X1 , . . . , Xn ; m, N, ε) always non-empty? The question is equivalent to the Connes
embedding problem.
Remark 13.13. If could replace the lim supN →∞ by limN →∞ , many results would follow.
Remark 13.14 (Properties of χ).
133
a. (One-variable formula) If µ = µX ,
ZZ
3 1
χ(X) = log |s − t| dµ(s) dµ(t) + + log 2π,
4 2
minus logarithmic energy.
c. (Subadditivity)
134
Remark 13.19 (Major applications).
a. Absence of Cartan subalgebras (Voiculescu 1996): if M has a Cartan subalgebra, then δ(M) ≤ 1.
So L(Fn ) does not come from a measurable equivalent relation.
b. Primality (Ge 1998): if M ' M1 ⊗M2 for infinite-dimensional M1 , M2 , then δ(M) ≤ 1. So L(Fn )
is prime.
135
13.3 Operator algebras.
The description in this section is very brief and imprecise, and is meant to give only a very vague idea of
the field. It should not be used as a reference.
The von Neumann algebras of all countable, amenable groups with the infinite conjugacy class property
are the same, namely the hyperfinite II1 -factor. Similarly,
W ∗ (M2 (C), tr)⊗∞ ' W ∗ (M3 (C), tr)⊗∞ .
?
This is why a positive answer to the isomorphism problem for L(F2 ) ' L(F3 ) is not unreasonable.
Exercise 11.20 allows us to combine free semicircular and circular operators into a single semicircular
matrix, while the results from Lecture 14 of [NS06] allow us to cut up a semicircular element by a family
of matrix units.
Definition 13.23. If M is a a II1 -factor, p ∈ M a projection with τ [p] = α, the amplification of M is
Mα = pMp.
Mα depends only on α and not on p. Can do this also for larger α.
Theorem 13.24. If N ≥ 2, n ≥ 2, L(Fn ) ' MN (C) ⊗ L(FN 2 (n−1)+1 ), so that
(L(Fn ))1/N ' L(FN 2 (n−1)+1 ).
Using this idea, Dykema and Rădulescu independently defined interpolated free group factors:
{L(Ft ) : t ∈ (1, ∞)} , L(Ft ) ∗ L(Fs ) ' L(Ft+s ).
Isomorphism problem: not solved. But know: either all L(Ft ), 1 < t ≤ ∞ are the same, or they are all
different.
Theorem 13.26 (Dykema). For n, k ≥ 2,
Mn (C) ∗ Mk (C) ' L(Fs ),
where (1 − n12 ) + (1 − k12 ) = s. More generally, formulas for free products of finite dimensional and even
hyperfinite algebras. Even with amalgamation.
136
Bibliography
137
[PV10] Mihai Popa and Victor Vinnikov, Non-commutative functions and non-commutative free Levy-
Hincin formula, arXiv:1007.1932v2 [math.OA], 2010.
[Spe98] Roland Speicher, Combinatorial theory of the free product with amalgamation and operator-
valued free probability theory, Mem. Amer. Math. Soc. 132 (1998), no. 627, x+88.
MR2266879 (2008k:46198)
[Sta97] Richard P. Stanley, Enumerative combinatorics. Vol. 1, Cambridge Studies in Advanced Math-
ematics, vol. 49, Cambridge University Press, Cambridge, 1997, With a foreword by Gian-
Carlo Rota, Corrected reprint of the 1986 original. MR1442260 (98a:05001)
[VDN92] D. V. Voiculescu, K. J. Dykema, and A. Nica, Free random variables, CRM Monograph Series,
vol. 1, American Mathematical Society, Providence, RI, 1992, A noncommutative probability
approach to free products with applications to random matrices, operator algebras and har-
monic analysis on free groups. MR1217253 (94c:46133)
[Vie84] Gérard Viennot, Une théorie combinatoire des polynômes orthogonaux généraux, Univ. Que-
bec, Montreal, Que. (unpublished notes), 1984.
[Vie85] , A combinatorial theory for general orthogonal polynomials with extensions and appli-
cations, Orthogonal polynomials and applications (Bar-le-Duc, 1984), Lecture Notes in Math.,
vol. 1171, Springer, Berlin, 1985, pp. 139–157. MR838979 (87g:42046)
[Voi02] Dan Voiculescu, Free entropy, Bull. London Math. Soc. 34 (2002), no. 3, 257–278.
MR838979 (87g:42046)
138