In a coming paper I will use Kac-Rice formula to study a question motivated by quantum information. So I decided it’d be a nice time to try to give a short account on the well known formula.
First a disclaimer, I am certainly not an expert on the specifics of the formula or the random geometry aspects that are tied to it. However, I’ve learnt a good amount of beautiful and enjoyable mathematics while working with it, so that I’d like to share a few very simple remarks about it.
What is Kac-Rice formula?
Let’s say with
is a (nice, smooth enough) process. Kac-Rice formula lets you access the average number of level crossings of this process. That is, for
, and if one denotes
Kac-Rice formula states, when it is defined,
is the Jacobian and
is the density of the gradient of
evaluated at
(which needs to exist and be continuous around zero).
In fact, it is much more general (see the Book of Adler and Taylor « Random fields and geometry » for technical aspects and a wonderful amount of beautiful geometry and probability). Let be an
dimensional Riemannian manifold with metric
, let
,
be two (nice and gentle enough) processes over
. Set, for (Borel)
,
, so that in words
is the number of level crossing by
given that
belongs to
. Well, then Kac-Rice formula on manifolds gives
What is it used for?
Of course this questions has many answers. Certainly, a bit of generalities can’t hurt. The story started in the 40’s, with the work of M. Kac and S.O. Rice. Kac worked out the formula to study the number of real roots of a random polynomial and published his work in 1943. On the other hand Rice was interested in the « singing of transmission lines ».
A big part of nowadays use of Kac-Rice formula is to estimate the number of critical points of a random function, with the underlying idea that random functions are good representatives of complicated functions. For instance, the hamiltonian of a glassy system, being a disordered system, will typically be complicated to describe and have a lot of metastable states. Similarly the cost function used in machine learning (ML), will be complicated and have many local maxima, minima and low/high index critical points. Understanding the number of critical points at different value of the energy (for glassy systems) or of the cost (ML) can tell us whether their is a chance at optimizing the function in a reasonable amount of time (where optimizing will mean reaching equilibrium in the context of physical systems and finding the « best » model in ML).
Easy examples
Let’s have a look at how to use the Kac-Rice formula on simple examples. Some of those examples were discussed with Guillaume Dubach and Camille Male during an informal and improvised presentation of Kac-Rice formula.1
Law of critical points of random degree 2 polynomials: Consider the following simple but still illustrative example. Set the random polynomial whose coefficients
are independent standard normal (Gaussian, centered, with variance one). Here I ask what is the average number of critical points of
such that
. There is an obvious way of doing that, which I do not discuss right now. I want to use Kac-Rice formula. So what I am interested in is really the number of
-level crossing of the derivative
over some interval
. That is the count of points
such that
. Of course, if
, the number of critical points is almost surely 1.
According to Kac-Rice formula we have To compute the average we need the density of the derivative at zero. This is easily established by noticing that
is a Gaussian random variable with variance
.
We also need the law of conditioned to the random variable
. Thanks to the fact that both
are independent Gaussian random variables, this is easily obtained, by computing the covariance of
:
and from there just computing the usual conditional covariance formula for , which leads to
Indeed, this is sufficient to determine the conditional law, as the conditioning only imposes linear relations between Gaussian random variables, so that the resulting conditional random variables are Gaussian and centered. So
Simplifying the problem a bit, we now let . Doing so leaves us with
Realizing that the density of the derivative at 0 is just , and putting everything together, we obtain from Kac-Rice formula that
In fact, this says that the critical point of a degree two polynomial is distributed according to Cauchy’s law of location parameter 0 and scale parameter .
This can be easily confirmed by realizing that the critical point of is almost surely given by
, which is a ratio of two independent centered Gaussian random variables and is known to give a Cauchy random variable.
Eigenvalues of the Gaussian Orthogonal Ensemble: We now study another simple example. We want to recover the eigenvalue distribution of the GOE from the Kac-Rice integral. This is a good example as it is a baby version of an important use case for Kac-Rice: the spherical p-spin model. Our example is the (almost) trivial case.
First a reminder. Let be a
real symmetric matrix. Then Courant-Fisher theorem tells us that the eigenvalues of
are the values of the function
at its critical points, for
. In fact since if
is a critical point, then
is another and
, counting critical points of
will lead us to double count the eigenvalues (one way to avoid that would be to work on the projective space instead of the unit vector sphere).
Keeping the above remark in mind, we now consider to be a $N\times N$ GOE matrix. Thas is elements
are all independent, and
.
According to Kac-Rice, up to a factor of 2, the average number of eigenvalues in the interval is given by
A first remark is that will not depend on
as the matrices
, for any orthogonal transformation
. So we will compute this average at
and
will just contribute the volume of the
dimensional sphere.
As earlier, we need the joint law of . We compute the joint law in a chart around
given by
using at our advantage that the covariant derivatives around
coincide with the
. This allows us to compute correlation using the remark that
. We find,
This allows us to identifiy the joint law of the joint law of noticing that the gradient is independent from
and that conditioned on
one has
where
is a
random GOE matrix,
while
.
Therefore according to Kac-Rice formula, the number of critical point of is given by
Using a classical trick, we can recognize, up to multiplicative constant that is the distribution function of eigenvalues of
by noticing that one can interpret
as the
eigenvalues of
(where the
‘s are the remaining
. This gives a good check of our Kac-Rice formula, as computing the average number of critical points of the random bilinear form
recovers the distribution of eigenvalues of
as should be expected from Courant-Fisher.
- I thank them for their patience, their nice remarks and their general curiosity ↩︎