T-I: Difference between revisions

From Disordered Systems Wiki
Jump to navigation Jump to search
 
(463 intermediate revisions by the same user not shown)
Line 1: Line 1:
This session of problems is entirely devoted to the Random Energy Model.  
<strong>Goal: </strong> understanding the energy landscape of the simplest spin-glass model, the Random Energy Model (REM). <br>
<strong> Techniques: </strong> probability theory.
<br>




<br>
== Large-N disordered systems: a probabilistic dictionary ==
<br>
<ul>


=== Problem 1: the energy landscape of the REM ===
<li> <ins>'''Notation.'''</ins> We will discuss disordered systems with <math> N </math> degrees of freedom (for instance, for a spin system on a lattice of size <math>L</math> in dimension <math> d</math>, <math> N = L^d </math>). The disorder can have various sources: (e.g., the interactions between the degrees of freedom are random). Since the systems are random, the quantities that describe their properties (the free energy, the number of configurations of the system that satisfy a certain property, the magnetization etc) are also random variables, with a distribution. In this discussion we denote these random variables generically with <math> X_N, Y_N </math> (where the subscript denotes the number of degrees of freedom) and with <math> P_{X_N}(x), P_{Y_N}(y) </math> their distribution. We denote with <math> \mathbb{E}({\cdot})</math> the average over the disorder (equivalently, over the distribution <math> P_{X_N}(x) </math> of the quantity, induced by the disorder). Statistical physics goal is to characterize the behavior of these quantities in the limit <math> N \to \infty</math>. In statistical physics, in general one deals with exponentially scaling quantities:
<math display="block">
X_N = e^{N Y_N}
</math>
</li>
<br>
 
<li> <ins>'''Typical and rare.'''</ins> The typical value of a random variable is the value at which its distribution peaks (it is the most probable value). Values at the tails of the distribution, where the probability density does not peak but it is small (for instance, vanishing with <math> N \to \infty </math>) are said to be rare. In general, average and typical value of a random variable may not coincide: this happens when the average is dominated by values of the random variable that lie in the tail of the distribution and are rare, associated to a small probability of occurrence.
<br>
<br>
<code>Short exercise. </code> Consider a random variable <math> X </math> with a log-normal distribution, meaning that <math> X=e^Y </math> where <math> Y </math> is a random variable with a Gaussian distribution with average <math> \mu </math> and variance <math> \sigma </math>. Show that  <math> X^{typ}=e^\mu </math> while  <math> \mathbb{E}({X})=e^{\mu+\frac{\sigma^2}{2}} </math>, and thus the average value is larger than the typical one.
</li>
<br>


We begin by characterizing the energy landscape of the REM. In particular, in this exercise we study the number <math> \mathcal{N}(E)dE </math> of configurations having energy  <math> E_\alpha \in [E, E+dE] </math>. This quantity is a random variable. For large <math> N </math>, we will show that its <em> typical value </em> scales as


<center><math>
<li> <ins>'''Self-averagingness.'''</ins> The physics of disordered systems is described by quantities that are distributed when <math> N </math> is finite: they take different values from sample to sample of the system, where a sample corresponds to one "realization" of the disorder. However, for relevant quantities sample to sample fluctuations are suppressed when <math> N \to \infty</math>. These quantities are said to be <ins>self-averaging </ins>.
  \mathcal{N}(E) = e^{N \Sigma\left( \frac{E}{N}\right) + o(N)}, \quad \quad \Sigma(\epsilon)
A random variable <math> Y_N </math> is self-averaging  when, in the limit <math> N \to \infty </math>, its distribution concentrates around the average, collapsing to a deterministic value:
=
\begin{cases}
\log 2- \epsilon^2 \quad &\text{ if } |\epsilon| \leq \sqrt{\log 2} \\
0 \quad &\text{ if } |\epsilon| >\sqrt{\log 2}
\end{cases}
</math></center>


The function  <math> \Sigma(\epsilon) </math> is the entropy of the model, and it is sketched in Fig. X. The point where the entropy vanishes, <math> \epsilon=- \sqrt{\log 2} </math>, is the energy density of the ground state, consistently with what we obtained in the lecture. The entropy is maximal at <math> \epsilon=0 </math>: the highest number of configurations have vanishing energy density.
<math display="block">
\lim_{N \to \infty} Y_N =\lim_{N \to \infty}  \mathbb{E}({Y_N}):= {Y_\infty}, \quad \quad \mathbb{E}({Y_N})=\int \, dy\, P_{Y_N}(y)\, y
</math>




<ol>
This happens when its fluctuations are small compared to the average, meaning that <sup>[[#Notes|[*] ]]</sup>
<li> <em> Averages: the annealed entropy.</em> We begin by computing the “annealed" entropy <math> \Sigma^A </math>, which is the function that controls the behaviour of the average number of configurations at a given energy, <math> \overline{\mathcal{N}(E)}= \text{exp}\left(N \Sigma^A\left( \frac{E}{N} \right)+ o(N)\right) </math>. Compute this function using the representation <math> \mathcal{N}(E)dE= \sum_{\alpha=1}^{2^N} \chi_\alpha(E) dE \;</math>   [with <math> \chi_\alpha(E)=1</math> if <math> E_\alpha \in [E, E+dE]</math> and  <math> \chi_\alpha(E)=0</math> otherwise], together with the distribution  <math> p(E)</math> of the energies of the REM configurations. When does <math> \Sigma^A </math> coincide with the entropy defined above (which we define as the “quenched” entropy in the following)?</li>
<math display="block">
</ol>
\lim_{N \to \infty} \frac{\mathbb{E}(Y_N^2)}{\mathbb{E}(Y_N)^2}=1.
</math>
For self-averaging quantities, in the limit <math> N \to \infty </math> the distribution collapses to a single value <math> {Y_\infty}</math>, that is both the limit of the average and typical value of the quantity: <ins><em> sample-to-sample </em> fluctuations are suppressed</ins> when <math>N </math> is large. When the random variable is not self-averaging, it remains distributed in the limit <math> N \to \infty </math>.
</li>
<br>
 
 
<code>Example 1. </code> Consider the partition function of a disordered system at inverse temperature <math> \beta</math>,  <math> Z_N(\beta) </math>. When <math> N </math> is large this random variable has an exponential scaling, <math> Z_N(\beta) \sim e^{-\beta N f_N(\beta)} </math>, where the variable <math> f_N(\beta) </math> is the free energy density. This scaling means that the random variable <math> \beta f_N=-N^{-1}\log Z_N </math> remains of <math> O(1) </math> when <math> N \to \infty </math>, and we can define its distribution in this limit. In all the disordered systems models we will consider in these lectures, the free-energy not only has a well defined distribution in the limit, but it is also self-averaging: it does not fluctuate from sample to sample when <math> N </math> is large, and similarly for all the thermodynamics observables, that can be obtained taking derivatives of the free energy.  The physics of the system does not depend on the particular sample.
<br>
<br>
<br>


<ol start="2">
<code>Example 2. </code> While intensive quantities like <math> f_N </math> are self-averaging, quantities scaling exponentially like the partition function <math> Z_N </math> are not necessarily so: in particular, we will see that they are not when the system is in a glassy phase. Another example is given in Problem 1 below, where <math> X_N \to \mathcal{N}_N(E) </math> and <math> Y_N \to S(E) </math>.
<li><em> Self-averaging.</em> For  <math> |\epsilon| \leq  \sqrt{\log 2} </math> the quantity <math> \mathcal{N} </math> is self-averaging: its distribution concentrates around the average value <math> \overline{\mathcal{N}} </math> when  <math> N \to \infty </math>. Show that this is the case by computing the second moment <math> \overline{\mathcal{N}^2} </math> and using the central limit theorem. Show that  this is no longer true in the region where the annealed entropy is negative: why does one expect fluctuations to be relevant in this region?</li>
   
</ol>
 
<br>
<br>
<br>


<ol start="3">
 
<li> <em> Rare events vs typical values.</em> For  <math> |\epsilon| > \sqrt{\log 2} </math> the annealed entropy is negative: the average number of configurations with those energy densities is exponentially small in <math> N </math>. This implies that the probability to get configurations with those energy is exponentially small in <math> N </math>: these configurations are rare. Do you have an idea of how to show this, using the expression for <math> \overline{\mathcal{N}}</math>? What is the typical value of <math> \mathcal{N} </math> in this region? Justify why the point where the entropy vanishes coincides with the ground state energy of the model.</li>
<li> <ins>'''Quenched averages.'''</ins> Consider exponentially scaling quantities like <math> X_N = e^{N Y_N} </math>, that occur very naturally in statistical physics. How to get <math> Y^{\text{typ}} </math> from <math> X_N </math>? On one hand, if we know how to compute <math>  X_N^{\text{typ}} </math>, then:
</ol>
<math display="block">
Y^{\text{typ}} = \lim_{N \to \infty} \frac{{\log X_N^{\text{typ}}}}{N}.
</math>
If <math> Y_N </math> is self-averaging, we can also use
<math display="block">
Y^{\text{typ}} =\lim_{N \to \infty} \mathbb{E}(Y_N)= \lim_{N \to \infty} \frac{\mathbb{E}(\log X_N)}{N}.
</math>
In the language of disordered systems, computing the typical value of <math> X_N </math> through the average of its logarithm  corresponds to performing a <ins> quenched average</ins>: from this average, one extracts the correct asymptotic value of the self-averaging quantity <math> Y_N </math>.
</li><br>
 
<li> <ins>'''Annealed averages.'''</ins> The quenched average does not necessarily coincide with the <ins> annealed average</ins>, defined as:
<math display="block">  
Y_{\text{a}} = \lim_{N \to \infty} \frac{\log \mathbb{E}(X_N)}{N}.
</math>
In fact, it always holds <math> \mathbb{E}(\log X_N) \leq \log \mathbb{E}(X_N)</math>  because of the concavity of the logarithm.
When the inequality is strict and quenched and annealed averages are not the same, it means that <math> X_N </math> is not self-averaging, and its average value is exponentially larger than the typical value (because the average is dominated by rare events). In this case, to get the correct limit of the self-averaging quantity <math> Y_N </math> one has to perform the quenched average.<sup>[[#Notes|[**] ]]</sup> This is what happens in many cases in disordered systems, for example in glassy phases. Sometimes, computing quenched averages requires ad hoc techniques (e.g., replica tricks).
</li>
</ul>
 
 
 
<div style="font-size:89%">
: <small>[*]</small> - See  [[Media:SelfAvMathNote.pdf| here]] for a note on the equivalence of these two criteria.
</div>
 
<div style="font-size:89%">
: <small>[**]</small> - Notice that the opposite is not true: one can have situations in which the partition function is not self-averaging, but still the quenched free energy coincides with the annealed one.
</div>
 
== Problems ==
This problem and the one of next week deal with the Random Energy Model (REM). The REM has been introduced in <sup>[[#Notes|[1] ]]</sup>. In the REM the system can take <math> M=2^N </math> configurations <math> \vec \sigma^\alpha=(\sigma^\alpha_1, \cdots, \sigma^\alpha_N)</math> with <math> \sigma^\alpha_i = \pm 1 </math>. To each  configuration <math>\alpha=1, \cdots, 2^N</math> is assigned a random energy <math>E_\alpha</math>. The random energies are independent, taken from a Gaussian distribution
<math> p(E) =( 2 \pi N)^{-1/2}e^{-\frac{E^2}{2 N}}.</math>
 
=== Problem 1: the energy landscape of the REM ===
 




Comment: this analysis of the landscape suggests that fluctuations due to the randomness become relevant when one looks at the bottom of their energy landscape, close to the ground state energy density. We show below that this intuition is correct, and corresponds to the fact that the partition function <math> Z </math> is not self-averaging in the low-T phase of the model.
In this problem we study the random variable <math> \mathcal{N}_N(E)dE </math>, that is the number of configurations having energy <math> E_\alpha \in [E, E+dE] </math>. For large <math> N </math> this variable scales exponentially <math>\mathcal{N}_N(E)dE \sim  e^{N S_N\left( E/N\right)}dE</math>. Let <math> \epsilon=E/N </math>. The goal of this Problem is to show that the asymptotic value of the entropy <math> S_N(\epsilon) </math>, that is self-averaging, is given by (henceforth, <math> S(\epsilon):=S_\infty(\epsilon) </math>):


=== Problem 2: the free energy and the freezing transition ===
[[File:Entropy REM.png|thumb|left|x140px|Entropy of the Random Energy Model]]


We now compute the equilibrium phase diagram of the model, and in particular the quenched free energy density <math>f </math> which controls the scaling of the typical value of the partition function, <math>Z \sim e^{-N \beta \, f +o(N) } </math> and which equals to
<math display="block">
<center><math>
S_\infty(\epsilon):=\lim_{N \to \infty} S_N(\epsilon)=\begin{cases}
f =  
\log 2- \frac{\epsilon^2}{2} \quad &\text{ if } |\epsilon| \leq \sqrt{2 \, \log 2} \\
\begin{cases}
- \infty \quad &\text{ if } |\epsilon| >\sqrt{2 \, \log 2}
&- \left( T \log 2 + \frac{1}{4 T}\right) \quad \text{if} \quad T \geq T_c\\
\end{cases}
& - \sqrt{\log 2} \quad \text{if} \quad T <T_c
</math>
\end{cases} \quad \quad T_c= (2 \sqrt{\log 2})^{-1}
<br>
</math></center>
At <math> T_c </math> a transition occurs, often called freezing transition: in the whole low-temperature phase, the free-energy is “frozen” at the value that it has at the critical temperature  <math>T= T_c </math>.


<ol>
<ol>
<li><em> The thermodynamical transition and the freezing.</em>  
<li> <em> Averages: the annealed entropy.</em> We begin by computing the annealed entropy <math> S_{\text{a}} </math>, which is defined by the average <math> \mathbb{E}({\mathcal{N}_N(E)} dE)= \text{exp}\left(N S_{\text{a}}\left( E/N \right)+ o(N)\right)dE </math>. Compute this function using the representation <math> \mathcal{N}_N(E)dE= \sum_{\alpha=1}^{2^N} \chi_\alpha(E) dE \;</math>   [with <math> \chi_\alpha(E)=1</math> if  <math> E_\alpha \in [E, E+dE]</math> and  <math> \chi_\alpha(E)=0</math> otherwise]. </li>
The partition function the REM reads
<math>
Z = \sum_{\alpha=1}^{2^N} e^{-\beta E_\alpha}= \int dE \, \mathcal{N}(E) e^{-\beta E}.
</math>  
Using the behaviour of the typical value of <math> \mathcal{N} </math> determined in Problem 1, derive the free energy of the model (hint: perform a saddle point calculation). What is the order of this thermodynamic transition?
</ol>
</ol>
</li>
<br>
<br>


<ol start="2">
<ol start="2">
<li> <em> Entropy.</em> What happens to the entropy of the model when the critical temperature is reached, and in the low temperature phase? What does this imply for the partition function <math> Z</math>?</li>
<li><em> Self-averaging.</em> For  <math> |\epsilon| \leq  \sqrt{2 \, \log 2} </math> the quantity <math> \mathcal{N}_N </math> is self-averaging: its distribution concentrates around the average value <math> \mathbb{E}(\mathcal{N}_N) </math> when  <math> N \to \infty </math>. Show this by computing the second moment  <math>\mathbb{E}(\mathcal{N}^2_N)</math>. Deduce that <math> S(\epsilon)= S_a(\epsilon)</math> when  <math> |\epsilon| \leq  \sqrt{2 \, \log 2} </math>. This property of being self-averaging is no longer true in the region where the annealed entropy is negative: why does one expect fluctuations to be relevant in this region?</li>
</ol>
<br>
 
<ol start="3">
<li> <em> Rare events.</em> For  <math> |\epsilon| > \sqrt{2 \, \log 2} </math> the annealed entropy is negative: the average number of configurations with those energy densities is exponentially small in <math> N </math>. This implies that the probability to get configurations with those energy is exponentially small in <math> N </math>: these configurations are rare. Do you have an idea of how to show this, using the expression for  <math> \mathbb{E}(\mathcal{N}_N)?</math> What is the typical value of <math> \mathcal{N}_N </math> in this region? Putting everything together, derive the form of the typical value of the entropy density. </li>
</ol>
</ol>
<br>
<br>


<ol start="3">
<ol start="3">
<li><em> Fluctuations, and back to average vs typical.</em> Similarly to the entropy, one can define an annealed free energy <math> f^A </math> from <math> \overline{Z}=e^{- N \beta f^A + o(N)} </math>: show that in the whole low-temperature phase this is smaller than the quenched free energy obtained above. Putting all together, explain why the average of the partition function in the low-T phase is dominated by rare events.
<li> <em> The energy landscape.</em> When <math>N </math> is large, if one extract one configuration at random (with a uniform measure over all configurations), what will be its energy? What is the equilibrium energy density of the system at infinite temperature? What is the typical value of the ground state energy? Deduce this from the entropy calculation: is this consistent with Alberto's lecture?  </li>
</ol></li>
</ol>


=== Problem 3: freezing as a localization/condensation transition ===
<!--'''Comment:''' this analysis of the landscape suggests that in the large  <math> N </math> limit, the fluctuations due to the randomness become relevant when one looks at the bottom of their energy landscape, close to the ground state energy. We show below that this intuition is correct, and corresponds to the fact that the partition function <math> Z </math> has an interesting behaviour at low temperature.-->


In this final exercise, we show how the freezing transition can be understood in terms of extreme valued statistics (discussed in the lecture) and localization. We consider the energies of the configurations and define <math> E_\alpha= - N \sqrt{\log 2} + \delta E_\alpha </math>, so that
<!--<ul>'''Example: typical vs average.''' Often, quantities like <math> Y_N </math> have a distribution that for large <math> N </math> takes the form <math> P_{Y_N}(y) \sim e^{-N^\alpha g(y)+ \text{subleading}} </math> where <math> g(y) </math> is some positive function and <math> \alpha>0 </math>. This is called a <ins> large deviation form </ins> for the probability distribution, with <ins> speed </ins> <math> N^\alpha </math>. This distribution is of <math> O(1) </math> for the value <math> y^{\text{typ}} </math> such that <math> g(y^{\text{typ}})=0 =g'(y^{\text{typ}})</math>: this value is the <ins> typical value </ins> of <math> Y_N </math> (asymptotically at large <math> N </math>); all the other values of <math> y </math> are associated to a probability that is exponentially small in <math> N^\alpha</math>: they are <ins>exponentially rare</ins>.  
<center><math>
{Z} = e^{ \beta N \sqrt{\log 2} }\sum_{\alpha=1}^{2^N} e^{-\beta \delta E_\alpha}= e^{ \beta N \sqrt{\log 2} }\sum_{\alpha=1}^{2^N} z_\alpha
</math></center>
We show that <math> Z </math> is a sum of random variables that become heavy tailed for <math> T < T_c </math>, implying that the central limit theorem is violated and this sum is dominated by few terms, the largest ones. This can be interpreted as the occurrence of localization.


Consider now an exponentially scaling quantity like <math> X_N = e^{N Y_N} </math>, and let’s fix <math> \alpha=1 </math>. The asymptotic typical values <math> x^{\text{typ}} </math> and  <math> y^{\text{typ}} </math> are related by:
<math display="block">
y^{\text{typ}}  =\lim_{N \to \infty} \frac{\log x^{\text{typ}} }{N},
</math>
so the scaling of <math> x^{\text{typ}} </math> is  <math> x^{\text{typ}}\sim e^{N y^{\text{typ}}} </math>. Let us now look at the scaling of the average.
The average of <math> X_N </math> can be computed with the <ins> saddle point approximation </ins> for large <math> N </math>:
<math display="block">
  \overline{X_N} =\int dy\, P_{Y_N}(y)\, e^{N y}= \int dy\, e^{N[y- g(y)]+o(N)} =e^{N [y^*-g(y*)]+ o(N) },
</math>
where <math> y^* </math> is the point maximising the shifted function <math> \tilde{g}(y)= y-g(y)</math>. In this example, <math> y^* \neq y^{\text{ty}} </math>: the asymptotic of the average value of <math> X_N </math> is different from the asymptotic of the typical value. In particular, the average is dominated by rare events, i.e. realisations in which <math> Y_N </math> takes the value <math> y^*</math>, whose probability of occurrence is exponentially small.
</ul>-->


<ol>
== Check out: key concepts ==
<li> <em> Heavy tails and concentration.</em> Compute the distribution of the variables <math> \delta E_\alpha </math> and show that for <math> (\delta E)^2/N \ll 1 </math> this is an exponential. Using this, compute the distribution of the  <math> z_\alpha </math> and show that it is a power law,
 
<center><math>
Self-averaging, average value vs typical value, rare events, quenched vs annealed.
p(z)= \frac{c}{z^{1+\mu}} \quad \quad \mu= \frac{2 \sqrt{\log 2}}{\beta}
 
</math></center>
== To know more ==
When <math> T < T_c </math>, one has  <math> \mu<1 </math>: the distribution of <math> z </math> becomes heavy tailed. What does this imply for the sum  <math> Z </math>? Is this consistent with the behaviour of the partition function and of the entropy discussed in Problem 2? Why can one talk about a localization or condensation transition?</li>
* Derrida. Random-energy model: limit of a family of disordered models [https://hal.science/hal-03285940v1/document]
</ol>
<br>


<ol start="2">
* A note on terminology:
<li><em> Inverse participation ratio.</em> The low temperature behaviour of the partition function an be characterized in terms of a standard measure of localization (or condensation), the Inverse Participation Ratio (IPR) defined as:
<div style="font-size:89%">
<center><math>
The terms “quenched” and “annealed” come from metallurgy and refer to the procedure in which you cool a very hot piece of metal: a system is quenched if it is cooled very rapidly (istantaneously changing its environment by putting it into cold water, for instance) and has to adjusts to this new fixed environment; annealed if it is cooled slowly, kept in (quasi)equilibrium with its changing environment at all times. Think now at how you compute the free energy of a disordered system, and at disorder as the environment. In the quenched protocol, you first compute the average over the configurations of the system (with the Boltzmann weight) keeping the disorder (environment) fixed, so the configurations have to adjust to the given disorder. Like the hot metal has to adjust itself to the new, fixed, cold water). Then you take the log and only afterwards average over the randomness, that plays the role of the environment (not even needed, at large <math>N</math>, if the free-energy is self-averaging). In the annealed protocol instead, the disorder (environment) and the configurations are treated on the same footing and adjust to each others, you average over both simultaneously. The quenched case corresponds to equilibrating at fixed disorder, and it is justified when there is a separation of timescales and the disorder evolved much slower than the variables (the cold water changes temperature much slower than the metal you put in it); the annealed one to considering the disorder as a dynamical variable, that co-evolves with the variables.
IPR= \frac{\sum_{\alpha=1}^{2^N} z_\alpha^2}{[\sum_{\alpha=1}^{2^N} z_\alpha]^2}= \sum_{\alpha=1}^{2^N} \omega_\alpha^2 \quad \quad \omega_\alpha=\frac{ z_\alpha}{\sum_{\alpha=1}^{2^N} z_\alpha}.
</div>
</math></center>
When <math> z </math> is power law distributed with exponent <math> \mu </math>, one can show (HOMEWORK!) that
<center><math>
IPR= \frac{\Gamma(2-\mu)}{\Gamma(\mu) \Gamma(1-\mu)}.
</math></center>
Discuss how this quantity changes across the transition at <math> \mu=1 </math>, and how this fits with what you expect in general in a localized phase.  
</li>
</ol>
<br>

Latest revision as of 11:53, 11 March 2026

Goal: understanding the energy landscape of the simplest spin-glass model, the Random Energy Model (REM).
Techniques: probability theory.



Large-N disordered systems: a probabilistic dictionary


  • Notation. We will discuss disordered systems with N degrees of freedom (for instance, for a spin system on a lattice of size L in dimension d, N=Ld). The disorder can have various sources: (e.g., the interactions between the degrees of freedom are random). Since the systems are random, the quantities that describe their properties (the free energy, the number of configurations of the system that satisfy a certain property, the magnetization etc) are also random variables, with a distribution. In this discussion we denote these random variables generically with XN,YN (where the subscript denotes the number of degrees of freedom) and with PXN(x),PYN(y) their distribution. We denote with 𝔼() the average over the disorder (equivalently, over the distribution PXN(x) of the quantity, induced by the disorder). Statistical physics goal is to characterize the behavior of these quantities in the limit N. In statistical physics, in general one deals with exponentially scaling quantities: XN=eNYN

  • Typical and rare. The typical value of a random variable is the value at which its distribution peaks (it is the most probable value). Values at the tails of the distribution, where the probability density does not peak but it is small (for instance, vanishing with N) are said to be rare. In general, average and typical value of a random variable may not coincide: this happens when the average is dominated by values of the random variable that lie in the tail of the distribution and are rare, associated to a small probability of occurrence.

    Short exercise. Consider a random variable X with a log-normal distribution, meaning that X=eY where Y is a random variable with a Gaussian distribution with average μ and variance σ. Show that Xtyp=eμ while 𝔼(X)=eμ+σ22, and thus the average value is larger than the typical one.

  • Self-averagingness. The physics of disordered systems is described by quantities that are distributed when N is finite: they take different values from sample to sample of the system, where a sample corresponds to one "realization" of the disorder. However, for relevant quantities sample to sample fluctuations are suppressed when N. These quantities are said to be self-averaging . A random variable YN is self-averaging when, in the limit N, its distribution concentrates around the average, collapsing to a deterministic value: limNYN=limN𝔼(YN):=Y,𝔼(YN)=dyPYN(y)y This happens when its fluctuations are small compared to the average, meaning that [*] limN𝔼(YN2)𝔼(YN)2=1. For self-averaging quantities, in the limit N the distribution collapses to a single value Y, that is both the limit of the average and typical value of the quantity: sample-to-sample fluctuations are suppressed when N is large. When the random variable is not self-averaging, it remains distributed in the limit N.

  • Example 1. Consider the partition function of a disordered system at inverse temperature β, ZN(β). When N is large this random variable has an exponential scaling, ZN(β)eβNfN(β), where the variable fN(β) is the free energy density. This scaling means that the random variable βfN=N1logZN remains of O(1) when N, and we can define its distribution in this limit. In all the disordered systems models we will consider in these lectures, the free-energy not only has a well defined distribution in the limit, but it is also self-averaging: it does not fluctuate from sample to sample when N is large, and similarly for all the thermodynamics observables, that can be obtained taking derivatives of the free energy. The physics of the system does not depend on the particular sample.

    Example 2. While intensive quantities like fN are self-averaging, quantities scaling exponentially like the partition function ZN are not necessarily so: in particular, we will see that they are not when the system is in a glassy phase. Another example is given in Problem 1 below, where XN𝒩N(E) and YNS(E).

  • Quenched averages. Consider exponentially scaling quantities like XN=eNYN, that occur very naturally in statistical physics. How to get Ytyp from XN? On one hand, if we know how to compute XNtyp, then: Ytyp=limNlogXNtypN. If YN is self-averaging, we can also use Ytyp=limN𝔼(YN)=limN𝔼(logXN)N. In the language of disordered systems, computing the typical value of XN through the average of its logarithm corresponds to performing a quenched average: from this average, one extracts the correct asymptotic value of the self-averaging quantity YN.

  • Annealed averages. The quenched average does not necessarily coincide with the annealed average, defined as: Ya=limNlog𝔼(XN)N. In fact, it always holds 𝔼(logXN)log𝔼(XN) because of the concavity of the logarithm. When the inequality is strict and quenched and annealed averages are not the same, it means that XN is not self-averaging, and its average value is exponentially larger than the typical value (because the average is dominated by rare events). In this case, to get the correct limit of the self-averaging quantity YN one has to perform the quenched average.[**] This is what happens in many cases in disordered systems, for example in glassy phases. Sometimes, computing quenched averages requires ad hoc techniques (e.g., replica tricks).


[*] - See here for a note on the equivalence of these two criteria.
[**] - Notice that the opposite is not true: one can have situations in which the partition function is not self-averaging, but still the quenched free energy coincides with the annealed one.

Problems

This problem and the one of next week deal with the Random Energy Model (REM). The REM has been introduced in [1] . In the REM the system can take M=2N configurations σα=(σ1α,,σNα) with σiα=±1. To each configuration α=1,,2N is assigned a random energy Eα. The random energies are independent, taken from a Gaussian distribution p(E)=(2πN)1/2eE22N.

Problem 1: the energy landscape of the REM

In this problem we study the random variable 𝒩N(E)dE, that is the number of configurations having energy Eα[E,E+dE]. For large N this variable scales exponentially 𝒩N(E)dEeNSN(E/N)dE. Let ϵ=E/N. The goal of this Problem is to show that the asymptotic value of the entropy SN(ϵ), that is self-averaging, is given by (henceforth, S(ϵ):=S(ϵ)):

Entropy of the Random Energy Model

S(ϵ):=limNSN(ϵ)={log2ϵ22 if |ϵ|2log2 if |ϵ|>2log2

  1. Averages: the annealed entropy. We begin by computing the annealed entropy Sa, which is defined by the average 𝔼(𝒩N(E)dE)=exp(NSa(E/N)+o(N))dE. Compute this function using the representation 𝒩N(E)dE=α=12Nχα(E)dE [with χα(E)=1 if Eα[E,E+dE] and χα(E)=0 otherwise].


  1. Self-averaging. For |ϵ|2log2 the quantity 𝒩N is self-averaging: its distribution concentrates around the average value 𝔼(𝒩N) when N. Show this by computing the second moment 𝔼(𝒩N2). Deduce that S(ϵ)=Sa(ϵ) when |ϵ|2log2. This property of being self-averaging is no longer true in the region where the annealed entropy is negative: why does one expect fluctuations to be relevant in this region?


  1. Rare events. For |ϵ|>2log2 the annealed entropy is negative: the average number of configurations with those energy densities is exponentially small in N. This implies that the probability to get configurations with those energy is exponentially small in N: these configurations are rare. Do you have an idea of how to show this, using the expression for 𝔼(𝒩N)? What is the typical value of 𝒩N in this region? Putting everything together, derive the form of the typical value of the entropy density.


  1. The energy landscape. When N is large, if one extract one configuration at random (with a uniform measure over all configurations), what will be its energy? What is the equilibrium energy density of the system at infinite temperature? What is the typical value of the ground state energy? Deduce this from the entropy calculation: is this consistent with Alberto's lecture?


Check out: key concepts

Self-averaging, average value vs typical value, rare events, quenched vs annealed.

To know more

  • Derrida. Random-energy model: limit of a family of disordered models [1]
  • A note on terminology:

The terms “quenched” and “annealed” come from metallurgy and refer to the procedure in which you cool a very hot piece of metal: a system is quenched if it is cooled very rapidly (istantaneously changing its environment by putting it into cold water, for instance) and has to adjusts to this new fixed environment; annealed if it is cooled slowly, kept in (quasi)equilibrium with its changing environment at all times. Think now at how you compute the free energy of a disordered system, and at disorder as the environment. In the quenched protocol, you first compute the average over the configurations of the system (with the Boltzmann weight) keeping the disorder (environment) fixed, so the configurations have to adjust to the given disorder. Like the hot metal has to adjust itself to the new, fixed, cold water). Then you take the log and only afterwards average over the randomness, that plays the role of the environment (not even needed, at large N, if the free-energy is self-averaging). In the annealed protocol instead, the disorder (environment) and the configurations are treated on the same footing and adjust to each others, you average over both simultaneously. The quenched case corresponds to equilibrating at fixed disorder, and it is justified when there is a separation of timescales and the disorder evolved much slower than the variables (the cold water changes temperature much slower than the metal you put in it); the annealed one to considering the disorder as a dynamical variable, that co-evolves with the variables.