L-1: Difference between revisions

From Disordered Systems Wiki
Jump to navigation Jump to search
Line 52: Line 52:
=== The Solution: Part 1  ===
=== The Solution: Part 1  ===


We provide different solutions of the Random Energy Model (REM). The first one focus on the statistics of the smallest energies among the one associated to the <math>2^N</math> configurations.
We provide different solutions of the Random Energy Model (REM). The first one focus on the statistics of the smallest energies among the ones associated to the <math>M=2^N</math> configurations.
 
Consider the  <math>M=2^N</math>  energies,  <math>M=2^N</math>  nameley independent random variables <math>(x_1,...,x_M)</math> drawn from the Gaussian distribution <math>p(x)</math>.
 
It is useful to  use the following notations:
* <math>P^<(x)=\int_{-\infty}^x dx' p(x')  \sim \frac{\sigma}{\sqrt{2 \pi}|x|}e^{-\frac{x^2}{2 \sigma^2}} \; </math> for  <math>x \to -\infty</math>. It  represents the probability to draw a number smalle than ''x''
* <math> P^>(x)=\int_x^{+\infty} dx' p(x') = 1- P^<(x) </math>. It represents the probabilitty to draw a number larger than ''x''.
 


====  Extreme value stattics for Gaussian variables ====
====  Extreme value stattics for Gaussian variables ====
We consider ''N'' independent random variables <math>(x_1,...,x_N)</math> drawn from the Gaussian distribution <math>p(x)</math>.
We denote
We denote
<center><math>y_N=\min(x_1,...,x_N)</math></center>
<center><math>y_M=\min(x_1,...,x_M)</math></center>
Our goal is to compute the cumulative distribution  <math>Q_M(y)=\text{Prob}(y_M> </math> for large ''M'' in the case of i.i.d .variables.
 
We need to understand two key relations:
* the first relation is exact:
<center><math>Q_M(y) = \left(P^<(y)\right)^M </math> </center>
* the second relation identifies the typical value of the minimum, namely <math> a_M  </math>:
<center><math>P_<(a_M) = 1/M </math> </center>
From the second relation you can get that, in the Gaussian case:
<center><math>a_M =2 \sigma \sqrt{\ln M}-\frac{1}{2}\sqrt{\ln(\lnM)} +O(1) </math> </center>


It is useful to  use the following notations for the cumulative distribution <math>P^<(x)=\int_{-\infty}^x dx' p(x') \sim </math> which represents the probabilitty to draw a number smalle than ''x'' and <math> P^>(x)=\int_x^{+\infty} dx' p(x') = 1- P^<(x) </math> which represents the probabilitty to draw a number larger tthan ''x''.
Close to <math> a_M </math>, where P_<(y) sim 1/M we can use the limt  <math>\lim_{M\to \int} (1+\frac{k}{M})^M =\exp(k)</math> and re-write the first relation:
  <center><math>Q_M(y) \sim \exp\left(-P^>(y)\right)


Let us denote by <math>q_N(y)</math> the distribution of <math>y_N</math> and by <math>Q_N(y)=\text{Prob}(y_N>y)</math> its  cumulative distribution.


* Write <math>Q_N(y)</math> in terms of <math>P^>(y) </math>. (Help: Start to write this relation for <math>N=2,3,...</math>).
This is the fundamental relation of Extreme statistics and we analyze its consequences in the large ''N'' limit where, analogously to the central limit theorem, extremes statistics  display universal features.
* In particular shows that in the  large ''N'' limit  we can write
  <center><math>Q_N(y) \sim \exp\left(-N  P^>(y)\right) </math></center>
Consider now the case <math>\lambda=1</math>
* Write  <math>P^>(x)</math> and <math>P^<(x)</math>. (Remember that <math>x</math>  is a positive number.)
* Write <math>Q_N(y)</math> and <math>q_N(y)</math>.
* Plot <math>q_N(y)</math> for different values of ''N''.




Line 78: Line 85:
<math>P^>( \tilde \tilde y)=\frac 23</math>
<math>P^>( \tilde \tilde y)=\frac 23</math>


* Justify that <math>a_N</math> can be estimated from
<center><math>P^>(a_N)=\frac 1N</math></center>
* Compute  <math>a_N</math> for the exponential distribution and justify that
<center><math>Q_N\left(y=a_N+z\right)</math></center>


In the large ''N'' limit, the distribution  <math>\pi(z)</math> becomes <math>N</math> independent.  
In the large ''N'' limit, the distribution  <math>\pi(z)</math> becomes <math>N</math> independent.  

Revision as of 17:02, 19 November 2023

Spin glass Transition

Experiments

Parlare dei campioni di rame dopati con il magnesio, marino o no: trovare due figure una di suscettivita e una di calore specifico, prova della transizione termodinamica.

Edwards Anderson model

We consider for simplicity the Ising version of this model.

Ising spins takes two values σ=±1 and live on a lattice of N sitees i=1,2,,N. The enregy is writteen as a sum between the nearest neighbours <i,j>:

E=<i,j>Jijσiσj

Edwards and Anderson proposed to study this model for couplings J that are i.i.d. random variables with zero mean. We set π(J) the coupling distribution indicate the avergage over the couplings called disorder average, with an overline:

J¯dJJπ(J)=0

It is crucial to assume J¯=0, otherwise the model displays ferro/antiferro order. We sill discuss two distributions:

  • Gaussian couplings: π(J)=exp(J2/2)/2π
  • Coin toss couplings, J=±1, selected with probability 1/2.

Edwards Anderson order parameter

The SK model

Sherrington and Kirkpatrik considered the fully connected version of the model with Gaussian couplings:

E=i,jJij2Nσiσj

At the inverse temperature β, the partion function of the model is

Z=α=12Nzα,withzα=eβEα

Here Eα is the energy associated to the configuration α. This model presents a thermodynamic transition at βc=??.

Random energy model

The solution of the SK is difficult. To make progress we first study the radnom energy model (REM) introduced by B. Derrida.

Derivation of the model

The REM neglects the correlations between the 2N configurations and assumes the Eα as iid variables.

  • Show that the energy distribution is
p(Eα)=12πσ2eEα22σ2

and determine σ2

The Solution: Part 1

We provide different solutions of the Random Energy Model (REM). The first one focus on the statistics of the smallest energies among the ones associated to the M=2N configurations.

Consider the M=2N energies, M=2N nameley independent random variables (x1,...,xM) drawn from the Gaussian distribution p(x).

It is useful to use the following notations:

  • P<(x)=xdxp(x)σ2π|x|ex22σ2 for x. It represents the probability to draw a number smalle than x
  • P>(x)=x+dxp(x)=1P<(x). It represents the probabilitty to draw a number larger than x.


Extreme value stattics for Gaussian variables

We denote

yM=min(x1,...,xM)

Our goal is to compute the cumulative distribution QM(y)=Prob(yM> for large M in the case of i.i.d .variables.

We need to understand two key relations:

  • the first relation is exact:
QM(y)=(P<(y))M
  • the second relation identifies the typical value of the minimum, namely aM:
P<(aM)=1/M

From the second relation you can get that, in the Gaussian case:

Failed to parse (unknown function "\lnM"): {\displaystyle a_M =2 \sigma \sqrt{\ln M}-\frac{1}{2}\sqrt{\ln(\lnM)} +O(1) }

Close to aM, where P_<(y) sim 1/M we can use the limt limM(1+kM)M=exp(k) and re-write the first relation:

QM(y)exp(MP>(y))Wewantnowtogiveanaturaldefinitionforthenumber<math>aN and bN.

Consider P>(y~)=12. If you draw N independent exponential variables, how many variables (in average) will be greater than y~? Repeat the same exercise with y~~ such that P>(y~~)=23


In the large N limit, the distribution π(z) becomes N independent.

  • Show that in this limit its cumulative takes the from
Π(z)=eez
Π(z)=eez

This is the cumulative distribution of the famous Gumbel distribution.

Let us remark that the precise definition of aN and bN fix the mean and the variance of the rescaled distribution π(z) At variance with the central limit case the mean will be different from zero and the variance different from one.

  • Compute the mean, the variance and the asymptotic behavior of the Gumbel distribution. Draw the distribution. Explain why z=0 is a special point

Generic case: Universality of the Gumbel distribution

The Gumbel distribution is the limit distribution of the maxima of a large class of function. We can say that the Gumbel distribution plays, for extreme statistics, the same role of the Gaussian distribution for the central limit theorem.

By contrast the behavior of aN and bN as a function of N strongly depend on the particular distributions p(x). We discuss here a family of distribution characterized by a fast decay for large x

p(x)cexα

where α>0 The key point is to be able to determine A(x) such that

P>(x)=exp(A(x))
  • For p(x)=ex shows A(x)=x

Otherwise A(x) should be determined asymptotically for large x

  • Show that A(x)=xα+(α1)logx+...
  • Show that in general A(aN)=logN+... and compute aN as a function of α for large N.
  • Show that the maximum distribution take the form
limNQN(y)=(y=aN+zA(aN))

with z Gumbel distributed

  • Identify bN and discuss its behavior as a function of α

Number

Bibliography

Bibliography

  • Theory of spin glasses, S. F. Edwards and P. W. Anderson, J. Phys. F: Met. Phys. 5 965, 1975