|   |   | 
| Line 130: | Line 130: | 
|  | In the limit of large  <math>N </math>, this integral can be computed with the saddle point method, and one gets |  | In the limit of large  <math>N </math>, this integral can be computed with the saddle point method, and one gets | 
|  | <center><math> |  | <center><math> | 
|  | \overline{Z} = e^{N \left[\Sigma(\epsilon^*)- \beta \epsilon^* \right]+ o(N)}, \quad \quad \epsilon^*= \text{argmax}_{|\epsilon| \leq \sqrt{\log 2}} \left(\Sigma(\epsilon)- \beta \epsilon \right)
 |  | {Z} = e^{N \left[\Sigma(\epsilon^*)- \beta \epsilon^* \right]+ o(N)}, \quad \quad \epsilon^*= \text{argmax}_{|\epsilon| \leq \sqrt{\log 2}} \left(\Sigma(\epsilon)- \beta \epsilon \right) | 
|  | </math></center> |  | </math></center> | 
|  | Using the expression of the entropy, we see that the function is stationary at  <math>\epsilon^*= -1/2T </math>, which belongs to the domain of integration whenever <math>T \geq T_c= 1/(2 \sqrt{\log 2}) </math>. This temperature identifies a transition point: for all values of <math>T < T_c </math>, the stationary point is outside the domain and thus <math>\epsilon^*</math> has to be chosen at the boundary of the domain, <math>\epsilon^*= -\sqrt{\log 2}</math>. |  | Using the expression of the entropy, we see that the function is stationary at  <math>\epsilon^*= -1/2T </math>, which belongs to the domain of integration whenever <math>T \geq T_c= 1/(2 \sqrt{\log 2}) </math>. This temperature identifies a transition point: for all values of <math>T < T_c </math>, the stationary point is outside the domain and thus <math>\epsilon^*</math> has to be chosen at the boundary of the domain, <math>\epsilon^*= -\sqrt{\log 2}</math>. | 
Spin glass Transition
Experiments
Parlare dei campioni di rame dopati con il magnesio, marino o no: trovare due figure una  di suscettivita e una di calore specifico, prova della transizione termodinamica.
Edwards Anderson model
We consider for simplicity the Ising version of this model.
Ising spins takes two values  and live on a lattice of
 and live on a lattice of  sitees
 sitees  . 
The enregy is writteen as a sum between the nearest neighbours <i,j>:
. 
The enregy is writteen as a sum between the nearest neighbours <i,j>: 
  
Edwards and Anderson proposed to study this model for couplings  that are i.i.d. random variables with zero mean.
We set
 that are i.i.d. random variables with zero mean.
We set  the coupling distribution indicate the avergage over the couplings called disorder average, with an overline:
 the coupling distribution indicate the avergage over the couplings called disorder average, with an overline: 
 
It is crucial to assume  , otherwise the model displays ferro/antiferro order. We sill discuss two distributions:
, otherwise the model displays ferro/antiferro order. We sill discuss two distributions:
- Gaussian couplings:  
- Coin toss couplings,  , selected  with probability , selected  with probability . .
Edwards Anderson order parameter
The SK model
Sherrington and Kirkpatrik considered the fully connected version of the model with Gaussian couplings:
  
At the inverse temperature  , the partion function of the model is
, the partion function of the model is
  
Here  is the energy associated to the configuration
 is the energy associated to the configuration   .
This model presents a thermodynamic transition at
.
This model presents a thermodynamic transition at  .
.
Random energy model
The solution of the SK is difficult. To make progress we first study the radnom energy model (REM) introduced by B. Derrida. 
Derivation of the model
The REM neglects the correlations between the  configurations and assumes the
 configurations and assumes the  as iid variables.
 as iid variables.
- Show that the energy distribution is
 
and determine  
The Solution: Part 1
We provide different solutions of the Random Energy Model (REM).  The first one focus on the statistics of the smallest energies among the ones associated to the  configurations.
 configurations.
Consider the   energies:
  energies:  . They are i.i.d. variables, drawn from the Gaussian distribution
. They are i.i.d. variables, drawn from the Gaussian distribution  .
It is useful to  use the following notations:
.
It is useful to  use the following notations:
 for for . It  represents the probability to find an energy smaller than E. . It  represents the probability to find an energy smaller than E.
 . It represents the probability to dfind an energy  larger than E. . It represents the probability to dfind an energy  larger than E.
Extreme value statistics for iid
We denote
 
Our goal is to compute the cumulative distribution   for large M and iid variables.
 for large M and iid variables.
We need to understand two key relations: 
- The first relation is exact:
 
 
- The second relation identifies the typical value of the minimum, namely  : :
 
 . 
Close to  , we expect
, we expect  . Hence, from the limit
. Hence, from the limit   we re-write the first relation:
 we re-write the first relation:
  
 
Moreover, if we define  we recover the famous Gumbel distribution:
 we recover the famous Gumbel distribution:
 
 
Exercise L1-A: the Gaussian case
Specify these results to the Guassian case and find
- the typical value of the minimum
%
 
 
- The expression  
- The expression of the Gumbel distribution for the Gaussian case
 
 
Density of states above the minimum
For a given disorder realization, we compute  , the number of configurations above the minimum, but with an energy smaller than
, the number of configurations above the minimum, but with an energy smaller than  .
.
![{\displaystyle {\text{Prob}}(d(x)=k)=M{\binom {M-1}{k}}\int dE\;p(E)[P^{>}(E)-P^{>}(E+x)]^{k}P^{>}(E+x)^{M-k-1}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/69668fb29a513e8d633235a7f24af7042a44ebdd) 
Taking the average   , we derive
, we derive
![{\displaystyle {\overline {d(x)}}=M(M-1)\int dE\;p(E)\left[P^{>}(E)-P^{>}(E+x)\right]P^{>}(E)^{M-2}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/af87af1ecaa2c149718802e2d0dc5e4395b2dc2b) 
Number
The landscape
To characterize the energy landscape of the REM, we can determine the number  of configurations having energy
 of configurations having energy  ![{\displaystyle E_{\alpha }\in [E,E+dE]}](https://wikimedia.org/api/rest_v1/media/math/render/svg/743293b01383f5322ecc6b6bb9268ca083af88f4) . This quantity is a random variable. For large
. This quantity is a random variable. For large  , its typical value is given by
, its typical value is given by
 
where  is the entropy of the model. A sketch of this function is in Fig. X. The point where the entropy vanishes,
 is the entropy of the model. A sketch of this function is in Fig. X. The point where the entropy vanishes,  , is the energy density of the ground state, consistently with what we obtained with extreme values statistics. The entropy is maximal at
, is the energy density of the ground state, consistently with what we obtained with extreme values statistics. The entropy is maximal at   : the highest number of configurations have vanishing energy density.
: the highest number of configurations have vanishing energy density.
- We begin by computing the average   . We set . We set , where , where is the annealed entropy. Write is the annealed entropy. Write with with if if![{\displaystyle E_{\alpha }\in [E,E+dE]}](https://wikimedia.org/api/rest_v1/media/math/render/svg/743293b01383f5322ecc6b6bb9268ca083af88f4) and and otherwise. Use this together with otherwise. Use this together with to obtain to obtain : when does this coincide with the entropy? : when does this coincide with the entropy?
- For    the quantity the quantity is self-averaging. This means that its distribution concentrates around the average value is self-averaging. This means that its distribution concentrates around the average value when when . Show that this is the case by computing the second moment . Show that this is the case by computing the second moment and using the central limit theorem. Show that  this is no longer true in the region where the annealed entropy is negative. and using the central limit theorem. Show that  this is no longer true in the region where the annealed entropy is negative.
- For   the annealed entropy is negative. This means that configurations with those energy are exponentially rare: the probability to find one is exponentially small in the annealed entropy is negative. This means that configurations with those energy are exponentially rare: the probability to find one is exponentially small in . Do you have an idea of how to show this, using the expression for . Do you have an idea of how to show this, using the expression for ? Why the entropy is zero in this region? Why the point where the entropy vanishes coincides with the ground state energy of the model? ? Why the entropy is zero in this region? Why the point where the entropy vanishes coincides with the ground state energy of the model?
this will be responsible of the fact that the partition function  is not self-averaging in the low-T phase, as we discuss below.
 is not self-averaging in the low-T phase, as we discuss below.
The free energy and the freezing transition
Let us compute the free energy  of the REM. The partition function reads
 of the REM. The partition function reads 
 
 
We have shown above the behaviour of the typical value of  for large
 for large  . The typical value of the partition function is
. The typical value of the partition function is
![{\displaystyle Z=\int _{-N{\sqrt {\log 2}}}^{N{\sqrt {\log 2}}}dE\,{\mathcal {N}}(E)e^{-\beta E}=\int _{-{\sqrt {\log 2}}}^{\sqrt {\log 2}}d\epsilon \,e^{N\left[\Sigma (\epsilon )-\beta \epsilon \right]+o(N)}}](https://wikimedia.org/api/rest_v1/media/math/render/svg/e145c9e0d4620a0d93bdc74127c00c29313c53ef) 
 
In the limit of large   , this integral can be computed with the saddle point method, and one gets
, this integral can be computed with the saddle point method, and one gets
![{\displaystyle {Z}=e^{N\left[\Sigma (\epsilon ^{*})-\beta \epsilon ^{*}\right]+o(N)},\quad \quad \epsilon ^{*}={\text{argmax}}_{|\epsilon |\leq {\sqrt {\log 2}}}\left(\Sigma (\epsilon )-\beta \epsilon \right)}](https://wikimedia.org/api/rest_v1/media/math/render/svg/3271e3ed43e1d4f43fb02db629ee1c652eba3e76) 
Using the expression of the entropy, we see that the function is stationary at   , which belongs to the domain of integration whenever
, which belongs to the domain of integration whenever  . This temperature identifies a transition point: for all values of
. This temperature identifies a transition point: for all values of  , the stationary point is outside the domain and thus
, the stationary point is outside the domain and thus  has to be chosen at the boundary of the domain,
 has to be chosen at the boundary of the domain,  .
.
The free energy becomes
 
Bibliography
Bibliography
- Theory of spin glasses, S. F. Edwards and P. W. Anderson, J. Phys. F: Met. Phys. 5 965, 1975