T-6: Difference between revisions

From Disordered Systems Wiki
Jump to navigation Jump to search
Line 86: Line 86:




<li> '''Quenches.''' We can generalize this protocol to higher <math>T</math>: we extract randomly the initial condition of the dynamics, and then we evolve the configuration following Langevin dynamics  
<li> '''Quenches.''' We can generalize this protocol to higher <math>T</math>: we extract randomly the initial condition of the dynamics, and then we evolve the configuration with Langevin dynamics (gradient descent + noise):
 
<math display="block">
\frac{d \vec{\sigma}(t)}{dt}=- \nabla_\perp E(\vec{\sigma})+ \vec{\eta}_\perp(t), \quad \quad \langle \eta_i(t) \eta_j(t')\rangle= 2 T \delta_{ij} \delta(t-t')
</math>
In Langevin dynamics, <math>\vec{\eta}_\perp(t)</math> a Gaussian vector at each time <math> t </math>, uncorrelated from the vectors at other times <math> t' \neq t </math>,  with zero average and some constant variance proportional to temperature. They represent effectively the action of a bath on the system.


   
   

Revision as of 15:58, 15 March 2026

Goal: Complete the characterisation of the energy landscape of the spherical p-spin.
Techniques: saddle point, random matrix theory.


Problems

Problem 6: the Hessian at the stationary points, and random matrix theory

This is a continuation of problem 5. To get the complexity of the spherical p-spin, it remains to compute the expectation value of the determinant of the Hessian matrix: this is the goal of this problem. We will do this exploiting results from random matrix theory discussion in the Tutorial and Exercise 4 .


  1. Gaussian Random matrices. Show that the matrix M, defined in Problem 5, is a GOE matrix, i.e. a matrix taken from the Gaussian Orthogonal Ensemble, meaning that it is a symmetric matrix with distribution PN(M)=ZN1exp(N4σ2TrM2) where ZN is a normalization. What is the value of σ2?


  1. Eigenvalue density and concentration. Let λα be the eigenvalues of the matrix M. Show that the following identity holds: 𝔼[|det(Mpϵ𝕀)|]=𝔼[exp((N1)dλρN1(λ)log|λpϵ|)],ρN1(λ)=1N1α=1N1δ(λλα) where ρN1(λ) is the empirical eigenvalue distribution. It can be shown that if M is a GOE matrix, the distribution of the empirical distribution has a large deviation form with speed N2, meaning that PN[ρ]=eN2g[ρ] where now g[] is a functional. Using a saddle point argument, show that this implies 𝔼[exp((N1)dλρN1(λ)log|λpϵ|)]=exp[Ndλρ(λ+pϵ)log|λ|+o(N)] where ρ(λ) is the typical value of the eigenvalue density, which satisfies g[ρ]=0.


  1. The semicircle and the complexity. The eigenvalue density of GOE matrices is self-averaging, and it equals to limNρN(λ)=limN𝔼[ρN(λ)]=ρ(λ)=12πσ24σ2λ2
      Combining all the results, show that the annealed complexity is Σa(ϵ)=12log[4e(p1)]ϵ22+Ip(ϵ),Ip(ϵ)=2πdx1(xϵϵth)2log|x|,ϵth=2p1p. The integral Ip(ϵ) can be computed explicitly, and one finds: Ip(ϵ)={ϵ2ϵth212ϵϵthϵ2ϵth21+log(ϵϵth+ϵ2ϵth21)log2ifϵϵthϵ2ϵth212log2ifϵ>ϵth Plot the annealed complexity, and determine numerically where it vanishes: why is this a lower bound or the ground state energy density?


  1. The threshold and the stability. Sketch ρ(λ+pϵ) for different values of ϵ; recalling that the Hessian encodes for the stability of the stationary points, show that there is a transition in the stability of the stationary points at the critical value of the energy density ϵth=2(p1)/p. When are the critical points stable local minima? When are they saddles? Why the stationary points at ϵ=ϵth are called marginally stable ?


Back to dynamics: quenches, and dynamical transitions

Non-rugged vs rugged energy landscapes.

Through Problems 5 and 6, we have shown that the energy landscape of the spherical p-spin model has exponentially many stationary points , and that there is a transition at the energy density ϵth: for ϵ>ϵth the stationary points are saddles, for ϵϵth they are local minima. Let us try to deduce something on the systems's dynamics out of this.


  • Gradient descent dynamics. The local minima are dynamically stable: if I do gradient descent, I get stuck in a local minimum and I exert a small perturbation to the configuration, gradient descent brings me back to the local minimum. In this sense, these configurations are trapping. Therefore, if I try to optimize the landscape, i.e. to reach the ground state, with gradient descent dynamics, I expect that I will not be able to reach the ground state easily, as I will be trapped by these local minima. In fact, for the spherical p-spin model it can be shown that starting the gradient descent dynamics from random initial conditions and evolving the configuration with gradient descent (possibly with infinitesimal noise), limtlimNE(σ(t))N=ϵthϵgs. The system gets stuck at the energy density level where local minima start to appear, and does not reach the deeper local minima.

  • Quenches. We can generalize this protocol to higher T: we extract randomly the initial condition of the dynamics, and then we evolve the configuration with Langevin dynamics (gradient descent + noise): dσ(t)dt=E(σ)+η(t),ηi(t)ηj(t)=2Tδijδ(tt) In Langevin dynamics, η(t) a Gaussian vector at each time t, uncorrelated from the vectors at other times tt, with zero average and some constant variance proportional to temperature. They represent effectively the action of a bath on the system.
  • Optimization by gradient descent. Suppose that we are interested in finding the configurations of minimal energy, starting from an arbitrary configuration σ0: we can implement a dynamics in which we progressively update the configuration moving towards lower and lower values of the energy, hoping to eventually converge to the ground state(s). The simplest dynamics of this sort is gradient descent, dσ(t)dt=E(σ) where E(σ) is the gradient of the landscape restricted to the sphere. The dynamics stops when it reaches a stationary point , a configuration where E(σ)=0. If the landscape has a convex structure, this will be the ground state; if the energy landscape is very non-convex like in glasses, the end point of this algorithm will be a local minimum at energies much higher than the ground state (see sketch).

  • Stationary points and complexity. To guess where gradient descent dynamics (or Langevin dynamics ) are expected to converge, it is useful to understand the distribution of the stationary points, i.e. the number 𝒩(ϵ) of such configuration having a given energy density ϵ. In fully-connected models, this quantity has an exponential scaling, 𝒩(ϵ)exp(NΣ(ϵ)), where Σ(ϵ) is the landscape’s complexity. [*] . Stationary points can be stable (local minima), or unstable (saddles or local maxima): their stability is encoded in the spectrum of the Hessian matrix 2E(σ): when all the eigenvalues of the Hessian are positive, the point is a local minimum (and a saddle otherwise).

  • [*] - This quantity looks similar to the entropy S(ϵ) we computed for the REM in Problem 1. However, while the entropy counts all configurations at a given energy density, the complexity Σ(ϵ) accounts only for the stationary points.


    Check out: key concepts

    Metastable states, Hessian matrices, random matrix theory, landscape’s complexity.