L-8: Difference between revisions

From Disordered Systems Wiki
Jump to navigation Jump to search
Line 88: Line 88:
For the log-normal distribution the mean <math> \overline{\Pi_N} = \exp[(\gamma-\gamma_2^2) N]</math> is larger than the median value  <math> \Pi_N^{\text{median}} = \exp(\gamma N)</math> (which is larger than the mode). Hence, <math> \Pi_N </math> is not self averaging, while <math> \ln \Pi_N </math> is self averaging. This is the reason why in the following we will take quenched  averages.
For the log-normal distribution the mean <math> \overline{\Pi_N} = \exp[(\gamma-\gamma_2^2) N]</math> is larger than the median value  <math> \Pi_N^{\text{median}} = \exp(\gamma N)</math> (which is larger than the mode). Hence, <math> \Pi_N </math> is not self averaging, while <math> \ln \Pi_N </math> is self averaging. This is the reason why in the following we will take quenched  averages.


=== Product of random matrices===
== Product of random matrices==


Instead of imposing Dirichelet boundary condition on the extremeties of the box we can impose Cauchy boundaries on one side of the box. Let's rewrite the previous eigensystem in the following form
Let's consider again the Anderson Model in 1d. The eigensystem is well defined in a box of size L with Dirichelet boundary condition on the extremeties of the box.
 
Here we will solve the second order differential equation  imposing instead Cauchy boundaries on one side of the box. Let's rewrite the previous eigensystem in the following form
<center> <math>
<center> <math>
  \begin{bmatrix}
  \begin{bmatrix}
Line 105: Line 107:
\end{bmatrix}
\end{bmatrix}
</math></center>
</math></center>
We can continue the recusion
We can continue the recursion
<center> <math>
<center> <math>
\begin{bmatrix}
\begin{bmatrix}
Line 123: Line 125:
\end{bmatrix}
\end{bmatrix}
</math></center>
</math></center>
It is useful to intruduce the transfer matrix
It is useful to introduce the transfer matrix
<center> <math>
<center> <math>
T_n =\begin{bmatrix}
T_n =\begin{bmatrix}
Line 132: Line 134:
and the product of random matrices
and the product of random matrices
<center> <math>
<center> <math>
M_n= T_n \cdot T_{n-1} \codt\ldots T_1  
M_n= T_n \cdot T_{n-1} \cdot\ldots T_1  
</math></center>
</math></center>

Revision as of 09:14, 17 March 2024

Goal: we will introduce the Anderson model, discuss the behaviour as a function of the dimension. In 1d localization can be connected to the product of random matrices.

Anderson model (tight bindind model)

We consider disordered non-interacting particles hopping between nearest neighbors sites on a lattice. The hamiltonian reads:

The single particle hamiltonian in 1d reads

For simplicity we set the hopping . The disorder are iid random variables drawn, uniformly from the box .

The final goal is to study the statistical properties of eigensystem

Density of states (DOS)

In 1d and in absence of disorder, the dispersion relation is . From the dispersion relation, we compute the density of states (DOS) :

In presence of disorder the DOS becomes larger, and display sample to sample fluctuations. One can consider its mean value, avergaed over disorder realization.

Eigenstates

In absence of disorder the eigenstates are plane waves delocalized along all the system. In presence of disorder, three situations can occur and to distinguish them it is useful to introduce the inverse participation ratio, IPR

The normalization imposes and .

  • Delocalized eigenstates In this case, . Hence, we expect
  • Localized eigenstates In this case, for sites and zero elsewhere. Hence, we expect
  • Multifractal eigenstates At the transition, namely at the mobility edge, an anomalous scaling is observed elsewhere. Hence, we expect

Here is q-dependent fractal dimension, smaller than .

Transfer matrices and Lyapunov exponents

Product of random variables and Central limit theorem

Consider a set of positive iid random variables with finite mean and variance and compute their product

For large N, the Central Limit Theorem predicts:

  • is a Gaussian number of zero mean and unit variance
  • are N indepent and can be written as

Log-normal distribution

The distribution of is log-normal

Quenched and Annealed averages

For the log-normal distribution the mean is larger than the median value (which is larger than the mode). Hence, is not self averaging, while is self averaging. This is the reason why in the following we will take quenched averages.

Product of random matrices

Let's consider again the Anderson Model in 1d. The eigensystem is well defined in a box of size L with Dirichelet boundary condition on the extremeties of the box.

Here we will solve the second order differential equation imposing instead Cauchy boundaries on one side of the box. Let's rewrite the previous eigensystem in the following form

We can continue the recursion

It is useful to introduce the transfer matrix

and the product of random matrices