Upload
others
View
5
Download
0
Embed Size (px)
Citation preview
Preliminary Monte-Carlo QMC
Introduction to Quantum Monte-Carlo
Francesco Sottile
Ecole Polytechnique and ETSF
ESNUM 9 June 2016
Preliminary Monte-Carlo QMC
Outline
Preminary (statistic) concepts
Monte-Carlo: means, samplings and Markov chains
Quantum Monte-Carlo: variational and diffusion MC
Preliminary Monte-Carlo QMC
Outline
Preminary (statistic) concepts
Monte-Carlo: means, samplings and Markov chains
Quantum Monte-Carlo: variational and diffusion MC
Preliminary Monte-Carlo QMC
Two theorems
Preliminary Monte-Carlo QMC
Two theorems
Law of large numbers
If you perform the same experiment a large number of times, theaverage of the results obtained should be close to the expectedvalue, and will tend to become closer as more trials are performed.
Central limit theorem
The mean of a sufficiently large number of independent randomvariables, each with finite mean and variance, will be approximatelynormally distributed.
Preliminary Monte-Carlo QMC
Two theorems
Law of large numbers
If you perform the same experiment a large number of times, theaverage of the results obtained should be close to the expectedvalue, and will tend to become closer as more trials are performed.
mean or expected value = µ = 〈x〉 =∑j
pjxj =
∫dx p(x) x
variance = σ2 =⟨(x − µ)2
⟩=∑j
pjx2j − µ2 =
∫dx p(x) x2 − µ2
Sn =x1 + x2 + ..+ xn
n−→ µ
Preliminary Monte-Carlo QMC
Two theorems
(Sn − µ) −→ N (0,σ2
n)
Central limit theorem
The mean of a sufficiently large number of independent randomvariables, each with finite mean and variance, will be approximatelynormally distributed.
Preliminary Monte-Carlo QMC
Two theorems
Law of large numbers
If you perform the same experiment a large number of times, theaverage of the results obtained should be close to the expectedvalue, and will tend to become closer as more trials are performed.
Central limit theorem
The mean of a sufficiently large number of independent randomvariables, each with finite mean and variance, will be approximatelynormally distributed.
Preliminary Monte-Carlo QMC
Two theorems
Large numbers + central limit
Sn −→ µ± ξ√n
Preliminary Monte-Carlo QMC
Pseudo-Random Number Generator (PRNG)
Two (+1) requests for good PRNG
• It has to be good: long period, good lattice structure, goodsequences, etc.
• It has to be fast.
* It has to be reproducible
Preliminary Monte-Carlo QMC
Pseudo-Random Number Generator (PRNG)
Two (+1) requests for good PRNG
• It has to be good: long period, good lattice structure, goodsequences, etc.
• It has to be fast.
* It has to be reproducible
Preliminary Monte-Carlo QMC
Random Numbers
• Today’s libraries give reliable uniform random numbers(∈ [0, 1]).
• We are able, by transformation from the uniform distribution,to create random numbers distributed according to other(simple) functions, like the Gaussian.
Preliminary Monte-Carlo QMC
Outline
Preminary (statistic) concepts
Monte-Carlo: means, samplings and Markov chains
Quantum Monte-Carlo: variational and diffusion MC
Preliminary Monte-Carlo QMC
Monte-Carlo sampling
y
A
BC
O x
X
X
X
X
X
X
1
π
4∼ Nhit
Ntot
Preliminary Monte-Carlo QMC
Monte-Carlo sampling
f(x)
O x1
1
x x1 2
π
4∼∫ √
1− x2dx ∼ V
N
N∑i
√1− x2
i
Preliminary Monte-Carlo QMC
Barely relevant Monte-Carlo sampling
I =
∫f (x)dx
advantages
• easy to implement
disadvantages
• converges only like O(
1√N
),
poorly compared to theSimpson’s method O
(1N4
)
Preliminary Monte-Carlo QMC
Barely relevant Monte-Carlo sampling
I =
∫· ·∫
f (x)ddx
advantages
• easy to implement
• converges still like O(
1√N
),
compared to the Simpson’s
method O(
1N4/d
)disadvantages
• We hit a lot of empty(or barely relevant)space
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0f (x)dx =
∫ 1
0
f (x)
p(x)p(x)dx
I =< f >=
⟨f
p
⟩p
How to choose p(x)?
• Choose p(x) to minimize the variance
σ = 0 ←− p(x) = f (x)I quite useless?
p(x) close to f (x), but simple enough to be sampled
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0f (x)dx =
∫ 1
0
f (x)
p(x)p(x)dx
I =< f >=
⟨f
p
⟩p
How to choose p(x)?
• Choose p(x) to minimize the variance
σ = 0 ←− p(x) = f (x)I quite useless?
p(x) close to f (x), but simple enough to be sampled
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0f (x)dx =
∫ 1
0
f (x)
p(x)p(x)dx
I =< f >=
⟨f
p
⟩p
How to choose p(x)?
• Choose p(x) to minimize the variance
σ = 0 ←− p(x) = f (x)I quite useless?
p(x) close to f (x), but simple enough to be sampled
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0
ex − 1
e − 1dx = 0.418
σ1 = 0.3009540σ2 = 0.0560286σ3 = 0.1380024σ4 = 0.1838806
0 0.2 0.4 0.6 0.8 10
0.5
1
1.5
2
2.5
3
f(x) =
ex-1
e - 1
Sample of a function f(x) using different probability distributions
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0
ex − 1
e − 1dx = 0.418
σ1 = 0.3009540σ2 = 0.0560286σ3 = 0.1380024σ4 = 0.1838806
0 0.2 0.4 0.6 0.8 10
0.5
1
1.5
2
2.5
3
f(x) =
ex-1
e - 1
p1(x) = 1
Sample of a function f(x) using different probability distributions
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0
ex − 1
e − 1dx = 0.418
σ1 = 0.3009540σ2 = 0.0560286σ3 = 0.1380024σ4 = 0.1838806
0 0.2 0.4 0.6 0.8 10
0.5
1
1.5
2
2.5
3
f(x) =
ex-1
e - 1
p1(x) = 1
p2(x) = 2 x
Sample of a function f(x) using different probability distributions
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0
ex − 1
e − 1dx = 0.418
σ1 = 0.3009540σ2 = 0.0560286σ3 = 0.1380024σ4 = 0.1838806
0 0.2 0.4 0.6 0.8 10
0.5
1
1.5
2
2.5
3
f(x) =
ex-1
e - 1
p1(x) = 1
p2(x) = 2 x
p3(x) = e
x/(e-1)
Sample of a function f(x) using different probability distributions
Preliminary Monte-Carlo QMC
Importance sampling
I =
∫ 1
0
ex − 1
e − 1dx = 0.418
σ1 = 0.3009540σ2 = 0.0560286σ3 = 0.1380024σ4 = 0.1838806
0 0.2 0.4 0.6 0.8 10
0.5
1
1.5
2
2.5
3
f(x) =
ex-1
e - 1
p1(x) = 1
p2(x) = 2 x
p3(x) = e
x/(e-1)
p4(x) = 3 x
2
Sample of a function f(x) using different probability distributions
Preliminary Monte-Carlo QMC
Importance sampling
√importance sampling is crucial in practice
√it relies on finding p(x)
× many-dimensions complex p(x) are difficult to find andto sample
One solution is Markov chains
Preliminary Monte-Carlo QMC
Importance sampling
√importance sampling is crucial in practice
√it relies on finding p(x)
× many-dimensions complex p(x) are difficult to find andto sample
One solution is Markov chains
Preliminary Monte-Carlo QMC
Importance sampling
√importance sampling is crucial in practice
√it relies on finding p(x)
× many-dimensions complex p(x) are difficult to find andto sample
One solution is Markov chains
Preliminary Monte-Carlo QMC
Importance sampling
√importance sampling is crucial in practice
√it relies on finding p(x)
× many-dimensions complex p(x) are difficult to find andto sample
One solution is Markov chains
Preliminary Monte-Carlo QMC
Markov Chains
Distribution function p(x)
e−βH∫e−βH
;|ψ|2∫|ψ|2
;e−S(x)∫e−S(x)
Two problems to overcome
• Hamiltonians, wavefunctions, actions are complicate(d-dimensional) functions (no way to find an analyticprimitive).
• they are normalized by an integral that has intrinsically thesame difficulty of the main integral
Preliminary Monte-Carlo QMC
Markov Chains
Markov Chains sequence
x1P−→ x2
P−→ x3..P−→ xn
x1, x2, .. random but not independent
Markov Chain operator P(x → y)
It is possible to demonstrate that, no matter how complicate p(x)
• P(x → y) generates a sequence that, at the end, isdistributed according to p(x)
• we don’t need to know P(x → y)
• we don’t need to know p(x), but just a function proportionalto p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Markov Chains sequence
x1P−→ x2
P−→ x3..P−→ xn
x1, x2, .. random but not independent
Markov Chain operator P(x → y)
It is possible to demonstrate that, no matter how complicate p(x)
• P(x → y) generates a sequence that, at the end, isdistributed according to p(x)
• we don’t need to know P(x → y)
• we don’t need to know p(x), but just a function proportionalto p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Markov Chains sequence
x1P−→ x2
P−→ x3..P−→ xn
x1, x2, .. random but not independent
Markov Chain operator P(x → y)
It is possible to demonstrate that, no matter how complicate p(x)
• P(x → y) generates a sequence that, at the end, isdistributed according to p(x)
• we don’t need to know P(x → y)
• we don’t need to know p(x), but just a function proportionalto p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Markov Chains sequence
x1P−→ x2
P−→ x3..P−→ xn
x1, x2, .. random but not independent
Markov Chain operator P(x → y)
It is possible to demonstrate that, no matter how complicate p(x)
• P(x → y) generates a sequence that, at the end, isdistributed according to p(x)
• we don’t need to know P(x → y)
• we don’t need to know p(x), but just a function proportionalto p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Markov Chains sequence
x1P−→ x2
P−→ x3..P−→ xn
x1, x2, .. random but not independent
Markov Chain operator P(x → y)
It is possible to demonstrate that, no matter how complicate p(x)
• P(x → y) generates a sequence that, at the end, isdistributed according to p(x)
• we don’t need to know P(x → y)
• we don’t need to know p(x), but just a function proportionalto p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Simple example: two levels system
Population of cityA and population of cityB.Every year: 40% of people of cityA moves to CityB; 30% of thecontrary. Initially the population is A and B, for cityA and cityB.So the second year will be(
A′
B ′
)=
(0.6A + 0.3B0.4A + 0.7B
)=
(0.6 0.30.4 0.7
)(AB
)
P =
(0.6 0.30.4 0.7
)is the stochastic matrix
Preliminary Monte-Carlo QMC
Markov Chains
Finding the converged stable distribution (of people)
• Iterate the Markov process, applying P to (A,B) to produce(A’,B’), then (A”,B”), etc.
• Considering that, at convergency PX = X . So the convergeddistribution is the eigenvector of the stochastic matrix, relatedto the unitary eigenvalue
This is the case in which we know the stochastic matrix and wefind the final distribution function.
But we want to generate sequences, according to a known finaldistribution function, without knowing P.
Preliminary Monte-Carlo QMC
Markov Chains
Finding the converged stable distribution (of people)
• Iterate the Markov process, applying P to (A,B) to produce(A’,B’), then (A”,B”), etc.
• Considering that, at convergency PX = X . So the convergeddistribution is the eigenvector of the stochastic matrix, relatedto the unitary eigenvalue
This is the case in which we know the stochastic matrix and wefind the final distribution function.
But we want to generate sequences, according to a known finaldistribution function, without knowing P.
Preliminary Monte-Carlo QMC
Markov Chains
Finding the converged stable distribution (of people)
• Iterate the Markov process, applying P to (A,B) to produce(A’,B’), then (A”,B”), etc.
• Considering that, at convergency PX = X . So the convergeddistribution is the eigenvector of the stochastic matrix, relatedto the unitary eigenvalue
This is the case in which we know the stochastic matrix and wefind the final distribution function.
But we want to generate sequences, according to a known finaldistribution function, without knowing P.
Preliminary Monte-Carlo QMC
Markov Chains
Detailed balance principle or microreversibility
It can be demonstrated that any stochastic matrix P converges toa distribution function p(x) if
p(x)P(x → y) = p(y)P(y → x)
Missing: how to conceive P, in order to generate this sequence.Missing: here it seems we have to know p(x).
Preliminary Monte-Carlo QMC
Markov Chains
Detailed balance principle or microreversibility
It can be demonstrated that any stochastic matrix P converges toa distribution function p(x) if
p(x)P(x → y) = p(y)P(y → x)
Missing: how to conceive P, in order to generate this sequence.Missing: here it seems we have to know p(x).
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
Method to generate a microreversible P(x → y)
• We are at x
• We propose a trial move xT according to a symmetricprobability distribution F (x → xT ) = F (xT → x)
• Accept the trial move xT (and so put y = xT ) with
probability min(
1, p(xT )p(x)
)
we don’t need the exact p(x),but just a function proportional to αp(x)
Preliminary Monte-Carlo QMC
Metropolis method
In practice
• F (x → xT ) is a Gaussian centered on x .
σ dynamically adjusted
• Accepting a trial move with probability p(xT )p(x) ?
Get a random ξ ∈ [0, 1].
Accept if p(xT )p(x) > ξ
Preliminary Monte-Carlo QMC
Metropolis method
In practice
• F (x → xT ) is a Gaussian centered on x .
σ dynamically adjusted
• Accepting a trial move with probability p(xT )p(x) ?
Get a random ξ ∈ [0, 1].
Accept if p(xT )p(x) > ξ
Preliminary Monte-Carlo QMC
Metropolis method
The M(RT)2 method is today used inmany different application,ranging fromnon-linear differential equations, tosimulation of galaxy formations: whatabout electronic structure calculation?
Preliminary Monte-Carlo QMC
Metropolis method
The M(RT)2 method is today used inmany different application,ranging fromnon-linear differential equations, tosimulation of galaxy formations: whatabout electronic structure calculation?
Preliminary Monte-Carlo QMC
Outline
Preminary (statistic) concepts
Monte-Carlo: means, samplings and Markov chains
Quantum Monte-Carlo: variational and diffusion MC
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Method to calculate the exact values of certain (ground-state)properties.
• Capable to reach high accuracy
• Wavefunction sampling is an alternative to brute forcewave-function representation (CI, CC) with advantages anddisadvantages
• QMC better scaling (N3 vs N6)• QMC subject to statistical errors
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Method to calculate the exact values of certain (ground-state)properties.
• Capable to reach high accuracy
• Wavefunction sampling is an alternative to brute forcewave-function representation (CI, CC) with advantages anddisadvantages
• QMC better scaling (N3 vs N6)• QMC subject to statistical errors
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Method to calculate the exact values of certain (ground-state)properties.
• Capable to reach high accuracy
• Wavefunction sampling is an alternative to brute forcewave-function representation (CI, CC) with advantages anddisadvantages
• QMC better scaling (N3 vs N6)• QMC subject to statistical errors
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Method to calculate the exact values of certain (ground-state)properties.
• Capable to reach high accuracy
• Wavefunction sampling is an alternative to brute forcewave-function representation (CI, CC) with advantages anddisadvantages
• QMC better scaling (N3 vs N6)• QMC subject to statistical errors
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Method to calculate the exact values of certain (ground-state)properties.
• Capable to reach high accuracy
• Wavefunction sampling is an alternative to brute forcewave-function representation (CI, CC) with advantages anddisadvantages
• QMC better scaling (N3 vs N6)• QMC subject to statistical errors
Preliminary Monte-Carlo QMC
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Variational Monte-Carlo
• Diffusion Monte-Carlo
• Path Integral Monte-Carlo, Reptation Monte-Carlo, Green’sfunctions Monte-Carlo
Preliminary Monte-Carlo QMC
Quantum Monte-Carlo
• Variational Monte-Carlo
• Diffusion Monte-Carlo
• Path Integral Monte-Carlo, Reptation Monte-Carlo, Green’sfunctions Monte-Carlo
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
Variational Theorem
Given
〈E 〉 =
∫dx ψ∗(x)Hψ(x)∫dx ψ∗(x)ψ(x)
the variational theorem states that 〈E 〉 ≥ E0.And 〈E 〉 = E0 if and only if ψ ∝ φ0
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
Idea of VMC
Let’s consider a trial ψT (x , {α}).∫dx ψ∗T (x , {α})HψT (x , {α})∫dx ψ∗T (x , {α})ψT (x , {α})
= E ({α}) ≥ E0
Minimizing E ({α}), with respect to the paramaters {α} will givean (upper) estimate of E0
Of course, we will use Monte-Carlo methods to calculate the3N-dimensional integrals
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
Idea of VMC
Let’s consider a trial ψT (x , {α}).∫dx ψ∗T (x , {α})HψT (x , {α})∫dx ψ∗T (x , {α})ψT (x , {α})
= E ({α}) ≥ E0
Minimizing E ({α}), with respect to the paramaters {α} will givean (upper) estimate of E0
Of course, we will use Monte-Carlo methods to calculate the3N-dimensional integrals
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
Idea of VMC
Let’s consider a trial ψT (x , {α}).∫dx ψ∗T (x , {α})HψT (x , {α})∫dx ψ∗T (x , {α})ψT (x , {α})
= E ({α}) ≥ E0
Minimizing E ({α}), with respect to the paramaters {α} will givean (upper) estimate of E0
Of course, we will use Monte-Carlo methods to calculate the3N-dimensional integrals
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
Idea of VMC
Let’s consider a trial ψT (x , {α}).∫dx ψ∗T (x , {α})HψT (x , {α})∫dx ψ∗T (x , {α})ψT (x , {α})
= E ({α}) ≥ E0
Minimizing E ({α}), with respect to the paramaters {α} will givean (upper) estimate of E0
Of course, we will use Monte-Carlo methods to calculate the3N-dimensional integrals
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
What we don’t do
Naıvely we might uniformly sample ψHψ and ψψ for the twointegrals. ∫
dx ψ∗THψT ;
∫dx ψ∗TψT
for any {α}.
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
What we do: importance sampling
〈E 〉 =
∫dx ψ∗THψT∫dx |ψ∗T |2
=
∫dx |ψT |2 HψT
ψT∫dx |ψT |2
=
〈E 〉 =
∫dx ρ(x , {α})EL(x , {α})
EL(x , {α}) =HψT (x , {α})ψT (x , {α})
ρ(x , {α}) =|ψT (x , {α})|2∫dx |ψT (x , {α})|2
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
VMC in practice
1. Let’s generate a number of copies of the system, each one withdifferent (random) electron coordinates: x ’s (the walkers).
2. Let’s choose a form for the ψT (x , {α})
3. Let’s use the Metropolis method to propagate such walkers.
4. We can monitor some observables during the Markov chain, likelocal energy, variance, etc.
5. When the walkers are distributed like |ψT |2, say at step L, wecalculate the local energy
E =1
N
L+N∑i=L
EL(xi , {α})
6. We change now α and we go to step 3
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
VMC in practice
1. Let’s generate a number of copies of the system, each one withdifferent (random) electron coordinates: x ’s (the walkers).
2. Let’s choose a form for the ψT (x , {α})
3. Let’s use the Metropolis method to propagate such walkers.
4. We can monitor some observables during the Markov chain, likelocal energy, variance, etc.
5. When the walkers are distributed like |ψT |2, say at step L, wecalculate the local energy
E =1
N
L+N∑i=L
EL(xi , {α})
6. We change now α and we go to step 3
Preliminary Monte-Carlo QMC
The trial wavefunction
Antisymmetric function for fermions
ψT (x) = D(x)J (x) =
ψT (x) =few∑ν
cν det[ψ↑ν,n(ri )
]det[ψ↓ν,m(rj)
]e−V (x).
V (x) =∑i
V1(ri ) +∑i ,j>i
V2(rij),
Preliminary Monte-Carlo QMC
The trial wavefunction
Jastrow factor for jellium spheres
J = exp
(N∑i=1
V1(ri )
)exp
N∑i<j
V2(rij )
exp(V (N)
)
V1(ri ) =20∑n=1
α(i)n j0
(nβri
)
V(λ)2 (rij ) =
a(λ)rij + c(λ)r2ij + e(λ)r3
ij
1 + b(λ)(ri )rij + d (λ)r2ij + r3
ij
with
b(λ)(ri ) = b(λ)0 + b
(λ)1 arctan
[ r2i − R2
b
K (λ)
]λ = A,P (antiparallel and parallel spins), and j0 is the zero-order spherical Bessel
function(j0(x) = sin x
x
).
V (N) = γ(PC )2 + δ(PS )2
PC =N∑i
ri e PS = 2N∑i
riS(i)z
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
• Variational Monte-Carlo gives high quality results (recover90% of correlation energy)
• but it is still approximate (relies on the choice of the trialwavefunction)
We now want extremely accurate (exact) results for theground-state energy
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
• Variational Monte-Carlo gives high quality results (recover90% of correlation energy)
• but it is still approximate (relies on the choice of the trialwavefunction)
We now want extremely accurate (exact) results for theground-state energy
Preliminary Monte-Carlo QMC
Variational Monte-Carlo
• Variational Monte-Carlo gives high quality results (recover90% of correlation energy)
• but it is still approximate (relies on the choice of the trialwavefunction)
We now want extremely accurate (exact) results for theground-state energy
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
ψ(t) =∑n
cne− i
~Entφn
Hφn = Enφn
cn =
∫dx φn(x)ψ(x , 0), n = 0, 1, 2, ..
In imaginary time,
ψ(τ) = c0e−E0τφ0 + c1e
−E1τφ1 + c2e−E2τφ2 + ..
τ→∞−−−→∝ φ0
We want a practical scheme to do this imaginary time evolutionand recover the ground-state energy
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
ψ(t) =∑n
cne− i
~Entφn
Hφn = Enφn
cn =
∫dx φn(x)ψ(x , 0), n = 0, 1, 2, ..
In imaginary time,
ψ(τ) = c0e−E0τφ0 + c1e
−E1τφ1 + c2e−E2τφ2 + ..
τ→∞−−−→∝ φ0
We want a practical scheme to do this imaginary time evolutionand recover the ground-state energy
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
ψ(t) =∑n
cne− i
~Entφn
Hφn = Enφn
cn =
∫dx φn(x)ψ(x , 0), n = 0, 1, 2, ..
In imaginary time,
ψ(τ) = c0e−E0τφ0 + c1e
−E1τφ1 + c2e−E2τφ2 + ..
τ→∞−−−→∝ φ0
We want a practical scheme to do this imaginary time evolutionand recover the ground-state energy
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
ψ(t) =∑n
cne− i
~Entφn
Hφn = Enφn
cn =
∫dx φn(x)ψ(x , 0), n = 0, 1, 2, ..
In imaginary time,
ψ(τ) = c0e−E0τφ0 + c1e
−E1τφ1 + c2e−E2τφ2 + ..
τ→∞−−−→∝ φ0
We want a practical scheme to do this imaginary time evolutionand recover the ground-state energy
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
First step: shift of energy
i~∂ψ(x , t)
∂t=
[− ~2
2m∇2 + (V (x)− ET )
]ψ(x , t)
ψ(x , t) =∑n
cne− i
~ (En−ET )tφn(x)
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
Second step: Wick rotation in time
~∂ψ(x , τ)
∂τ=
[− ~2
2m∇2 + (V (x)− ET )
]ψ(x , τ)
ψ(x , τ) =∑n
cne− (En−ET )
~ tφn(x)
Role of ET
• ET > E0 the wavefunction will diverge exponentially fast
• ET < E0 the wavefunction will vanish exponentially fast
• ET = E0 the wavefunction will exponentially converge to φ0!
We want a practical method that,starting from an initial wave-function, performs an imaginarytime iteration, permitting succes-sive adjustements to ET , such thatat the end, the stationary solutioncorresponds to ET (τ →∞) = E0
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
Second step: Wick rotation in time
~∂ψ(x , τ)
∂τ=
[− ~2
2m∇2 + (V (x)− ET )
]ψ(x , τ)
ψ(x , τ) =∑n
cne− (En−ET )
~ tφn(x)
Role of ET
• ET > E0 the wavefunction will diverge exponentially fast
• ET < E0 the wavefunction will vanish exponentially fast
• ET = E0 the wavefunction will exponentially converge to φ0!
We want a practical method that,starting from an initial wave-function, performs an imaginarytime iteration, permitting succes-sive adjustements to ET , such thatat the end, the stationary solutioncorresponds to ET (τ →∞) = E0
Preliminary Monte-Carlo QMC
Diffusion Monte-Carlo
Second step: Wick rotation in time
~∂ψ(x , τ)
∂τ=
[− ~2
2m∇2 + (V (x)− ET )
]ψ(x , τ)
ψ(x , τ) =∑n
cne− (En−ET )
~ tφn(x)
Role of ET
• ET > E0 the wavefunction will diverge exponentially fast
• ET < E0 the wavefunction will vanish exponentially fast
• ET = E0 the wavefunction will exponentially converge to φ0!
We want a practical method that,starting from an initial wave-function, performs an imaginarytime iteration, permitting succes-sive adjustements to ET , such thatat the end, the stationary solutioncorresponds to ET (τ →∞) = E0
Preliminary Monte-Carlo QMC
DMC: practical scheme
First step: generation of walkers
Let’s generate Nw replicas of the systems sampled from the initialwavefunction ψT (x , 0)
ψ(x , 0) =Nw∑i=1
wiδ(x − xi )
Second step: writing the propagator
The integral form of the imaginary time Schrodinger equationinvolves the Green’s function
ψ(x ′, τ + δτ) =
∫dx G (x , x ′, δτ)ψ(x , τ)
Preliminary Monte-Carlo QMC
DMC: practical scheme
First step: generation of walkers
Let’s generate Nw replicas of the systems sampled from the initialwavefunction ψT (x , 0)
ψ(x , 0) =Nw∑i=1
wiδ(x − xi )
Second step: writing the propagator
The integral form of the imaginary time Schrodinger equationinvolves the Green’s function
ψ(x ′, τ + δτ) =
∫dx G (x , x ′, δτ)ψ(x , τ)
Preliminary Monte-Carlo QMC
DMC: practical scheme
The propagator G
• Only diffusive term
∂ψ(x , τ)
∂τ= −D ∇2ψ(x , τ)
GD(x , x ′, δτ) = e−(x−x′)2
2δτ
• Only rate-term (branching)
∂ψ(x , τ)
∂τ= (V (x)− ET )ψ(x , τ)
GB(x , x , δτ) = e−(V (x)−ET )δτ
M = int[e−(V (x)−ET )δτ + ξ
]
Preliminary Monte-Carlo QMC
DMC: practical scheme
The propagator G
• Only diffusive term
∂ψ(x , τ)
∂τ= −D ∇2ψ(x , τ)
GD(x , x ′, δτ) = e−(x−x′)2
2δτ
• Only rate-term (branching)
∂ψ(x , τ)
∂τ= (V (x)− ET )ψ(x , τ)
GB(x , x , δτ) = e−(V (x)−ET )δτ
M = int[e−(V (x)−ET )δτ + ξ
]
Preliminary Monte-Carlo QMC
DMC: practical scheme
The propagator G
∂ψ(x , τ)
∂τ=[−D ∇2 + (V (x)− ET )
]ψ(x , τ)
G (x , x ′, δτ) = GD(x , x ′, δτ) · GB(x , x ′, δτ) +O(δτ)2
= e−(x − x ′)2
2δτ− (V (x)− ET ) δτ
+O(δτ)2
• diffuse a walker, and accept x ′ with probability
min(
1, ψ(x ′)G(x ,x ′)ψ(x)G(x ′,x)
)• remove or proliferate the walker according to the multiplicity,
calculated with the branching term
Preliminary Monte-Carlo QMC
DMC: practical scheme
The propagator G
∂ψ(x , τ)
∂τ=[−D ∇2 + (V (x)− ET )
]ψ(x , τ)
G (x , x ′, δτ) = GD(x , x ′, δτ) · GB(x , x ′, δτ) +O(δτ)2
= e−(x − x ′)2
2δτ− (V (x)− ET ) δτ
+O(δτ)2
• diffuse a walker, and accept x ′ with probability
min(
1, ψ(x ′)G(x ,x ′)ψ(x)G(x ′,x)
)• remove or proliferate the walker according to the multiplicity,
calculated with the branching term
Preliminary Monte-Carlo QMC
DMC: practical scheme
First step: generation of walkers
Let’s generate Nw replicas of the systems sampled from the initialwavefunction ψT (x , 0)
ψ(x , 0) =Nw∑i=1
wiδ(x − xi )
Second step: writing the propagator
The integral form of the imaginary time Schrodinger equationinvolves the Green’s function
ψ(x ′, τ + δτ) =
∫dx G (x , x ′, δτ)ψ(x , τ)
Preliminary Monte-Carlo QMC
DMC: practical scheme
Third step: calculate quantity of interest
Calculate quantity of interest (at this step) averaging on thewalkers.
E0(x , τ) =Nw∑i=1
EL(xi , τ) =Hψ(x , τ)
ψ(x , τ)
Fourth step: adjust trial energy
EnewT =
ET + E0(x , τ)
2
We continue to propagate until E0 = ET , exact result for theground-state energy.
Preliminary Monte-Carlo QMC
DMC: practical scheme
This is exactly how things ...do not work
Preliminary Monte-Carlo QMC
DMC: two issues
Fluctuations
Branching term causes large fluctuations in the number of walkers,preventing convergency. Solution: importance sampling.
Interpretation
ψT has to be positive everywhere (which is not the case) to beinterpreted as walkers distribution density. Solution: fixed-nodesapproximation.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
f (x , τ) = ψT (x)ψ(x , τ)
f (x , 0) = |ψT (x)|2
so let’s get some walkers from our previous VariationalMonte-Carlo calculation.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
f (x , τ) = ψT (x)ψ(x , τ)
f (x , 0) = |ψT (x)|2
so let’s get some walkers from our previous VariationalMonte-Carlo calculation.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
−∂f (x , τ)
∂τ=[−D∇2 + (EL(x)− ET )
]f (x , τ) + D∇ [f (x , τ)v(x)]
with
EL(x) =HψT
ψTand v(x) =
∇ψT
ψT
• Branching term is now related to the local energy, rather thanto the potential.
• A new term appears, a drift term, for which the relativeGreen’s function can be easily evaluated, i.e.G (x , x ′, δτ) = δ(x − x ′ − v(x)δτ).
so our final equation becomes a drifted diffusion process (Brownianmotion within an external field) + branching.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
−∂f (x , τ)
∂τ=[−D∇2 + (EL(x)− ET )
]f (x , τ) + D∇ [f (x , τ)v(x)]
with
EL(x) =HψT
ψTand v(x) =
∇ψT
ψT
• Branching term is now related to the local energy, rather thanto the potential.
• A new term appears, a drift term, for which the relativeGreen’s function can be easily evaluated, i.e.G (x , x ′, δτ) = δ(x − x ′ − v(x)δτ).
so our final equation becomes a drifted diffusion process (Brownianmotion within an external field) + branching.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
−∂f (x , τ)
∂τ=[−D∇2 + (EL(x)− ET )
]f (x , τ) + D∇ [f (x , τ)v(x)]
with
EL(x) =HψT
ψTand v(x) =
∇ψT
ψT
• Branching term is now related to the local energy, rather thanto the potential.
• A new term appears, a drift term, for which the relativeGreen’s function can be easily evaluated, i.e.G (x , x ′, δτ) = δ(x − x ′ − v(x)δτ).
so our final equation becomes a drifted diffusion process (Brownianmotion within an external field) + branching.
Preliminary Monte-Carlo QMC
DMC: Importance Sampling
−∂f (x , τ)
∂τ=[−D∇2 + (EL(x)− ET )
]f (x , τ) + D∇ [f (x , τ)v(x)]
with
EL(x) =HψT
ψTand v(x) =
∇ψT
ψT
• Branching term is now related to the local energy, rather thanto the potential.
• A new term appears, a drift term, for which the relativeGreen’s function can be easily evaluated, i.e.G (x , x ′, δτ) = δ(x − x ′ − v(x)δτ).
so our final equation becomes a drifted diffusion process (Brownianmotion within an external field) + branching.
Preliminary Monte-Carlo QMC
DMC: Walkers evolution
Preliminary Monte-Carlo QMC
DMC: Fixed-nodes approximation
Fixed-nodes: howto?
f (x , τ) = ψT (x)ψ(x , τ) ≥ 0
This is possible if the nodes(ψ(x , τ))=nodes(ψT (x)) all along theimaginary time evolution.Walker refusal= during the diffusive-drifted process a newposition is proposed for the walker. If this position changes thesign of the wave-function, the move is refused.This is an approximation and implies a (small) error.
Preliminary Monte-Carlo QMC
DMC: Fixed-nodes approximation
Fixed-nodes: howto?
f (x , τ) = ψT (x)ψ(x , τ) ≥ 0
This is possible if the nodes(ψ(x , τ))=nodes(ψT (x)) all along theimaginary time evolution.Walker refusal= during the diffusive-drifted process a newposition is proposed for the walker. If this position changes thesign of the wave-function, the move is refused.This is an approximation and implies a (small) error.
Preliminary Monte-Carlo QMC
DMC: Fixed-nodes approximation
Fixed-nodes: howto?
f (x , τ) = ψT (x)ψ(x , τ) ≥ 0
This is possible if the nodes(ψ(x , τ))=nodes(ψT (x)) all along theimaginary time evolution.Walker refusal= during the diffusive-drifted process a newposition is proposed for the walker. If this position changes thesign of the wave-function, the move is refused.This is an approximation and implies a (small) error.
Preliminary Monte-Carlo QMC
DMC: Fixed-nodes approximation
Fixed-nodes: howto?
f (x , τ) = ψT (x)ψ(x , τ) ≥ 0
This is possible if the nodes(ψ(x , τ))=nodes(ψT (x)) all along theimaginary time evolution.Walker refusal= during the diffusive-drifted process a newposition is proposed for the walker. If this position changes thesign of the wave-function, the move is refused.This is an approximation and implies a (small) error.
Preliminary Monte-Carlo QMC
DMC: Fixed-nodes approximation
• FN-DMC uniquely depends on the nodes of the trialwave-function
• Even within Fixed-Nodes the accuracy of DMC is very high,and comparable to much more cumbersom methods like CI orCC
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
DMC: Can we release the nodes?• An antisymmetric function can be written as a difference
between to positive functions, f1 and f2
• Let’s associate a set of Walkers to f1, called W1, and a set ofWalkers to f2, called W2
• The Schrodinger eq. is linear, so we can perform imaginarytime iteration of these two set of walkers separately.
• The W1 and W2 wavefunctions have a not-negligiblesuperposition with the bosonic ground-state, lower in energy.
• If no numerical errors, the bosonic part of W1 and W2 cancelout, and we have an exact result.
• Since numerical errors are present, after few steps ⇒ bosoniccatastrophe
Preliminary Monte-Carlo QMC
Fermion MC
Preliminary Monte-Carlo QMC
Fermion Sign Problem: unresolved
Several solution / year BUT
• They have errors
• They are known not to work
• They have uncontrolled approximations
• Scaling not demonstrated
Preliminary Monte-Carlo QMC
Fermion Sign Problem: unresolved
Several solution / year BUT
• They have errors
• They are known not to work
• They have uncontrolled approximations
• Scaling not demonstrated
Preliminary Monte-Carlo QMC
Fermion Sign Problem: unresolved
Several solution / year BUT
• They have errors
• They are known not to work
• They have uncontrolled approximations
• Scaling not demonstrated
Preliminary Monte-Carlo QMC
Fermion Sign Problem: unresolved
Several solution / year BUT
• They have errors
• They are known not to work
• They have uncontrolled approximations
• Scaling not demonstrated
Preliminary Monte-Carlo QMC
Fermion Sign Problem: unresolved
Several solution / year BUT
• They have errors
• They are known not to work
• They have uncontrolled approximations
• Scaling not demonstrated
Preliminary Monte-Carlo QMC
QMC for solids: many issues
• Bloch theorem only valid in one-particle theories, not formany-body wavefunctions.
Consequences: twisted boundary conditions, supercells,finite-size error, Ewald sums not exact.
• Kinetic energy is large for deep electron, so pseudo-potentialare mandatory.
• Non-local pseudo-potential worsen the fermion-sign problem.