Next: 3.3 Variational Monte Carlo
Up: 3. Quantum Monte Carlo
Previous: 3.1 Introduction
  Contents
Subsections
Monte Carlo methods were first developed as a method for estimating
integrals that could not be evaluated analytically. Although many
statistical techniques are now included in the category of ``Monte
Carlo methods''[16,17], the method
used in this thesis is principally Monte Carlo integration.
A straightforward application of Monte Carlo is the evaluation of
definite integrals. Consider the one dimensional integral
|
(3.1) |
By application of the mean value theorem of calculus, the integral may
be approximated by
|
(3.2) |
where the points fully cover the range of integration. In the
limit of a large number of points , tends to the exact value
.
A conventional choice for the points would be a uniform
grid. More accurate methods of numeric quadrature, such as Simpson's
rule and Gaussian quadrature[18]
use a weighted average of the points:
|
(3.3) |
These methods are highly effective for low dimensional integrals. The
computational cost of evaluating an integral, to a fixed accuracy, of
dimensionality increases as . An alternate and more efficient
approach is to select the points randomly, from a given
probability distribution by Monte Carlo methods.
If points are selected at random over the interval , the Monte
Carlo estimate of the integral, equation
, becomes
|
(3.4) |
where denotes the mean value of over the set of
sampled points . By the central limit
theorem, [19] the set of all possible sums over
different will have a Gaussian distribution. The standard
deviation of the different values of is a measure of
the uncertainty in the integral's value
|
(3.5) |
The probability that is within
is
and the probability of being with is . This
error decays as independent of the dimensionality
of the integral, unlike grid based methods which have a strong
dimensional dependence.
Simple schemes for Monte Carlo integration can suffer from low
efficiency. Many functions of interest have significant weight in only
a few regions. For example, most of the contributions to an integral
of a simple Gaussian are located near the central peak. In a simple
Monte Carlo integration scheme, points are sampled uniformly, wasting
considerable effort sampling the tails of the Gaussian. Techniques for
overcoming this problem act to increase the density of points in
regions of interest and hence improve the overall efficiency. These
techniques are called importance sampling.
Points are sampled from a non-uniform
distribution , where is always positive and chosen to
approximate over the region of interest. The integral may now
be evaluated by selecting points from the probability distribution
,
|
(3.6) |
so that the integral may be evaluated
|
(3.7) |
where
and the points are generated from the
probability distribution .
Provided is well described by , remains close to
unity, so that the standard deviation defined in equation
3.5 is greatly reduced. (The best possible choice
for is .) The integration procedure is therefore more
efficient, although the error still decays as .
In this thesis, the Metropolis algorithm[20] is
used in the importance sampling of multi-dimensional integrals. The
Metropolis algorithm generates a random walk of points
distributed according to a required probability distribution. From an
initial ``position'' in phase or configuration space, a proposed ``move''
is generated and the move either accepted or rejected according to the
Metropolis algorithm. By taking a sufficient number of trial steps all
of phase space is explored and the Metropolis algorithm ensures that
the points are distributed according to the required probability
distribution.
The Metropolis algorithm may be derived by considering the
probability density, at two points in configuration space and
. Configuration space is specified by the integration
variables and limits. For one dimensional integration, for
. Averaged over many trial steps, the
average number of samples at and should be equal,
.
The probability for a trial move from to defined
and we require
|
(3.8) |
If the probability of accepting a move from to is
then the total probability of accepting a
move from to is
. Therefore, at equilibrium
|
(3.9) |
The Metropolis algorithm corresponds to choosing
|
(3.10) |
For the Metropolis algorithm to be valid, it is essential that the
random walk is ergodic, that is any point in configuration
space may be reached from any other point . In some applications of
the Metropolis algorithm, parts of configuration space may be
difficult to reach. Long simulations or a modification of the
algorithm are then necessary. Many methods have been developed to cope
with this ``slowing down'', [16] but for the
applications presented here the Metropolis algorithm has been found
both adequate and sufficient for the problems investigated.
Next: 3.3 Variational Monte Carlo
Up: 3. Quantum Monte Carlo
Previous: 3.1 Introduction
  Contents
© Paul Kent