background image

THE BEGINNING 

of the

MONTE CARLO METHOD

by N. Metropolis

T

he year was 1945. Two earth-

shaking events took place: the
successful test at Alamogordo
and the building of the first elec-

tronic computer. Their combined impact
was to modify qualitatively the nature of
global interactions between Russia and
the West. No less perturbative were the
changes wrought in all of academic re-
search and in applied science. On a less
grand scale these events brought about a
renascence of a mathematical technique
known to the old guard as statistical sam-
pling; in its new surroundings and owing
to its nature, there was no denying its new
name of the Monte Carlo method.

This essay attempts to describe the de-

tails that led to this renascence and the
roles played by the various actors. It is
appropriate that it appears in an issue ded-
icated to Stan Ulam.

Los Alamos Science  Special Issue 1987

Some Background

Most of us have grown so blase about

computer developments and capabilities
-even some that are spectacular窶杯hat
it is difficult to believe or imagine there
was a time when we suffered the noisy,
painstakingly slow, electromechanical de-
vices that chomped away on punched
cards. Their saving grace was that they
continued working around the clock, ex-
cept for maintenance and occasional re-
pair (such as removing a dust particle
from a relay gap). But these machines
helped enormously with the routine, rela-
tively simple calculations that led to Hi-
roshima.

The ENIAC. 

During this wartime pe-

riod, a team of scientists, engineers, and
technicians was working furiously on the

first electronic computer窶杯he ENIAC窶
at the University of Pennsylvania in Phil-
adelphia. Their mentors were Physicist
First Class John Mauchly and Brilliant
Engineer Presper Eckert. Mauchly, fa-
miliar with Geiger counters in physics
laboratories, had realized that if electronic
circuits could count, then they could do
arithmetic and hence solve, inter 

alia, 

dif-

ference equations窶蚤t almost incredible
speeds! When he窶囘 seen a seemingly
limitless array of women cranking out
firing tables with desk calculators, he窶囘
been inspired to propose to the Ballistics
Research Laboratory at Aberdeen that an

electronic computer be built to deal with
these calculations.

John von Neumann, Professor of Math-

ematics at the Institute for Advanced
Study, was a consultant to Aberdeen and
to Los Alamos.

For a whole host of

125

background image

Monte Carlo

reasons, he had become seriously inter-
ested in the thermonuclear problem being
spawned at that time in Los Alamos by
a friendly fellow-Hungarian scientist, Ed-
ward Teller, and his group. Johnny (as he
was affectionately called) let it be known

that construction of the ENIAC was near-
ing completion, and he wondered whether
Stan Frankel and I would be interested
in preparing a preliminary computational
model of a thermonuclear reaction for the
ENIAC. He felt he could convince the
authorities at Aberdeen that our problem
could provide a more exhaustive test of

the computer than mere firing-table com-
putations. (The designers of the ENIAC
had wisely provided for the capability of
much more ambitious versions of firing
tables than were being arduously com-
puted by hand, not to mention other quite
different applications.) Our response to
von Neumann窶冱 suggestion was enthusi-
astic, and his heuristic arguments were
accepted by the authorities at Aberdeen.

In March, 1945, Johnny, Frankel, and I

visited the Moore School of Electrical En-
gineering at the University of Pennsylva-
nia for an advance glimpse of the ENIAC.

We were impressed. Its physical size
was overwhelming窶敗ome 18,000 double

triode vacuum tubes in a system with
500,000 solder joints. No one ever had

such a wonderful toy!

The staff was dedicated and enthusi-

astic; the friendly cooperation is still re-
membered. The prevailing spirit was akin
to that in Los Alamos. What a pity that a
war seems necessary to launch such revo-
lutionary scientific endeavors. The com-

ponents used in the ENIAC were joint-
army-navy (JAN) rejects. This fact not
only emphasizes the genius of Eckert and
Mauchly and their staff, but also suggests
that the ENIAC was technically realizable
even before we entered the war in Decem-
ber, 1941.

After becoming saturated with indoc-

trination about the general and detailed
structure of the ENIAC, Frankel and I re-
turned to Los Alamos to work on a model

126

that was realistically calculable. (There
was a small interlude at Alamogordo!)
The war ended before we completed our
set of problems, but it was agreed that we

continue working.

Anthony Turkevich

joined the team and contributed substan-

tially to all aspects of the work. More-
over, the uncertainty of the first phase of
the postwar Los Alamos period prompted
Edward Teller to urge us not only to com-
plete the thermonuclear computations but
to document and provide a critical review
of the results.

The Spark. 

The review of the ENIAC

results was held in the spring of 1946
at Los Alamos. In addition to Edward
Teller, the principals included Enrico Fer-
mi, John von Neumann, and the Direc-
tor, Norris Bradbury. Stanley Frankel,
Anthony Turkevich, and I described the
ENIAC, the calculations, and the con-
clusions. Although the model was rel-
atively simple, the simplifications were
taken into account and the extrapolated

results were cause for guarded optimism
about the feasibility of a thermonuclear
weapon.

Among the attendees was Stan Ulam,

who had rejoined the Laboratory after
a brief time on the mathematics faculty
at the University of Southern California.
Ulam窶冱 personality would stand out in
any community, even where 窶彡haracters窶

abounded. His was an informal nature; he
would drop in casually, without the usual
amenities. He preferred to chat, more or
less at leisure, rather than to dissertate.
Topics would range over mathematics,

physics, world events, local news, games
of chance, quotes from the classics窶蚤ll
treated somewhat episodically but always
with a meaningful point. His was a mind
ready to provide a critical link.

During his wartime stint at the Labora-

tory, Stan had become aware of the elec-
tromechanical computers used for implo-
sion studies, so he was duly impressed,
along with many other scientists, by the
speed and versatility of the ENIAC. In ad-

Stanislaw Ulam

dition, however, Stan窶冱 extensive mathe-
matical background made him aware that
statistical sampling techniques had fallen
into desuetude because of the length and
tediousness of the calculations. But with

this miraculous development of the
ENIAC窶蚤long with the applications Stan
must have been pondering窶琶t occurred to
him that statistical techniques should be
resuscitated, and he discussed this idea
with von Neumann. Thus was triggered
the spark that led to the Monte Carlo
method.

The Method

The spirit of this method was consis-

tent with Stan窶冱 interest in random pro-
cesses窶杷rom the simple to the sublime.
He relaxed playing solitaire; he was stim-
ulated by playing poker; he would cite

the times he drove into a filled parking
lot at the same moment someone was ac-
commodatingly leaving. More seriously,
he created the concept of 窶徑ucky num-
bers,窶 whose distribution was much like
that of prime numbers; he was intrigued
by the theory of branching processes and

background image

Monte Carlo

contributed much to its development, in-
cluding its application during the war to
neutron multiplication in fission devices.
For a long time his collection of research
interests included pattern development in
two-dimensional games played according
to very simple rules. Such work has lately
emerged as a cottage industry known as
cellular automata.

John von Neumann saw the relevance

of Ulam窶冱 suggestion and, on March 11,

1947, sent a handwritten letter to Robert

Richtmyer, the Theoretical Division lead-
er (see 窶彜tan Ulam, John von Neumann,
and the Monte Carlo Method窶). His let-
ter included a detailed outline of a pos-
sible statistical approach to solving the
problem of neutron diffusion in fission-
able material.

Johnny窶冱 interest in the method was

contagious and inspiring. His seemingly
relaxed attitude belied an intense interest
and a well-disguised impatient drive. His
talents were so obvious and his coopera-

tive spirit so stimulating that he garnered
the interest of many of us. It was at that
time that I suggested an obvious name
for the statistical method窶蚤 suggestion
not unrelated to the fact that Stan had an
uncle who would borrow money from rel-
atives because he 窶徊ust had to go to Monte
Carlo.窶 The name seems to have endured.

The spirit of Monte Carlo is best con-

veyed by the example discussed in von
Neumann窶冱 letter to Richtmyer. Consider
a spherical core of fissionable material

surrounded by a shell of tamper material.
Assume some initial distribution of neu-

trons in space and in velocity but ignore
radiative and hydrodynamic effects. The

idea is to now follow the development
of a large number of individual neutron
chains as a consequence of scattering, ab-
sorption, fission, and escape.

At each stage a sequence of decisions

has to be made based on statistical prob-
abilities appropriate to the physical and
geometric factors. The first two decisions
occur at time 

t = O, 

when a neutron is se-

lected to have a certain velocity and a cer-

tain spatial position. The next decisions
are the position of the first collision and
the nature of that collision. If it is deter-
mined that a fission occurs, the number of
emerging neutrons must be decided upon,
and each of these neutrons is eventually
followed in the same fashion as the first.
If the collision is decreed to be a scatter-
ing, appropriate statistics are invoked to
determine the new momentum of the neu-

John von Neumann

tron. When the neutron crosses a material
boundary, the parameters and characteris-
tics of the new medium are taken into ac-
count. Thus, a genealogical history of an
individual neutron is developed. The pro-
cess is repeated for other neutrons until a

statistically valid picture is generated.

Random Numbers.  How are the vari-
ous decisions made? To start with, the

computer must have a source of uni-
formly distributed psuedo-random num-
bers. A much used algorithm for gener-
ating such numbers is the so-called von
Neumann 窶徇iddle-square digits.窶 Here,
an arbitrary n-digit integer is squared,

creating a 2n-digit product. A new in-
teger is formed by extracting the middle
n-digits from the product. This process
is iterated over and over, forming a chain

127

background image

Monte Carlo

I

I

1

example, see the section entitled 窶弋he
Monte Carlo Method窶 in 窶廣 Primer on
Probability, Measure, and the Laws of
Large Numbers.窶)

Since its inception,

many international conferences have been
held on the various applications of the
method.

Recently, these range from

the conference, 窶廴onte Carlo Methods
and Applications in Neutronics, Photon-
ics, and Statistical Physics,窶 at Cadarache
Castle, France, in the spring of 1985 to

the latest at Los Alamos, 窶廡rontiers of
Quantum Monte Carlo,窶 in September,

1985.

Putting the Method into Practice

Let me return to the historical account.

In late 1947 the ENIAC was to be moved
to its permanent home at the Ballistics
Research Laboratory in Maryland. What
a gargantuan task! Few observers were
of the opinion that it would ever do an-

other multiplication or even an addition.
It is a tribute to the patience and skill
of Josh Gray and Richard Merwin, two

fearless uninitiated, that the move was a
success. One salutary effect of the inter-
ruption for Monte Carlo was that another
distinguished physicist took this occasion
to resume his interest in statistical studies.

Enrico Fermi helped create modern

physics.

Here, we focus on his inter-

est in neutron diffusion during those ex-
citing times in Rome in the early thir-
ties. According to Emilio Segre, Fermi窶冱
student and collaborator, 窶廡ermi had in-
vented, but of course not named, the
present Monte Carlo method when he was
studying the moderation of neutrons in
Rome. He did not publish anything on

the subject, but he used the method to

solve many problems with whatever cal-
culating facilities he had, chiefly a small

mechanical adding machine.窶*

In a recent conversation with Segre, I

Company from From X-Rays to Quarks by Emilio
Segre.

128

learned that Fermi took great delight in
astonishing his Roman colleagues with
his remarkably accurate, 窶徼oo-good-to-be-
lieve窶 predictions of experimental results.
After indulging himself, he revealed that
his 窶徃uesses窶 were really derived from
the statistical sampling techniques that he
used to calculate with whenever insomnia
struck in the wee morning hours! And
so it was that nearly fifteen years earlier,
Fermi had independently developed the
Monte Carlo method.

Enrico Fermi

It was then natural for Fermi, during

the hiatus in the ENIAC operation, to
dream up a simple but ingenious  ana-

log  device to implement studies in neu-

tron transport.

He persuaded his friend

and collaborator Percy King, while on a
hike one Sunday morning in the moun-
tains surrounding Los Alamos, to build
such an instrument窶罵ater affectionately
called the FERMIAC (see the accompa-
nying photo).

The FERMIAC developed neutron ge-

nealogies in two dimensions, that is, in a
plane, by generating the site of the 窶從ext
collision. 窶

Each generation was based

on a choice of parameters that charac-
terized the particular material being tra-

versed. When a material boundary was
crossed, another choice was made appro-
priate to the new material. The device
could accommodate two neutron energies,
referred to as 窶徭low窶 and 窶彷ast.窶 Once
again, the Master had just the right feel
for what was meaningful and relevant to
do in the pursuit of science.

The First Ambitious Test.  Much to
the amazement of many 窶彳xperts,窶 the
ENIAC survived the vicissitudes of its
200-mile journey. In the meantime Rich-
ard Clippinger, a staff member at Ab-
erdeen, had suggested that the ENIAC
had sufficient flexibility to permit its con-
trols to be reorganized into a more conve-
nient (albeit static) stored-program mode
of operation.

This mode would have a

capacity of 1800 instructions from a vo-
cabulary of about 60 arithmetical and log-
ical operations. The previous method of
programming might be likened to a gi-
ant plugboard, that is to say, to a can
of worms. Although implementing the
new approach is an interesting story, suf-
fice it to say that Johnny窶冱 wife, Klari,
and I designed the new controls in about
two months and completed the implemen-

tation in a fortnight. We then had the
opportunity of using the ENIAC for the
first ambitious test of the Monte Carlo
method窶蚤 variety of problems in neu-
tron transport done in collaboration with
Johnny.

Nine problems were computed corre-

sponding to various configurations of ma-

terials, initial distributions of neutrons,
and running times.

These problems, as

yet, did not include hydrodynamic or ra-
diative effects, but complex geometries
and realistic neutron-velocity spectra
were handled easily. The neutron histo-
ries were subjected to a variety of statisti-
cal analyses and comparisons with other
approaches.

Conclusions about the effi-

cacy of the method were quite favorable.
It seemed as though Monte Carlo was
here to stay.

Not long afterward, other Laboratory

background image

Monte Carlo

staff members made their pilgrimages to
ENIAC to run Monte Carlo problems.
These included J. Calkin, C. Evans, and
F. Evans, who studied a thermonuclear
problem using a cylindrical model as well
as the simpler spherical one. B. Suydam
and R. Stark tested the concept of artifi-
cial viscosity on time-dependent shocks;
they also, for the first time, tested and
found satisfactory an approach to hydro-
dynamics using a realistic equation of
state in spherical geometry. Also, the dis-
tinguished (and mysterious) mathemati-
cian C. J. Everett was taking an inter-
est in Monte Carlo that would culminate
in a series of outstanding publications in
collaboration with E. Cashwell. Mean-
while, Richtmyer was very actively run-
ning Monte Carlo problems on the so-
called SSEC during its brief existence at
IBM in New York.

In many ways, as one looks back, it

was among the best of times.

Rapid Growth. 

Applications discussed

in the literature were many and varied
and spread quickly. By midyear 1949 a

symposium on the Monte Carlo method,
sponsored by the Rand Corporation, the
National Bureau of Standards, and the
Oak Ridge Laboratory, was held in Los
Angeles. Later, a second symposium was
organized by members of the Statistical
Laboratory at the University of Florida in
Gainesville.

In early 1952a new computer, the MA-

NIAC, became operational at Los Ala-
mos. Soon after Anthony Turkevich led
a study of the nuclear cascades that result
when an accelerated particle collides with
a nucleus. The incoming particle strikes
a nucleon, experiencing either an elastic
or an inelastic scattering, with the latter
event producing a pion. In this study par-
ticles and their subsequent collisions were
followed until all particles either escaped
from the nucleus or their energy dropped
below some threshold value. The 窶彳xper-
iment窶 was repeated until sufficient statis-
tics were accumulated. A whole series of
target nuclei and incoming particle ener-
gies was examined.

Another computational problem run on

the MANIAC was a study of equations

THE FERMIAC

The Monte Carlo trolley, or FERMIAC, was

invented by Enrico Fermi and constructed

by Percy King. The drums on the trolley

were set according to the material being tra-

versed and a random choice between fast

and slow neutrons. Another random digit

was used to determine the direction of mo-

tion, and a third was selected to give the dis-

tance to the next collision. The trolley was

then operated by moving it across a two-

dimensional scale drawing of the nuclear

device or reactor assembly being studied.

The trolley drew a path as it rolled, stopping

for changes in drum settings whenever a

material boundary was crossed. This infant

computer was used for about two years to

determine, among other things, the change

in neutron population with time in numerous

types of nuclear systems.

of state based on the two-dimensional
motion of hard spheres. The work was
a collaborative effort with the Tellers,
Edward and Mici, and the Rosenbluths,
Marshall and Arianna (see 窶廴onte Carlo
at Work窶). During this study a strategy
was developed that led to greater com-
puting efficiency for equilibrium systems
obeying the Boltzmann distribution func-
tion. According to this strategy, if a sta-
tistical 窶徇ove窶 of a particle in the sys-
tem resulted in a decrease in the energy
of the system, the new configuration was
accepted. On the other hand, if there was
an increase in energy, the new configu-
ration was accepted only if it survived a
game of chance biased by a Boltzmann
factor. Otherwise, the old configuration
became a new statistic.

It is interesting to look back over two-

score years and note the emergence, rather

early on, of experimental mathematics,
a natural consequence of the electronic
computer.

The role of the Monte Carlo

method in reinforcing such mathematics
seems self-evident. When display units
were introduced, the temptation to exper-

129

background image

Monte Carlo

iment became almost irresistible, at least
for the fortunate few who enjoyed the lux-
ury of a hands-on policy. When shared-
time operations became realistic, exper-
imental mathematics came of age. At
long last, mathematics achieved a certain
parity-the twofold aspect of experiment
and theory-that all other sciences enjoy.

It is, in fact, the coupling of the sub-

tleties of the human brain with rapid
and reliable calculations, both arithmeti-
cal and logical, by the modern computer
that has stimulated the development of
experimental mathematics. This develop-

ment will enable us to achieve Olympian
heights.

The Future

So 

far I have summarized the rebirth

of statistical sampling under the rubric
of Monte Carlo. What of the future窶
perhaps even a not too distant future?

The miracle of the chip, like most mir-

acles, is almost unbelievable. Yet the fan-

tastic performances achieved to date have
not quieted all users. At the same time we

are reaching upper limits on the comput-
ing power of a single processor.

One bright facet of the miracle is the

lack of macroscopic moving parts, which
makes the chip a very reliable bit of
hardware. Such reliability suggests par-
allel processing.

The thought here is

not a simple extension to two, or even
four or eight, processing systems. Such
extensions are adiabatic transitions that,
to be sure, should be part of the im-
mediate, short-term game plan. Rather,
the thought is massively parallel opera-
tions with thousands of interacting pro-
cessors-even millions!

Already commercially available is one

computer, the Connection Machine, with
65,536 simple processors working in par-
allel. The processors are linked in such
a way that no processor in the array is
more than twelve wires away from an-
other and the processors are pairwise con-
nected by a number of equally efficient

routes, making communication both flex-
ible and efficient. The computer has been
used on such problems as turbulent fluid
flow, imaging processing (with features
analogous to the human visual system),
document retrieval, and 窶彡ommon-sense窶
reasoning in artificial intelligence.

One natural application of massive par-

allelism would be to the more ambitious
Monte Carlo problems already upon us.
To achieve good statistics in Monte Carlo
calculations, a large number of 窶徂istories窶
need to be followed. Although each his-
tory has its own unique path, the under-
lying calculations for all paths are highly
parallel in nature.

Still, the magnitude of the endeavor

to compute on massively parallel devices
must not be underestimated. Some of the
tools and techniques needed are:

A high-level language and new archi-
tecture able to deal with the demands
of such a sophisticated language (to the
relief of the user);
Highly efficient operating systems and
compilers;
Use of modern combinatorial theory,
perhaps even new principles of logic,
in the development of elegant, compre-
hensive architectures;
A fresh look at numerical analysis and
the preparation of new algorithms (we
have been mesmerized by serial com-
putation and purblind to the sophistica-
tion and artistry of parallelism).
Where will all this lead? If one were

to wax enthusiastic, perhaps窶破ust per-
haps窶蚤 simplified model of the brain
might be studied. These studies, in turn,
might provide feedback to computer ar-
chitects designing the new parallel struc-
tures.

Such matters fascinated Stan Ulam. He

often mused about the nature of memory
and how it was implemented in the brain.
Most important, though, his own brain
possessed the fertile imagination needed
to make substantive contributions to the
very important pursuit of understanding
intelligence. 

Further Reading

S. Ulam, R. D. Richtmyer, and J. von Neumann.

1947. Statistical methods in neutron diffusion. Los

Alamos Scientific Laboratory report LAMS窶551.
This reference contains the von Neumann letter dis-
cussed in the present article.

N. Metropolis and S. Ulam. 1949. The Monte
Carlo method.  Journal of the American Statistical
Association 44:335-341.

S. Ulam. 1950. Random processes and transforma-
tions. Proceedings of the International Congress of
Mathematicians 2:264-275.

Los Alamos Scientific Laboratory. 1966. Fermi in-
vention rediscovered at LASL.  The Atom, October,
pp. 7-11.

C. C. Hurd. 1985. A note on early Monte Carlo
computations and scientific meetings. Annals of the
History of Computing 7:141窶155.

W. Daniel Hillis. 1987. The connection machine.
Scientific American,  June, pp. 108窶1 15.

N. Metropolis 

received his B.S. (1937) and his

Ph.D. ( 1941) in physics at the University of Chi-
cago. He arrived in Los Alamos, April 1943, as
a member of the original staff of fifty scientists.
After the war he returned to the faculty of the
University of Chicago as Assistant Professor. He

came back to Los Alamos in 1948 to form the
group that designed and built MANIAC I and II. (He
chose the name MANIAC in the hope of stopping
the rash of such acronyms for machine names, but
may have, instead, only further stimulated such use.)
From 1957 to 1965 he was Professor of Physics
at the University of Chicago and was the founding
Director of its Institute for Computer Research. In

1965 he returned to Los Alamos where he was made

a Laboratory Senior Fellow in 1980. Although he
retired recently, he remains active as a Laboratory
Senior Fellow Emeritus.

130


Document Outline