Jump to content

Monte Carlo method

From Wikipedia, the free encyclopedia
(Redirected fromMonte carlo method)

The approximation of anormal distributionwith a Monte Carlo method

Monte Carlo methods,orMonte Carlo experiments,are a broad class ofcomputationalalgorithmsthat rely onrepeatedrandom samplingto obtain numerical results. The underlying concept is to userandomnessto solve problems that might bedeterministicin principle. The name comes from theMonte Carlo Casinoin Monaco, where the primary developer of the method, physicistStanislaw Ulam,was inspired by his uncle's gambling habits.

Monte Carlo methods are mainly used in three distinct problem classes: optimization, numerical integration, and generating draws from a probability distribution. They can also be used to model phenomena with significant uncertainty in inputs, such as calculating the risk of a nuclear power plant failure. Monte Carlo methods are often implemented using computer simulations, and they can provide approximate solutions to problems that are otherwise intractable or too complex to analyze mathematically.

Monte Carlo methods are widely used in various fields of science, engineering, and mathematics, such as physics, chemistry, biology, statistics, artificial intelligence, finance, and cryptography. They have also been applied to social sciences, such as sociology, psychology, and political science. Monte Carlo methods have been recognized as one of the most important and influential ideas of the 20th century, and they have enabled many scientific and technological breakthroughs.

Monte Carlo methods also have some limitations and challenges, such as the trade-off between accuracy and computational cost, the curse of dimensionality, the reliability of random number generators, and the verification and validation of the results.

Overview

[edit]

Monte Carlo methods vary, but tend to follow a particular pattern:

  1. Define a domain of possible inputs
  2. Generate inputs randomly from aprobability distributionover the domain
  3. Perform adeterministiccomputation of the outputs
  4. Aggregate the results
Monte Carlo method applied to approximating the value ofπ

For example, consider aquadrant (circular sector)inscribed in aunit square.Given that the ratio of their areas isπ/4,the value ofπcan be approximated using a Monte Carlo method:[1]

  1. Draw a square, theninscribea quadrant within it
  2. Uniformlyscatter a given number of points over the square
  3. Count the number of points inside the quadrant, i.e. having a distance from the origin of less than 1
  4. The ratio of the inside-count and the total-sample-count is an estimate of the ratio of the two areas,π/4.Multiply the result by 4 to estimateπ.

In this procedure the domain of inputs is the square that circumscribes the quadrant. One can generate random inputs by scattering grains over the square then perform a computation on each input (test whether it falls within the quadrant). Aggregating the results yields our final result, the approximation ofπ.

There are two important considerations:

  1. If the points are not uniformly distributed, then the approximation will be poor.
  2. The approximation is generally poor if only a few points are randomly placed in the whole square. On average, the approximation improves as more points are placed.

Uses of Monte Carlo methods require large amounts of random numbers, and their use benefitted greatly frompseudorandom number generators,which are far quicker to use than the tables of random numbers that had been previously used for statistical sampling.

Application

[edit]

Monte Carlo methods are often used inphysicalandmathematicalproblems and are most useful when it is difficult or impossible to use other approaches. Monte Carlo methods are mainly used in three problem classes:[2]optimization,numerical integration,and generating draws from aprobability distribution.

In physics-related problems, Monte Carlo methods are useful for simulating systems with manycoupleddegrees of freedom,such as fluids, disordered materials, strongly coupled solids, and cellular structures (seecellular Potts model,interacting particle systems,McKean–Vlasov processes,kinetic models of gases).

Other examples include modeling phenomena with significantuncertaintyin inputs such as the calculation ofriskin business and, in mathematics, evaluation of multidimensionaldefinite integralswith complicatedboundary conditions.In application to systems engineering problems (space,oil exploration,aircraft design, etc.), Monte Carlo–based predictions of failure,cost overrunsand schedule overruns are routinely better than human intuition or alternative "soft" methods.[3]

In principle, Monte Carlo methods can be used to solve any problem having a probabilistic interpretation. By thelaw of large numbers,integrals described by theexpected valueof some random variable can be approximated by taking theempirical mean(a.k.a.the 'sample mean') of independent samples of the variable. When theprobability distributionof the variable is parameterized, mathematicians often use aMarkov chain Monte Carlo(MCMC) sampler.[4][5][6]The central idea is to design a judiciousMarkov chainmodel with a prescribedstationary probability distribution.That is, in the limit, the samples being generated by the MCMC method will be samples from the desired (target) distribution.[7][8]By theergodic theorem,the stationary distribution is approximated by theempirical measuresof the random states of the MCMC sampler.

In other problems, the objective is generating draws from a sequence of probability distributions satisfying a nonlinear evolution equation. These flows of probability distributions can always be interpreted as the distributions of the random states of aMarkov processwhose transition probabilities depend on the distributions of the current random states (seeMcKean–Vlasov processes,nonlinear filtering equation).[9][10]In other instances, a flow of probability distributions with an increasing level of sampling complexity arise (path spaces models with an increasing time horizon, Boltzmann–Gibbs measures associated with decreasing temperature parameters, and many others). These models can also be seen as the evolution of the law of the random states of a nonlinear Markov chain.[10][11]A natural way to simulate these sophisticated nonlinear Markov processes is to sample multiple copies of the process, replacing in the evolution equation the unknown distributions of the random states by the sampledempirical measures.In contrast with traditional Monte Carlo and MCMC methodologies, thesemean-field particletechniques rely on sequential interacting samples. The terminologymean fieldreflects the fact that each of thesamples(a.k.a.particles, individuals, walkers, agents, creatures, or phenotypes) interacts with the empirical measures of the process. When the size of the system tends to infinity, these random empirical measures converge to the deterministic distribution of the random states of the nonlinear Markov chain, so that the statistical interaction between particles vanishes.

Simple Monte Carlo

[edit]

Suppose one wants to know theexpected valueμof apopulation(and knows thatμexists), but does not have a formula available to compute it. The simple Monte Carlo method gives an estimate forμby runningnsimulations and averaging the simulations’ results. It has no restrictions on the probability distribution of the inputs to the simulations, requiring only that the inpus are randomly generated and are independent of each other and thatμexists. A sufficiently largenwill produce a value formthat is arbitrarily close toμ;more formally, it will be the case that, for anyε> 0, |μm| ≤ε.[12]

Typically, the algorithm to obtainmis

s= 0;
fori= 1 tondo
run the simulation for theithtime, giving resultri;
s=s+ri;
repeat
m=s/n;

An example

[edit]

Suppose we want to know how many times we should expect to throw three eight-sided dice for the total of the dice throws to be at leastT.We know the expected value exists. The dice throws are randomly distributed and independent of each other. So simple Monte Carlo is applicable:

s= 0;
fori= 1 tondo
throw the three dice untilTis met or first exceeded;ri= the number of throws;
s=s+ri;
repeat
m=s/n;

Ifnis large enough,mwill be withinεofμfor anyε> 0.

Determining a sufficiently largen

[edit]

General formula

[edit]

Letε= |μm| > 0. Choose the desiredconfidence level– the percent chance that, when the Monte Carlo algorithm completes,mis indeed withinεofμ.Let z be thez-scorecorresponding to that confidence level.

Lets2be the estimated variance, sometimes called the “sample” variance; it is the variance of the results obtained from a relatively small numberkof “sample” simulations. Choose ak;Driels and Shin observe that “even for sample sizes an order of magnitude lower than the number required, the calculation of that number is quite stable."[13]

The following algorithm computess2in one pass while minimizing the possibility that accumulated numerical error produces erroneous results:[12]

s1= 0;
run the simulation for the first time, producing resultr1;
m1=r1;//miis the mean of the firstisimulations
fori = 2 tokdo
run the simulation for theithtime, producing resultri;
δi=ri-mi-1;
mi= mi-1+ (1/i)δi;
si= si-1+ ((i- 1)/i)(δi)2;
repeat
s2=sk/(k- 1);

Note that, when the algorithm completes,mkis the mean of thekresults.

nis sufficiently large when

[12][13]

Ifnk,thenmk= m;sufficient sample simulations were done to ensure thatmkis withinεofμ.Ifn>k,thennsimulations can be run “from scratch,” or, since k simulations have already been done, one can just runnkmore simulations and add their results into those from the sample simulations:

s=mk*k;
for i =k+ 1 tondo
run the simulation for theithtime, giving resultri;
s=s+ri;
m=s/n;

A formula when simulations' results are bounded

[edit]

An alternate formula can be used in the special case where all simulation results are bounded above and below.

Choose a value forεthat is twice the maximum allowed difference betweenμandm.Let 0 <δ< 100 be the desired confidence level, expressed as a percentage. Let every simulation resultr1,r2,…ri,…rnbe such thataribfor finiteaandb.To have confidence of at leastδthat |μm| <ε/2, use a value fornsuch that

For example, ifδ= 99%, thenn≥ 2(ba)2ln(2/0.01)/ε2≈ 10.6(ba)2/ε2.[12]

Computational costs

[edit]

Despite its conceptual and algorithmic simplicity, the computational cost associated with a Monte Carlo simulation can be staggeringly high. In general the method requires many samples to get a good approximation, which may incur an arbitrarily large total runtime if the processing time of a single sample is high.[14]Although this is a severe limitation in very complex problems, theembarrassingly parallelnature of the algorithm allows this large cost to be reduced (perhaps to a feasible level) throughparallel computingstrategies in local processors, clusters, cloud computing, GPU, FPGA, etc.[15][16][17][18]

History

[edit]

Before the Monte Carlo method was developed, simulations tested a previously understood deterministic problem, and statistical sampling was used to estimate uncertainties in the simulations. Monte Carlo simulations invert this approach, solving deterministic problems usingprobabilisticmetaheuristics(seesimulated annealing).

An early variant of the Monte Carlo method was devised to solve theBuffon's needle problem,in whichπcan be estimated by dropping needles on a floor made of parallel equidistant strips. In the 1930s,Enrico Fermifirst experimented with the Monte Carlo method while studying neutron diffusion, but he did not publish this work.[19]

In the late 1940s,Stanislaw Ulaminvented the modern version of the Markov Chain Monte Carlo method while he was working on nuclear weapons projects at theLos Alamos National Laboratory.In 1946, nuclear weapons physicists at Los Alamos were investigating neutron diffusion in the core of a nuclear weapon.[19]Despite having most of the necessary data, such as the average distance a neutron would travel in a substance before it collided with an atomic nucleus and how much energy the neutron was likely to give off following a collision, the Los Alamos physicists were unable to solve the problem using conventional, deterministic mathematical methods. Ulam proposed using random experiments. He recounts his inspiration as follows:

The first thoughts and attempts I made to practice [the Monte Carlo Method] were suggested by a question which occurred to me in 1946 as I was convalescing from an illness and playing solitaires. The question was what are the chances that a Canfield solitaire laid out with 52 cards will come out successfully? After spending a lot of time trying to estimate them by pure combinatorial calculations, I wondered whether a more practical method than "abstract thinking" might not be to lay it out say one hundred times and simply observe and count the number of successful plays. This was already possible to envisage with the beginning of the new era of fast computers, and I immediately thought of problems of neutron diffusion and other questions of mathematical physics, and more generally how to change processes described by certain differential equations into an equivalent form interpretable as a succession of random operations. Later [in 1946], I described the idea toJohn von Neumann,and we began to plan actual calculations.[20]

Being secret, the work of von Neumann and Ulam required a code name.[21]A colleague of von Neumann and Ulam,Nicholas Metropolis,suggested using the nameMonte Carlo,which refers to theMonte Carlo CasinoinMonacowhere Ulam's uncle would borrow money from relatives to gamble.[19] Monte Carlo methods were central to thesimulationsrequired for further postwar development of nuclear weapons, including the design of the H-bomb, though severely limited by the computational tools at the time. Von Neumann,Nicholas Metropolisand others programmed theENIACcomputer to perform the first fully automated Monte Carlo calculations, of afission weaponcore, in the spring of 1948.[22]In the 1950s Monte Carlo methods were used atLos Alamosfor the development of thehydrogen bomb,and became popularized in the fields ofphysics,physical chemistry,andoperations research.TheRand Corporationand theU.S. Air Forcewere two of the major organizations responsible for funding and disseminating information on Monte Carlo methods during this time, and they began to find a wide application in many different fields.

The theory of more sophisticated mean-field type particle Monte Carlo methods had certainly started by the mid-1960s, with the work ofHenry P. McKean Jr.on Markov interpretations of a class of nonlinear parabolic partial differential equations arising in fluid mechanics.[23][24]An earlier pioneering article byTheodore E. Harrisand Herman Kahn, published in 1951, used mean-fieldgenetic-type Monte Carlo methods for estimating particle transmission energies.[25]Mean-field genetic type Monte Carlo methodologies are also used as heuristic natural search algorithms (a.k.a.metaheuristic) in evolutionary computing. The origins of these mean-field computational techniques can be traced to 1950 and 1954 with the work ofAlan Turingon genetic type mutation-selection learning machines[26]and the articles byNils Aall Barricelliat theInstitute for Advanced StudyinPrinceton, New Jersey.[27][28]

Quantum Monte Carlo,and more specificallydiffusion Monte Carlo methodscan also be interpreted as a mean-field particle Monte Carlo approximation ofFeynmanKacpath integrals.[29][30][31][32][33][34][35]The origins of Quantum Monte Carlo methods are often attributed to Enrico Fermi andRobert Richtmyerwho developed in 1948 a mean-field particle interpretation of neutron-chain reactions,[36]but the first heuristic-like and genetic type particle algorithm (a.k.a. Resampled or Reconfiguration Monte Carlo methods) for estimating ground state energies of quantum systems (in reduced matrix models) is due to Jack H. Hetherington in 1984.[35]In molecular chemistry, the use of genetic heuristic-like particle methodologies (a.k.a. pruning and enrichment strategies) can be traced back to 1955 with the seminal work ofMarshall N. RosenbluthandArianna W. Rosenbluth.[37]

The use ofSequential Monte Carloin advancedsignal processingandBayesian inferenceis more recent. It was in 1993, that Gordon et al., published in their seminal work[38]the first application of a Monte Carloresamplingalgorithm in Bayesian statistical inference. The authors named their algorithm 'the bootstrap filter', and demonstrated that compared to other filtering methods, their bootstrap algorithm does not require any assumption about that state-space or the noise of the system. Another pioneering article in this field was Genshiro Kitagawa's, on a related "Monte Carlo filter",[39]and the ones by Pierre Del Moral[40]and Himilcon Carvalho, Pierre Del Moral, André Monin and Gérard Salut[41]on particle filters published in the mid-1990s. Particle filters were also developed in signal processing in 1989–1992 by P. Del Moral, J. C. Noyer, G. Rigal, and G. Salut in the LAAS-CNRS in a series of restricted and classified research reports with STCAN (Service Technique des Constructions et Armes Navales), the IT company DIGILOG, and theLAAS-CNRS(the Laboratory for Analysis and Architecture of Systems) on radar/sonar and GPS signal processing problems.[42][43][44][45][46][47]These Sequential Monte Carlo methodologies can be interpreted as an acceptance-rejection sampler equipped with an interacting recycling mechanism.

From 1950 to 1996, all the publications on Sequential Monte Carlo methodologies, including the pruning and resample Monte Carlo methods introduced in computational physics and molecular chemistry, present natural and heuristic-like algorithms applied to different situations without a single proof of their consistency, nor a discussion on the bias of the estimates and on genealogical and ancestral tree based algorithms. The mathematical foundations and the first rigorous analysis of these particle algorithms were written by Pierre Del Moral in 1996.[40][48]

Branching type particle methodologies with varying population sizes were also developed in the end of the 1990s by Dan Crisan, Jessica Gaines and Terry Lyons,[49][50][51]and by Dan Crisan, Pierre Del Moral and Terry Lyons.[52]Further developments in this field were described in 1999 to 2001 by P. Del Moral, A. Guionnet and L. Miclo.[30][53][54]

Definitions

[edit]

There is no consensus on howMonte Carloshould be defined. For example, Ripley[55]defines most probabilistic modeling asstochastic simulation,withMonte Carlobeing reserved forMonte Carlo integrationand Monte Carlo statistical tests.Sawilowsky[56]distinguishes between asimulation,a Monte Carlo method, and a Monte Carlo simulation: a simulation is a fictitious representation of reality, a Monte Carlo method is a technique that can be used to solve a mathematical or statistical problem, and a Monte Carlo simulation uses repeated sampling to obtain the statistical properties of some phenomenon (or behavior).

Here are the examples:

  • Simulation: Drawingonepseudo-random uniform variable from the interval [0,1] can be used to simulate the tossing of a coin: If the value is less than or equal to 0.50 designate the outcome as heads, but if the value is greater than 0.50 designate the outcome as tails. This is a simulation, but not a Monte Carlo simulation.
  • Monte Carlo method: Pouring out a box of coins on a table, and then computing the ratio of coins that land heads versus tails is a Monte Carlo method of determining the behavior of repeated coin tosses, but it is not a simulation.
  • Monte Carlo simulation: Drawinga large numberof pseudo-random uniform variables from the interval [0,1] at one time, or once at many different times, and assigning values less than or equal to 0.50 as heads and greater than 0.50 as tails, is aMonte Carlo simulationof the behavior of repeatedly tossing a coin.

Kalos and Whitlock[57]point out that such distinctions are not always easy to maintain. For example, the emission of radiation from atoms is a natural stochastic process. It can be simulated directly, or its average behavior can be described by stochastic equations that can themselves be solved using Monte Carlo methods. "Indeed, the same computer code can be viewed simultaneously as a 'natural simulation' or as a solution of the equations by natural sampling."

Convergence of the Monte Carlo simulation can be checked with theGelman-Rubin statistic.

Monte Carlo and random numbers

[edit]

The main idea behind this method is that the results are computed based on repeated random sampling and statistical analysis. The Monte Carlo simulation is, in fact, random experimentations, in the case that, the results of these experiments are not well known. Monte Carlo simulations are typically characterized by many unknown parameters, many of which are difficult to obtain experimentally.[58]Monte Carlo simulation methods do not always requiretruly random numbersto be useful (although, for some applications such asprimality testing,unpredictability is vital).[59]Many of the most useful techniques use deterministic,pseudorandomsequences, making it easy to test and re-run simulations. The only quality usually necessary to make goodsimulationsis for the pseudo-random sequence to appear "random enough" in a certain sense.

What this means depends on the application, but typically they should pass a series of statistical tests. Testing that the numbers areuniformly distributedor follow another desired distribution when a large enough number of elements of the sequence are considered is one of the simplest and most common ones. Weak correlations between successive samples are also often desirable/necessary.

Sawilowsky lists the characteristics of a high-quality Monte Carlo simulation:[56]

  • the (pseudo-random) number generator has certain characteristics (e.g. a long "period" before the sequence repeats)
  • the (pseudo-random) number generator produces values that pass tests for randomness
  • there are enough samples to ensure accurate results
  • the proper sampling technique is used
  • the algorithm used is valid for what is being modeled
  • it simulates the phenomenon in question.

Pseudo-random number samplingalgorithms are used to transform uniformly distributed pseudo-random numbers into numbers that are distributed according to a givenprobability distribution.

Low-discrepancy sequencesare often used instead of random sampling from a space as they ensure even coverage and normally have a faster order of convergence than Monte Carlo simulations using random or pseudorandom sequences. Methods based on their use are calledquasi-Monte Carlo methods.

In an effort to assess the impact of random number quality on Monte Carlo simulation outcomes, astrophysical researchers tested cryptographically secure pseudorandom numbers generated via Intel'sRDRANDinstruction set, as compared to those derived from algorithms, like theMersenne Twister,in Monte Carlo simulations of radio flares frombrown dwarfs.No statistically significant difference was found between models generated with typical pseudorandom number generators and RDRAND for trials consisting of the generation of 107random numbers.[60]

Monte Carlo simulation versus "what if" scenarios

[edit]

There are ways of using probabilities that are definitely not Monte Carlo simulations – for example, deterministic modeling using single-point estimates. Each uncertain variable within a model is assigned a "best guess" estimate. Scenarios (such as best, worst, or most likely case) for each input variable are chosen and the results recorded.[61]

By contrast, Monte Carlo simulations sample from aprobability distributionfor each variable to produce hundreds or thousands of possible outcomes. The results are analyzed to get probabilities of different outcomes occurring.[62]For example, a comparison of a spreadsheet cost construction model run using traditional "what if" scenarios, and then running the comparison again with Monte Carlo simulation andtriangular probability distributionsshows that the Monte Carlo analysis has a narrower range than the "what if" analysis.[example needed]This is because the "what if" analysis gives equal weight to all scenarios (seequantifying uncertainty in corporate finance), while the Monte Carlo method hardly samples in the very low probability regions. The samples in such regions are called "rare events".

Applications

[edit]

Monte Carlo methods are especially useful for simulating phenomena with significantuncertaintyin inputs and systems with manycoupleddegrees of freedom. Areas of application include:

Physical sciences

[edit]

Monte Carlo methods are very important incomputational physics,physical chemistry,and related applied fields, and have diverse applications from complicatedquantum chromodynamicscalculations to designingheat shieldsandaerodynamicforms as well as in modeling radiation transport for radiation dosimetry calculations.[63][64][65]Instatistical physics,Monte Carlo molecular modelingis an alternative to computationalmolecular dynamics,and Monte Carlo methods are used to computestatistical field theoriesof simple particle and polymer systems.[37][66]Quantum Monte Carlomethods solve themany-body problemfor quantum systems.[9][10][29]Inradiation materials science,thebinary collision approximationfor simulatingion implantationis usually based on a Monte Carlo approach to select the next colliding atom.[67]In experimentalparticle physics,Monte Carlo methods are used for designingdetectors,understanding their behavior and comparing experimental data to theory. Inastrophysics,they are used in such diverse manners as to model bothgalaxyevolution[68]and microwave radiation transmission through a rough planetary surface.[69]Monte Carlo methods are also used in theensemble modelsthat form the basis of modernweather forecasting.

Engineering

[edit]

Monte Carlo methods are widely used in engineering forsensitivity analysisand quantitativeprobabilisticanalysis inprocess design.The need arises from the interactive, co-linear and non-linear behavior of typical process simulations. For example,

Climate change and radiative forcing

[edit]

TheIntergovernmental Panel on Climate Changerelies on Monte Carlo methods inprobability density functionanalysis ofradiative forcing.[73]

Computational biology

[edit]

Monte Carlo methods are used in various fields ofcomputational biology,for example forBayesian inference in phylogeny,or for studying biological systems such as genomes, proteins,[74]or membranes.[75] The systems can be studied in the coarse-grained orab initioframeworks depending on the desired accuracy. Computer simulations allow monitoring of the local environment of a particularmoleculeto see if somechemical reactionis happening for instance. In cases where it is not feasible to conduct a physical experiment,thought experimentscan be conducted (for instance: breaking bonds, introducing impurities at specific sites, changing the local/global structure, or introducing external fields).

Computer graphics

[edit]

Path tracing,occasionally referred to as Monte Carlo ray tracing, renders a 3D scene by randomly tracing samples of possible light paths. Repeated sampling of any given pixel will eventually cause the average of the samples to converge on the correct solution of therendering equation,making it one of the most physically accurate 3D graphics rendering methods in existence.

Applied statistics

[edit]

The standards for Monte Carlo experiments in statistics were set by Sawilowsky.[76]In applied statistics, Monte Carlo methods may be used for at least four purposes:

  1. To compare competing statistics for small samples under realistic data conditions. Althoughtype I errorand power properties of statistics can be calculated for data drawn from classical theoretical distributions (e.g.,normal curve,Cauchy distribution) forasymptoticconditions (i. e,infinite sample size and infinitesimally small treatment effect), real data often do not have such distributions.[77]
  2. To provide implementations ofhypothesis teststhat are more efficient than exact tests such aspermutation tests(which are often impossible to compute) while being more accurate than critical values forasymptotic distributions.
  3. To provide a random sample from the posterior distribution inBayesian inference.This sample then approximates and summarizes all the essential features of the posterior.
  4. To provide efficient random estimates of the Hessian matrix of the negative log-likelihood function that may be averaged to form an estimate of theFisher informationmatrix.[78][79]

Monte Carlo methods are also a compromise between approximate randomization and permutation tests. An approximaterandomization testis based on a specified subset of all permutations (which entails potentially enormous housekeeping of which permutations have been considered). The Monte Carlo approach is based on a specified number of randomly drawn permutations (exchanging a minor loss in precision if a permutation is drawn twice—or more frequently—for the efficiency of not having to track which permutations have already been selected).

Artificial intelligence for games

[edit]

Monte Carlo methods have been developed into a technique calledMonte-Carlo tree searchthat is useful for searching for the best move in a game. Possible moves are organized in asearch treeand many random simulations are used to estimate the long-term potential of each move. A black box simulator represents the opponent's moves.[80]

The Monte Carlo tree search (MCTS) method has four steps:[81]

  1. Starting at root node of the tree, select optimal child nodes until a leaf node is reached.
  2. Expand the leaf node and choose one of its children.
  3. Play a simulated game starting with that node.
  4. Use the results of that simulated game to update the node and its ancestors.

The net effect, over the course of many simulated games, is that the value of a node representing a move will go up or down, hopefully corresponding to whether or not that node represents a good move.

Monte Carlo Tree Search has been used successfully to play games such asGo,[82]Tantrix,[83]Battleship,[84]Havannah,[85]andArimaa.[86]

Design and visuals

[edit]

Monte Carlo methods are also efficient in solving coupled integral differential equations of radiation fields and energy transport, and thus these methods have been used inglobal illuminationcomputations that produce photo-realistic images of virtual 3D models, with applications invideo games,architecture,design,computer generatedfilms,and cinematic special effects.[87]

Search and rescue

[edit]

TheUS Coast Guardutilizes Monte Carlo methods within its computer modeling softwareSAROPSin order to calculate the probable locations of vessels duringsearch and rescueoperations. Each simulation can generate as many as ten thousand data points that are randomly distributed based upon provided variables.[88]Search patterns are then generated based upon extrapolations of these data in order to optimize the probability of containment (POC) and the probability of detection (POD), which together will equal an overall probability of success (POS). Ultimately this serves as a practical application ofprobability distributionin order to provide the swiftest and most expedient method of rescue, saving both lives and resources.[89]

Finance and business

[edit]

Monte Carlo simulation is commonly used to evaluate the risk and uncertainty that would affect the outcome of different decision options. Monte Carlo simulation allows the business risk analyst to incorporate the total effects of uncertainty in variables like sales volume, commodity and labor prices, interest and exchange rates, as well as the effect of distinct risk events like the cancellation of a contract or the change of a tax law.

Monte Carlo methods in financeare often used toevaluate investments in projectsat a business unit or corporate level, or other financial valuations. They can be used to modelproject schedules,where simulations aggregate estimates for worst-case, best-case, and most likely durations for each task to determine outcomes for the overall project.[90]Monte Carlo methods are also used in option pricing, default risk analysis.[91][92]Additionally, they can be used to estimate the financial impact of medical interventions.[93]

Law

[edit]

A Monte Carlo approach was used for evaluating the potential value of a proposed program to help female petitioners in Wisconsin be successful in their applications forharassmentanddomestic abuse restraining orders.It was proposed to help women succeed in their petitions by providing them with greater advocacy thereby potentially reducing the risk ofrapeandphysical assault.However, there were many variables in play that could not be estimated perfectly, including the effectiveness of restraining orders, the success rate of petitioners both with and without advocacy, and many others. The study ran trials that varied these variables to come up with an overall estimate of the success level of the proposed program as a whole.[94]

Library science

[edit]

Monte Carlo approach had also been used to simulate the number of book publications based on bookgenrein Malaysia. The Monte Carlo simulation utilized previous published National Book publication data and book's price according to book genre in the local market. The Monte Carlo results were used to determine what kind of book genre that Malaysians are fond of and was used to compare book publications betweenMalaysiaandJapan.[95]

Other

[edit]

Nassim Nicholas Talebwrites about Monte Carlo generators in his 2001 bookFooled by Randomnessas a real instance of thereverse Turing test:a human can be declared unintelligent if their writing cannot be told apart from a generated one.

Use in mathematics

[edit]

In general, the Monte Carlo methods are used in mathematics to solve various problems by generating suitable random numbers (see alsoRandom number generation) and observing that fraction of the numbers that obeys some property or properties. The method is useful for obtaining numerical solutions to problems too complicated to solve analytically. The most common application of the Monte Carlo method is Monte Carlo integration.

Integration

[edit]
Monte-Carlo integration works by comparing random points with the value of the function.
Errors reduce by a factor of.

Deterministicnumerical integrationalgorithms work well in a small number of dimensions, but encounter two problems when the functions have many variables. First, the number of function evaluations needed increases rapidly with the number of dimensions. For example, if 10 evaluations provide adequate accuracy in one dimension, then10100points are needed for 100 dimensions—far too many to be computed. This is called thecurse of dimensionality.Second, the boundary of a multidimensional region may be very complicated, so it may not be feasible to reduce the problem to aniterated integral.[96]100dimensionsis by no means unusual, since in many physical problems, a "dimension" is equivalent to adegree of freedom.

Monte Carlo methods provide a way out of this exponential increase in computation time. As long as the function in question is reasonablywell-behaved,it can be estimated by randomly selecting points in 100-dimensional space, and taking some kind of average of the function values at these points. By thecentral limit theorem,this method displaysconvergence—i.e., quadrupling the number of sampled points halves the error, regardless of the number of dimensions.[96]

A refinement of this method, known asimportance samplingin statistics, involves sampling the points randomly, but more frequently where the integrand is large. To do this precisely one would have to already know the integral, but one can approximate the integral by an integral of a similar function or use adaptive routines such asstratified sampling,recursive stratified sampling,adaptive umbrella sampling[97][98]or theVEGAS algorithm.

A similar approach, thequasi-Monte Carlo method,useslow-discrepancy sequences.These sequences "fill" the area better and sample the most important points more frequently, so quasi-Monte Carlo methods can often converge on the integral more quickly.

Another class of methods for sampling points in a volume is to simulate random walks over it (Markov chain Monte Carlo). Such methods include theMetropolis–Hastings algorithm,Gibbs sampling,Wang and Landau algorithm,and interacting type MCMC methodologies such as thesequential Monte Carlosamplers.[99]

Simulation and optimization

[edit]

Another powerful and very popular application for random numbers in numerical simulation is innumerical optimization.The problem is to minimize (or maximize) functions of some vector that often has many dimensions. Many problems can be phrased in this way: for example, acomputer chessprogram could be seen as trying to find the set of, say, 10 moves that produces the best evaluation function at the end. In thetraveling salesman problemthe goal is to minimize distance traveled. There are also applications to engineering design, such asmultidisciplinary design optimization.It has been applied with quasi-one-dimensional models to solve particle dynamics problems by efficiently exploring large configuration space. Reference[100]is a comprehensive review of many issues related to simulation and optimization.

Thetraveling salesman problemis what is called a conventional optimization problem. That is, all the facts (distances between each destination point) needed to determine the optimal path to follow are known with certainty and the goal is to run through the possible travel choices to come up with the one with the lowest total distance. If instead of the goal being to minimize the total distance traveled to visit each desired destination but rather to minimize the total time needed to reach each destination, this goes beyond conventional optimization since travel time is inherently uncertain (traffic jams, time of day, etc.). As a result, to determine the optimal path a different simulation is required: optimization to first understand the range of potential times it could take to go from one point to another (represented by a probability distribution in this case rather than a specific distance) and then optimize the travel decisions to identify the best path to follow taking that uncertainty into account.

Inverse problems

[edit]

Probabilistic formulation ofinverse problemsleads to the definition of aprobability distributionin the model space. This probability distribution combinespriorinformation with new information obtained by measuring some observable parameters (data). As, in the general case, the theory linking data with model parameters is nonlinear, the posterior probability in the model space may not be easy to describe (it may be multimodal, some moments may not be defined, etc.).

When analyzing an inverse problem, obtaining a maximum likelihood model is usually not sufficient, as normally information on the resolution power of the data is desired. In the general case many parameters are modeled, and an inspection of themarginal probabilitydensities of interest may be impractical, or even useless. But it is possible to pseudorandomly generate a large collection of models according to theposterior probability distributionand to analyze and display the models in such a way that information on the relative likelihoods of model properties is conveyed to the spectator. This can be accomplished by means of an efficient Monte Carlo method, even in cases where no explicit formula for thea prioridistribution is available.

The best-known importance sampling method, the Metropolis algorithm, can be generalized, and this gives a method that allows analysis of (possibly highly nonlinear) inverse problems with complexa prioriinformation and data with an arbitrary noise distribution.[101][102]

Philosophy

[edit]

Popular exposition of the Monte Carlo Method was conducted by McCracken.[103]The method's general philosophy was discussed byElishakoff[104]and Grüne-Yanoff and Weirich.[105]

See also

[edit]

References

[edit]

Citations

[edit]
  1. ^Kalos & Whitlock 2008.
  2. ^Kroese, D. P.; Brereton, T.; Taimre, T.; Botev, Z. I. (2014). "Why the Monte Carlo method is so important today".WIREs Comput Stat.6(6): 386–392.doi:10.1002/wics.1314.S2CID18521840.
  3. ^Hubbard, Douglas; Samuelson, Douglas A. (October 2009)."Modeling Without Measurements".OR/MS Today:28–33.
  4. ^Metropolis, Nicholas; Rosenbluth, Arianna W.; Rosenbluth, Marshall N.; Teller, Augusta H.; Teller, Edward (June 1, 1953). "Equation of State Calculations by Fast Computing Machines".The Journal of Chemical Physics.21(6): 1087–1092.Bibcode:1953JChPh..21.1087M.doi:10.1063/1.1699114.ISSN0021-9606.OSTI4390578.S2CID1046577.
  5. ^Hastings, W. K. (April 1, 1970). "Monte Carlo sampling methods using Markov chains and their applications".Biometrika.57(1): 97–109.Bibcode:1970Bimka..57...97H.doi:10.1093/biomet/57.1.97.ISSN0006-3444.S2CID21204149.
  6. ^Liu, Jun S.; Liang, Faming; Wong, Wing Hung (March 1, 2000). "The Multiple-Try Method and Local Optimization in Metropolis Sampling".Journal of the American Statistical Association.95(449): 121–134.doi:10.1080/01621459.2000.10473908.ISSN0162-1459.S2CID123468109.
  7. ^Spall, J. C. (2003). "Estimation via Markov Chain Monte Carlo".IEEE Control Systems Magazine.23(2): 34–45.doi:10.1109/MCS.2003.1188770.
  8. ^Hill, Stacy D.; Spall, James C. (2019). "Stationarity and Convergence of the Metropolis-Hastings Algorithm: Insights into Theoretical Aspects".IEEE Control Systems Magazine.39:56–67.doi:10.1109/MCS.2018.2876959.S2CID58672766.
  9. ^abKolokoltsov, Vassili (2010).Nonlinear Markov processes.Cambridge University Press.p. 375.
  10. ^abcDel Moral, Pierre (2013).Mean field simulation for Monte Carlo integration.Chapman & Hall/CRC Press.p. 626.Monographs on Statistics & Applied Probability
  11. ^Del Moral, P.; Doucet, A.; Jasra, A. (2006). "Sequential Monte Carlo samplers".Journal of the Royal Statistical Society, Series B.68(3): 411–436.arXiv:cond-mat/0212648.doi:10.1111/j.1467-9868.2006.00553.x.S2CID12074789.
  12. ^abcdOwen, Art B. (2013).Monte Carlo Theory, Methods and Examples(PDF).Work in progress. pp. 15–36.{{cite book}}:CS1 maint: date and year (link)
  13. ^abDriels, Morris R.; Shin, Young S. (April 2004)."Determining the number of Iterations for Monte Carlo Simulations of Weapon Effectiveness".Naval Postgraduate School Technical Report(March 2003 - March 2004): 10–11.
  14. ^Shonkwiler, R. W.; Mendivil, F. (2009).Explorations in Monte Carlo Methods.Springer.
  15. ^Atanassova, E.; Gurov, T.; Karaivanova, A.; Ivanovska, S.; Durchova, M.; Dimitrov, D. (2016). "On the parallelization approaches for Intel MIC architecture".AIP Conference Proceedings.1773(1): 070001.Bibcode:2016AIPC.1773g0001A.doi:10.1063/1.4964983.
  16. ^Cunha Jr, A.; Nasser, R.; Sampaio, R.; Lopes, H.; Breitman, K. (2014). "Uncertainty quantification through the Monte Carlo method in a cloud computing setting".Computer Physics Communications.185(5): 1355–1363.arXiv:2105.09512.Bibcode:2014CoPhC.185.1355C.doi:10.1016/j.cpc.2014.01.006.S2CID32376269.
  17. ^Wei, J.; Kruis, F.E. (2013). "A GPU-based parallelized Monte-Carlo method for particle coagulation using an acceptance–rejection strategy".Chemical Engineering Science.104:451–459.Bibcode:2013ChEnS.104..451W.doi:10.1016/j.ces.2013.08.008.
  18. ^Lin, Y.; Wang, F.; Liu, B. (2018). "Random number generators for large-scale parallel Monte Carlo simulations on FPGA".Journal of Computational Physics.360:93–103.Bibcode:2018JCoPh.360...93L.doi:10.1016/j.jcp.2018.01.029.
  19. ^abcMetropolis 1987.
  20. ^Eckhardt 1987.
  21. ^abMazhdrakov, Benov & Valkanov 2018,p. 250.
  22. ^Haigh, Thomas; Priestley, Mark; Rope, Crispin (2014)."Los Alamos Bets on ENIAC: Nuclear Monte Carlo Simulations, 1947-1948".IEEE Annals of the History of Computing.36(3): 42–63.doi:10.1109/MAHC.2014.40.S2CID17470931.
  23. ^McKean, Henry P. (1967). "Propagation of chaos for a class of non-linear parabolic equations".Lecture Series in Differential Equations, Catholic Univ.7:41–57.
  24. ^McKean, Henry P. (1966)."A class of Markov processes associated with nonlinear parabolic equations".Proc. Natl. Acad. Sci. USA.56(6): 1907–1911.Bibcode:1966PNAS...56.1907M.doi:10.1073/pnas.56.6.1907.PMC220210.PMID16591437.
  25. ^Herman, Kahn; Theodore, Harris E. (1951)."Estimation of particle transmission by random sampling"(PDF).Natl. Bur. Stand. Appl. Math. Ser.12:27–30.
  26. ^Turing, Alan M. (1950). "Computing machinery and intelligence".Mind.LIX(238): 433–460.doi:10.1093/mind/LIX.236.433.
  27. ^Barricelli, Nils Aall(1954). "Esempi numerici di processi di evoluzione".Methodos:45–68.
  28. ^Barricelli, Nils Aall(1957). "Symbiogenetic evolution processes realized by artificial methods".Methodos:143–182.
  29. ^abDel Moral, Pierre (2004).Feynman–Kac formulae. Genealogical and interacting particle approximations.Probability and Its Applications. Springer. p. 575.ISBN9780387202686.Series: Probability and Applications
  30. ^abDel Moral, P.; Miclo, L. (2000)."Branching and interacting particle systems approximations of Feynman–Kac formulae with applications to non-linear filtering".Séminaire de Probabilités XXXIV.Lecture Notes in Mathematics. Vol. 1729. Berlin: Springer. pp. 1–145.doi:10.1007/BFb0103798.ISBN978-3-540-67314-9.MR1768060.
  31. ^Del Moral, Pierre; Miclo, Laurent (2000)."A Moran particle system approximation of Feynman–Kac formulae".Stochastic Processes and Their Applications.86(2): 193–216.doi:10.1016/S0304-4149(99)00094-0.
  32. ^Del Moral, Pierre (2003)."Particle approximations of Lyapunov exponents connected to Schrödinger operators and Feynman–Kac semigroups".ESAIM Probability & Statistics.7:171–208.doi:10.1051/ps:2003001.
  33. ^Assaraf, Roland; Caffarel, Michel; Khelif, Anatole (2000)."Diffusion Monte Carlo Methods with a fixed number of walkers"(PDF).Phys. Rev. E.61(4): 4566–4575.Bibcode:2000PhRvE..61.4566A.doi:10.1103/physreve.61.4566.PMID11088257.Archived fromthe original(PDF)on November 7, 2014.
  34. ^Caffarel, Michel; Ceperley, David; Kalos, Malvin (1993). "Comment on Feynman–Kac Path-Integral Calculation of the Ground-State Energies of Atoms".Phys. Rev. Lett.71(13): 2159.Bibcode:1993PhRvL..71.2159C.doi:10.1103/physrevlett.71.2159.PMID10054598.
  35. ^abHetherington, Jack H. (1984). "Observations on the statistical iteration of matrices".Phys. Rev. A.30(2713): 2713–2719.Bibcode:1984PhRvA..30.2713H.doi:10.1103/PhysRevA.30.2713.
  36. ^Fermi, Enrique; Richtmyer, Robert D. (1948)."Note on census-taking in Monte Carlo calculations"(PDF).LAM.805(A).Declassified report Los Alamos Archive
  37. ^abRosenbluth, Marshall N.; Rosenbluth, Arianna W. (1955)."Monte-Carlo calculations of the average extension of macromolecular chains".J. Chem. Phys.23(2): 356–359.Bibcode:1955JChPh..23..356R.doi:10.1063/1.1741967.S2CID89611599.
  38. ^Gordon, N.J.; Salmond, D.J.; Smith, A.F.M. (April 1993). "Novel approach to nonlinear/non-Gaussian Bayesian state estimation".IEE Proceedings F - Radar and Signal Processing.140(2): 107–113.doi:10.1049/ip-f-2.1993.0015.ISSN0956-375X.S2CID12644877.
  39. ^Kitagawa, G. (1996). "Monte carlo filter and smoother for non-Gaussian nonlinear state space models".Journal of Computational and Graphical Statistics.5(1): 1–25.doi:10.2307/1390750.JSTOR1390750.
  40. ^abDel Moral, Pierre (1996)."Non Linear Filtering: Interacting Particle Solution"(PDF).Markov Processes and Related Fields.2(4): 555–580. Archived fromthe original(PDF)on March 4, 2016.RetrievedJune 11,2015.
  41. ^Carvalho, Himilcon; Del Moral, Pierre; Monin, André; Salut, Gérard (July 1997)."Optimal Non-linear Filtering in GPS/INS Integration"(PDF).IEEE Transactions on Aerospace and Electronic Systems.33(3): 835–850.Bibcode:1997ITAES..33..835C.doi:10.1109/7.599254.S2CID27966240.Archived fromthe original(PDF)on November 10, 2022.RetrievedJune 11,2015.
  42. ^P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: An unified framework for particle solutions". LAAS-CNRS, Toulouse, Research Report no. 91137, DRET-DIGILOG- LAAS/CNRS contract, April (1991).
  43. ^P. Del Moral, G. Rigal, and G. Salut. "Nonlinear and non Gaussian particle filters applied to inertial platform repositioning." LAAS-CNRS, Toulouse, Research Report no. 92207, STCAN/DIGILOG-LAAS/CNRS Convention STCAN no. A.91.77.013, (94p.) September (1991).
  44. ^P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle resolution in filtering and estimation: Experimental results". Convention DRET no. 89.34.553.00.470.75.01, Research report no.2 (54p.), January (1992).
  45. ^P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle resolution in filtering and estimation: Theoretical results". Convention DRET no. 89.34.553.00.470.75.01, Research report no.3 (123p.), October (1992).
  46. ^P. Del Moral, J.-Ch. Noyer, G. Rigal, and G. Salut. "Particle filters in radar signal processing: detection, estimation and air targets recognition". LAAS-CNRS, Toulouse, Research report no. 92495, December (1992).
  47. ^P. Del Moral, G. Rigal, and G. Salut. "Estimation and nonlinear optimal control: Particle resolution in filtering and estimation". Studies on: Filtering, optimal control, and maximum likelihood estimation. Convention DRET no. 89.34.553.00.470.75.01. Research report no.4 (210p.), January (1993).
  48. ^Del Moral, Pierre (1998)."Measure Valued Processes and Interacting Particle Systems. Application to Non Linear Filtering Problems".Annals of Applied Probability.8(2) (Publications du Laboratoire de Statistique et Probabilités, 96-15 (1996) ed.): 438–495.CiteSeerX10.1.1.55.5257.doi:10.1214/aoap/1028903535.
  49. ^Crisan, Dan; Gaines, Jessica; Lyons, Terry (1998). "Convergence of a branching particle method to the solution of the Zakai".SIAM Journal on Applied Mathematics.58(5): 1568–1590.doi:10.1137/s0036139996307371.S2CID39982562.
  50. ^Crisan, Dan; Lyons, Terry (1997)."Nonlinear filtering and measure-valued processes".Probability Theory and Related Fields.109(2): 217–244.doi:10.1007/s004400050131.S2CID119809371.
  51. ^Crisan, Dan; Lyons, Terry (1999)."A particle approximation of the solution of the Kushner–Stratonovitch equation".Probability Theory and Related Fields.115(4): 549–578.doi:10.1007/s004400050249.S2CID117725141.
  52. ^Crisan, Dan; Del Moral, Pierre; Lyons, Terry (1999)."Discrete filtering using branching and interacting particle systems"(PDF).Markov Processes and Related Fields.5(3): 293–318.
  53. ^Del Moral, Pierre; Guionnet, Alice (1999). "On the stability of Measure Valued Processes with Applications to filtering".C. R. Acad. Sci. Paris.39(1): 429–434.
  54. ^Del Moral, Pierre; Guionnet, Alice (2001)."On the stability of interacting processes with applications to filtering and genetic algorithms".Annales de l'Institut Henri Poincaré.37(2): 155–194.Bibcode:2001AIHPB..37..155D.doi:10.1016/s0246-0203(00)01064-5.
  55. ^Ripley 1987
  56. ^abSawilowsky 2003
  57. ^Kalos & Whitlock 2008
  58. ^Shojaeefard, M.H.; Khalkhali, A.; Yarmohammadisatri, Sadegh (2017). "An efficient sensitivity analysis method for modified geometry of Macpherson suspension based on Pearson Correlation Coefficient".Vehicle System Dynamics.55(6): 827–852.Bibcode:2017VSD....55..827S.doi:10.1080/00423114.2017.1283046.S2CID114260173.
  59. ^Davenport 1992
  60. ^Route, Matthew (August 10, 2017)."Radio-flaring Ultracool Dwarf Population Synthesis".The Astrophysical Journal.845(1): 66.arXiv:1707.02212.Bibcode:2017ApJ...845...66R.doi:10.3847/1538-4357/aa7ede.S2CID118895524.
  61. ^Vose 2008,p. 13.
  62. ^Vose 2008,p. 16.
  63. ^Jia, Xun; Ziegenhein, Peter; Jiang, Steve B (2014)."GPU-based high-performance computing for radiation therapy".Physics in Medicine and Biology.59(4): R151–R182.Bibcode:2014PMB....59R.151J.doi:10.1088/0031-9155/59/4/R151.PMC4003902.PMID24486639.
  64. ^Hill, R.; Healy, B.; Holloway, L.; Kuncic, Z.; Thwaites, D.; Baldock, C. (March 2014). "Advances in kilovoltage x-ray beam dosimetry".Physics in Medicine and Biology.59(6): R183–R231.Bibcode:2014PMB....59R.183H.doi:10.1088/0031-9155/59/6/R183.PMID24584183.S2CID18082594.
  65. ^Rogers, D.W.O. (2006). "Fifty years of Monte Carlo simulations for medical physics".Physics in Medicine and Biology.51(13): R287–R301.Bibcode:2006PMB....51R.287R.doi:10.1088/0031-9155/51/13/R17.PMID16790908.S2CID12066026.
  66. ^Baeurle 2009
  67. ^Möller, W.; Eckstein, W. (March 1, 1984). "Tridyn — A TRIM simulation code including dynamic composition changes".Nuclear Instruments and Methods in Physics Research Section B: Beam Interactions with Materials and Atoms.2(1): 814–818.Bibcode:1984NIMPB...2..814M.doi:10.1016/0168-583X(84)90321-5.
  68. ^MacGillivray & Dodd 1982
  69. ^Golden 1979
  70. ^G. A. Bird, Molecular Gas Dynamics, Clarendon, Oxford (1976)
  71. ^Dietrich, S.; Boyd, I. (1996)."A Scalar optimized parallel implementation of the DSMC technique".Journal of Computational Physics.126(2): 328–42.Bibcode:1996JCoPh.126..328D.doi:10.1006/jcph.1996.0141.
  72. ^Chen, Shang-Ying; Hsu, Kuo-Chin; Fan, Chia-Ming (March 15, 2021). "Improvement of generalized finite difference method for stochastic subsurface flow modeling".Journal of Computational Physics.429:110002.Bibcode:2021JCoPh.42910002C.doi:10.1016/J.JCP.2020.110002.S2CID228828681.
  73. ^Climate Change 2013 The Physical Science Basis(PDF).Cambridge University Press.2013. p. 697.ISBN978-1-107-66182-0.RetrievedJuly 6,2023.
  74. ^Ojeda et al. 2009.
  75. ^Milik & Skolnick 1993.
  76. ^Cassey; Smith (2014). "Simulating confidence for the Ellison-Glaeser Index".Journal of Urban Economics.81:93.doi:10.1016/j.jue.2014.02.005.
  77. ^Sawilowsky & Fahoome 2003
  78. ^Spall, James C. (2005). "Monte Carlo Computation of the Fisher Information Matrix in Nonstandard Settings".Journal of Computational and Graphical Statistics.14(4): 889–909.CiteSeerX10.1.1.142.738.doi:10.1198/106186005X78800.S2CID16090098.
  79. ^Das, Sonjoy; Spall, James C.; Ghanem, Roger (2010). "Efficient Monte Carlo computation of Fisher information matrix using prior information".Computational Statistics & Data Analysis.54(2): 272–289.doi:10.1016/j.csda.2009.09.018.
  80. ^Chaslot, Guillaume; Bakkes, Sander; Szita, Istvan; Spronck, Pieter."Monte-Carlo Tree Search: A New Framework for Game AI"(PDF).Sander.landofsand.com.RetrievedOctober 28,2017.
  81. ^"Monte Carlo Tree Search - About".Archived fromthe originalon November 29, 2015.RetrievedMay 15,2013.
  82. ^Chaslot, Guillaume M. J. -B; Winands, Mark H. M.; Van Den Herik, H. Jaap (2008). "Parallel Monte-Carlo Tree Search".Computers and Games.Lecture Notes in Computer Science. Vol. 5131. pp. 60–71.CiteSeerX10.1.1.159.4373.doi:10.1007/978-3-540-87608-3_6.ISBN978-3-540-87607-6.
  83. ^Bruns, Pete.Monte-Carlo Tree Search in the game of Tantrix: Cosc490 Final Report(PDF)(Report).
  84. ^Silver, David; Veness, Joel."Monte-Carlo Planning in Large POMDPs"(PDF).0.cs.ucl.ac.uk.Archived fromthe original(PDF)on July 18, 2016.RetrievedOctober 28,2017.
  85. ^Lorentz, Richard J. (2011). "Improving Monte–Carlo Tree Search in Havannah".Computers and Games.Lecture Notes in Computer Science. Vol. 6515. pp. 105–115.Bibcode:2011LNCS.6515..105L.doi:10.1007/978-3-642-17928-0_10.ISBN978-3-642-17927-3.
  86. ^Jakl, Tomas."Arimaa challenge – comparison study of MCTS versus alpha-beta methods"(PDF).Arimaa.com.RetrievedOctober 28,2017.
  87. ^Szirmay-Kalos 2008.
  88. ^"How the Coast Guard Uses Analytics to Search for Those Lost at Sea".Dice Insights.January 3, 2014.
  89. ^Stone, Lawrence D.; Kratzke, Thomas M.; Frost, John R."Search Modeling and Optimization in USCG's Search and Rescue Optimal Planning System (SAROPS)"(PDF).Ifremer.fr.RetrievedOctober 28,2017.
  90. ^"Project Risk Simulation (BETA)".risk.octigo.pl.RetrievedMay 21,2024.
  91. ^Carmona, René; Del Moral, Pierre; Hu, Peng; Oudjane, Nadia (2012). "An Introduction to Particle Methods with Financial Applications". In Carmona, René A.; Moral, Pierre Del; Hu, Peng; et al. (eds.).Numerical Methods in Finance.Springer Proceedings in Mathematics. Vol. 12. Springer Berlin Heidelberg. pp. 3–49.CiteSeerX10.1.1.359.7957.doi:10.1007/978-3-642-25746-9_1.ISBN978-3-642-25745-2.
  92. ^Kroese, D. P.; Taimre, T.; Botev, Z. I. (2011).Handbook of Monte Carlo Methods.John Wiley & Sons.
  93. ^Arenas, Daniel J.; Lett, Lanair A.; Klusaritz, Heather; Teitelman, Anne M. (2017)."A Monte Carlo simulation approach for estimating the health and economic impact of interventions provided at a student-run clinic".PLOS ONE.12(12): e0189718.Bibcode:2017PLoSO..1289718A.doi:10.1371/journal.pone.0189718.PMC5746244.PMID29284026.
  94. ^Elwart, Liz; Emerson, Nina; Enders, Christina; Fumia, Dani; Murphy, Kevin (December 2006)."Increasing Access to Restraining Orders for Low Income Victims of Domestic Violence: A Cost-Benefit Analysis of the Proposed Domestic Abuse Grant Program"(PDF).State Bar of Wisconsin.Archived fromthe original(PDF)on November 6, 2018.RetrievedDecember 12,2016.
  95. ^Dahlan, Hadi Akbar (October 29, 2021)."Perbandingan Penerbitan dan Harga Buku Mengikut Genre di Malaysia dan Jepun Menggunakan Data Akses Terbuka dan Simulasi Monte Carlo"(PDF).Kajian Malaysia.39(2): 179–202.doi:10.21315/km2021.39.2.8.S2CID240435973.
  96. ^abPress et al. 1996
  97. ^MEZEI, M (December 31, 1986). "Adaptive umbrella sampling: Self-consistent determination of the non-Boltzmann bias".Journal of Computational Physics.68(1): 237–248.Bibcode:1987JCoPh..68..237M.doi:10.1016/0021-9991(87)90054-4.
  98. ^Bartels, Christian; Karplus, Martin (December 31, 1997). "Probability Distributions for Complex Systems: Adaptive Umbrella Sampling of the Potential Energy".The Journal of Physical Chemistry B.102(5): 865–880.doi:10.1021/jp972280j.
  99. ^Del Moral, Pierre; Doucet, Arnaud; Jasra, Ajay (2006). "Sequential Monte Carlo samplers".Journal of the Royal Statistical Society, Series B.68(3): 411–436.arXiv:cond-mat/0212648.doi:10.1111/j.1467-9868.2006.00553.x.S2CID12074789.
  100. ^Spall, J. C. (2003),Introduction to Stochastic Search and Optimization: Estimation, Simulation, and Control,Wiley, Hoboken, NJ.http://www.jhuapl.edu/ISSO
  101. ^Mosegaard & Tarantola 1995
  102. ^Tarantola 2005
  103. ^McCracken, D. D., (1955) The Monte Carlo Method, Scientific American, 192(5), pp. 90-97
  104. ^Elishakoff, I., (2003) Notes on Philosophy of the Monte Carlo Method, International Applied Mechanics, 39(7), pp.753-762
  105. ^Grüne-Yanoff, T., & Weirich, P. (2010). The philosophy and epistemology of simulation: A review, Simulation & Gaming, 41(1), pp. 20-50

Sources

[edit]
[edit]