Re-posted from: https://albi3ro.github.io/M4/numerics/Monte-Carlo.html
Monte Carlo- Random Numbers to Improve Calculations
When one hears “Monte Carlo”, most people might think of something like this:
Monte Carlo, Monaco: known for extremely large amounts of money, car racing, no income taxes,and copious gambling.
In addition to Monaco, Europe, Las Vegas decided to host a Monte Carlo-themed casino as well. So during the Manhattan project, when the best minds in the United States were camped out in the New Mexican desert, they had plenty of inspiration from Las Vegas, and plenty of difficult problems to work on in the form of quantifying the inner workings of nuclei. Enrico Fermi first played with these ideas, but Stanislaw Ulam invented the modern Monte Carlo Markov Chain later.
At the same time, these scientists now had computers at their disposal. John von Neumann programmed Ulam’s algorithm onto ENIAC, Electronic Numerical Integrator and Computer, the very first electronic, general purpose computer, even though it did still run on vacuum tubes.
That still doesn’t answer, why do random numbers actually help us solve problems?
Imagine you are visiting a new city for the first time, maybe Monte Carlo. You only have a day or two, and you want to really get to know the city. You have two options for your visit
- Hit the tourist sites you researched online
- Wander around. Try and communicate with the locals. Find an out-of-the-way restaurant and sample food not tamed for foreigners. Watch people interact. Get lost.
Both are legitimate ways to see the city. But depending on what you want, you might choose a different option. The same goes for exploring physics problems. Sometimes you want to go in and study just everything you knew about beforehand, but sometimes you need to wander around, not hit everything, but get a better feeling for what everything might be like.
Buffon’s Needle: Calculation of π
Even back in the 18th century, Georges-Louis Leclerc, Comte de Buffon posed a problem in geometric probability. Nowdays, we use a slightly different version of that problem to calculate π and illustrate Monte Carlo simulations.
Suppose we have a square dartboard, and someone with really bad, completely random aim, even though he/she always at least hits inside the dartboard. We then inscribe a circle inside that dartboard. After an infinity number of hits, what is the ratio between hits in the circle, and hits in the square?
Onto the Code!
#Pkg.update() #Pkg.add("PyPlot") using PyPlot
We will generate our random numbers on the unit interval. Thus the radius in our circumstance is $.5$.
Write a function
incircle(r2) such that if
r2 is in the circle, it returns true, else, it returns false. We will use this with the julia function
r2 is the radius squared, and already centered around the middle of the unit circle
function incircle(r2) if r2<.25 return true else return false end end
#The number of darts we will throw at the board. We will see how accurate different numbers are N=[10,25,50,75,100,250,500,750,1000,2500,5000,7500,10000]; # We will perform each number multiple times in order to calculate error bars M=15;
πapprox=zeros(Float64,length(N),M); for ii in 1:length(N) #initialize an array of proper size X=zeros(Float64,N[ii],2); for jj in 1:M #popular our array with random numbers on the unit interval rand!(X); #calculate their radius squared R2=(X[:,1]-.5).^2+(X[:,2]-.5).^2 # 4*number in circle / total number πapprox[ii,jj]=4.*length(filter(incircle,R2))/N[ii]; end end
# Get our averages and standard deviations πave=sum(πapprox,2)/M; πstd=std(πapprox,2);
So that was a nice, short little piece of code. Lets plot it now to see means.
title("Monte Carlo Estimation of π") ylabel("π estimate") xlabel("N points") plot(N,π*ones(N)); for j in 1:M scatter(N,πapprox[:,j],marker="o",color="green"); end ax=gca() errorbar(N,π*ones(N),yerr=πstd,color="red",fmt="o") ax[:set_xscale]("log"); ax[:set_xlim]([5,5*10^4]);
When we have fewer numbers of points, our estimates vary much more wildly, and much further from 3.1415926 .
But, at least, the guesses from our different runs all seem equally distributed around the correct value, so it seems we have no systematic error.
As we get up to $10^4$, our estimate starts getting much more accurate and consistent.
title("Dependence of Monte Carlo Error on Number of Points") ylabel("standard deviation") xlabel("N points") semilogx(N,πstd,marker="o");
So what we guessed in the first plot about dispersion in estimate, we quantify here in this plot. When we only have 10 darts, the guesses vary by up to .3, but when we get down to 1,000 trials, we are starting to be consistent to .0002
title("Overall Averages") xlabel("N steps") ylabel("Average of 15 runs") semilogx(N,π*ones(N)); semilogx(N,πave,marker="o");
Now lets just make a graphical representation of what we’ve been doing this whole time. Plot our points on unit square, and color the ones inside a circle a different color.
X=zeros(Float64,1000); Y=zeros(Float64,1000); rand!(X); rand!(Y); R2=(X-.5).^2+(Y-.5).^2; Xc=; Yc= for ii in 1:length(X) if R2[ii]<.25 push!(Xc,X[ii]); push!(Yc,Y[ii]); end end title("The dartboard") xlim(0,1) ylim(0,1) scatter(X,Y); scatter(Xc,Yc,color="red");
That’s all folks!
Now here’s a picture of some pie to congratulate you on calculating π.