Paper

Randomized quasi-Monte Carlo simulation of fast-ion thermalization

, and

Published 31 July 2012 © 2012 IOP Publishing Ltd
, , Citation L J Höök et al 2013 Comput. Sci. Discov. 5 014010 DOI 10.1088/1749-4699/5/1/014010

1749-4699/5/1/014010

Abstract

This work investigates the applicability of the randomized quasi-Monte Carlo method for simulation of fast-ion thermalization processes in fusion plasmas, e.g. for simulation of neutral beam injection and radio frequency heating. In contrast to the standard Monte Carlo method, the quasi-Monte Carlo method uses deterministic numbers instead of pseudo-random numbers and has a statistical weak convergence close to $\mathcal {O}(N^{-1})$ , where N is the number of markers. We have compared different quasi-Monte Carlo methods for a neutral beam injection scenario, which is solved by many realizations of the associated stochastic differential equation, discretized with the Euler–Maruyama scheme. The statistical convergence of the methods is measured for time steps up to 214.

Export citation and abstract BibTeX RIS

1. Introduction

The quasi-Monte Carlo method is a well-established method for improving the statistical convergence for problems solved with the standard Monte Carlo method. It is commonly used for improving high-dimensional integration, but there is growing research on the applicability of the method for simulation of diffusion processes. Simulation of diffusion is achieved by moving particles in the Lagrangian framework according to the associated stochastic differential equation (SDE). Examples in fusion plasma physics are the simulation of neutral beam injection [1, 2] and radio frequency heating [3]. In the simulation, two sources of error appear; the time discretization error from the numerical scheme of the SDE and the statistical error from the finite number of particles describing the density. The standard Monte Carlo method gives a statistical convergence of αNβ with β = −1/2. To improve this value one often tries to reduce the value of α with some sort of variance reduction method. Common methods are the importance sampling and control-variate methods. In the plasma community a variance reduction concept known as the δf-method is commonly used, which is a family of different methods based on combinations of the importance sampling and control-variate methods [49]. Instead of reducing α, the noise can be reduced by improving the order of convergence, β. This can be achieved by replacing the pseudo-random numbers in the Monte Carlo method with deterministic numbers with the low-discrepancy property. An early application of low-discrepancy sequences is the quiet start method in particle-in-cell simulations [10, 11] where quasi-random points are used for sampling the initial particle distribution.

The asymptotic convergence of the quasi-Monte Carlo method is,

Equation (1)

where s is the number of dimensions. For modest dimensions, the log(N)s term can be ignored. The convergence is dependent on the number of dimensions, forcing a greater number of particles to be used as the number of dimensions increase. For integration problems over the unit hypercube we require at least a $N \geqslant \exp (s)$ number of particles, (from ∂NN−1log(N)s = 0), for (1) to converge. The number of particles required explodes as the number of dimensions increase. This suggests that the quasi-Monte Carlo method is only efficient for small dimensions and suffers from the 'curse of dimensionality'. Fortunately it has been shown in [12] and references therein, that the quasi-Monte Carlo method has better convergence in practice than theory predicts. This can partly be explained by the fact that a large class of integrands have a so called low effective dimension; the function to be integrated has most of its variation located in few dimensions.

The purpose of this paper is to test the applicability of the quasi-Monte Carlo method for simulation of the fast-ion thermalization process of a neutral beam injection scenario. In this paper, the Brownian bridge method and the method used in [13], called 'the sorting and mixing method' herein, is tested for the unscrambled and scrambled Faure sequence, which is a quasi-random sequence in the family of (t,s)-sequences. In section 2 we will briefly introduce the Fokker–Planck equation and its connection to SDEs. In section 3 the quasi-Monte Carlo method is introduced where the concept of low-discrepancy sequence, randomization and effective dimensions are briefly touched upon. The sorting and mixing method and the Brownian bridge method are presented in sections 3.6 and 3.7. In section 4, we derive the SDE for the fast-ion thermalization process, which is followed by the simulation results in section 5 and ends with conclusions.

2. The Fokker–Planck equation

Consider the Fokker–Planck equation for particle interactions in s coordinates x = x1,...,xs,

Equation (2)

with initial condition f(y,0) = δ(x − y) where Ai and $B_{ij} = \sum _{l=1}^s \sigma _{il} \sigma _{jl}$ are the drift vector and diffusion tensor respectively, both independent of time. The 'characteristics' of (2) are described by an Itô SDE

Equation (3)

where dW(t) is the Wiener process also known as Brownian motion, with normal distributed components having zero mean and dt variance. Many realizations of the SDE give a distribution of particles, which is the solution of (2). A common time-discretization method of the above SDE is the Euler–Maruyama scheme [14]. Let i = {1,...,I} and define the end time by T = IΔt and assume $X \in \mathbb {R}$ ; then the Euler–Maruyama scheme of (3) is,

Equation (4)

where ΔWi = Wi+1 − Wi are normally distributed pseudo-random numbers with zero mean and (ti+1 − ti)-variance. In the following section we will discuss how the pseudo-random numbers can be replaced with quasi-random sequences.

3. The quasi-Monte Carlo method

The naive approach of replacing the pseudo-random numbers with quasi-random numbers will not work for simulation of SDEs since the quasi-random numbers are deterministic and the correlation will introduce artificial drift of the the particles. However, it is possible to use the quasi-random points, if combined with methods that break the correlation.When using the quasi-Monte Carlo method for SDEs the term dimension has a different meaning than the physical number of dimensions. The number of dimensions of a process, X(t), which solves a certain SDE is the number of time steps times the number of physical dimensions. From here onward we let s denote the physical number of dimensions. This is most easily explained by the following example.

3.1. Example

Consider a one-dimensional process X(t) satisfying the scalar SDE dX = σ(X) dW with initial condition X0 = x. We would like to simulate this SDE for three time steps using the discretization of Euler–Maruyama. Unrolling the SDE we obtain,

where ξiN(0,1), are normally distributed random numbers with zero mean and unit variance. Clearly after three time steps the process X is dependent on three random numbers. Now assume that we are interested in calculating an expected value of the process 〈g(X(t3))〉 for a known function g(·). We know that the normally distributed random numbers ξi are related to uniformly distributed numbers ziU(0,1) over the unit interval [0, 1) by the inverse cumulative normal distribution ξ = Φ−1(z). Using this in the above sequence we obtain,

The expected value of g(X3) is now a four-dimensional integration,

Equation (5)

From this example we clearly see how the number of dimensions is dependent on the number of physical dimensions and the number of time steps. An estimate of the expected value is obtained by sampling the hypercube with a finite number of particles,

Equation (6)

If the hypercube is sampled with quasi-random points, we obtain a more accurate estimate of the expected value than if we would use pseudo-random points. This is because the quasi-random points are more uniformly distributed than pseudo-random numbers and do not form clusters, which is a common effect for pseudo-random numbers.

3.2. Low discrepancy sequences

Two common classes of quasi-random sequences are (digital) nets/sequences and lattice rules [15]. In this paper we have used the Faure sequence [15], which is a (0,s)-sequence in base b, where b is the smallest prime number greater than the number of dimensions s. Similar to pseudo-random numbers the law of large numbers must hold for these sequences. A deterministic version of the law of large numbers is provided by the Koksma–Hlawka inequality, which gives a bound on the quadrature error,

Equation (7)

where VHK(g) is bounded variation of g in the sense of Hardy and Krause [15]. The exact definition of the Hardy and Krause variation contains higher order derivatives of g. Therefore the smoothness of g determines the size of VHK and is small for smooth functions. D*N is called the discrete star discrepancy and it measures how uniformly the samples are distributed over the s-dimensional unit cube. A sequence with a small D*N is called a low-discrepancy sequence. The discrete star discrepancy is defined as,

Equation (8)

where $\mathbb {I}$ is the indicator function and is 1 if Zj is inside the box [0,p)s and zero otherwise. The second term, vol measures the volume of the s-dimensional box [0,p)s. Equation (8) measures the maximum difference between the number of points in the box and the volume of the box over all possible boxes with one corner in the origin. An illustration of the star discrepancy in two dimensions is given in figure 1.

Figure 1.

Figure 1. Illustration of the discrepancy for the first 50 points of the two-dimensional Faure sequence in base 3. The area of the smallest box plotted in the figure is 0.32 = 0.09 and the number of points in this box is 5, which gives an estimate of $1/N\sum \mathbb {I}_{\mathrm {box}} = 5/50 = 0.1$ . The discrepancy for this box is |0.1–0.09| = 0.01. Similarly the area of the second smallest box is 0.42 = 0.16 and the number of points is 8, which gives 8/50 = 0.16 and a discrepancy of 0.

Standard image

3.3. Randomization

Scrambling is a randomization method for obtaining new uncorrelated (digital) nets or sequences generated from a common mother net/sequence and can partly break the correlation between particles. There are many different randomization methods, the simplest method is so called linear scrambling also known as the generalized Faure for scrambled Faure sequences, [16, 17]. The idea of scrambling is to combine the benefit of calculating sample variance, as in the standard Monte Carlo method, with the improved convergence of the quasi-Monte Carlo method. Scrambling shuffles the position of the points and is achieved by randomizing the digits in the b-adic expansion with a matrix vector multiplication of a matrix with random elements. In short, the nth s-dimensional quasi-random point is generated by,

where $\Phi _{b}(\mathbf {n}) = \frac {n_0}{b} + \frac {n_1}{b^2} + \cdots + \frac {n_m}{b^m}$ , is the so called radical inverse. The vector n = [nm,...,n0] is the b-adic representation of the scalar n, (n = nmbm + ··· + n1b + n0). The matrix Ps is the sth power of the Pascal matrix modulo b where the (k,l)-element of P is equal to $\left({{l-1}\atop {k-1}}\right){\mathrm { mod}}\,b$ . The matrices, L(s) are independent lower triangular matrices with diagonal elements selected randomly from {1,...,b − 1} and the other elements chosen randomly from {0,...,b − 1}. For a more in depth treatment on randomization we refer the reader to [15, 1820] (see figure 2).

Figure 2.

Figure 2. Two-dimensional projection plot (dim 37, dim 38) of the first hundred points from the 38-dimensional Faure sequence: (a) before, (b) after scrambling and (c) a plot of hundred pseudo-random points.

Standard image

3.4. Effective dimension

High-dimensional problems depend primarily on a few important dimensions, thus reducing the 'effective dimensionality'. This property was analyzed in [21, 22] in terms of the truncated ANOVA decomposition fd of a function f. The effective dimension can be defined as the lowest possible dimension of fd for which,

Equation (9)

We refer the reader to [23] for a detailed description on effective dimensions.

3.5. Quasi-Monte Carlo for diffusion

The quasi-Monte Carlo method for simulation of diffusion can be used in two ways. In the first concept, the Brownian motion is simulated using an s × I-dimensional sequence. Here I is the number of time steps. The other alternative is to use a 2s-dimensional sequence. Successive realizations of the Brownian motion are obtained by shifting the index pointer to the sequence with N, e.g. use the N first points for the first time-step and use the [N + 1,2N] points for the second time-step and so forth. What is important for both methods is that the realizations should be uncorrelated. In the beginning of this section, we mentioned the importance of breaking the correlation and pointed out that scrambling can partly break the correlation. This is only true in high dimensions. The difference in the auto-correlation between elements of 50 one-dimensional scrambled Faure points and the auto-correlation between elements of one 50-dimensional scrambled Faure point is illustrated in figures 3(a) and (b). The figures show that correlation is very strong between points in the one-dimensional case but not very strong in the 50-dimensional case, which is of the same order as for a standard pseudo-random generator. Scrambling is apparently not sufficient for breaking correlation in the low-dimensional case without using more advanced techniques. In [13], a method for resolving this issue was suggested. It is based on sorting and mixing the particles in a clever way such that decorrelation is ensured, and is the topic of the next section.

Figure 3.

Figure 3. Difference in auto-correlation between: (a) 50 one-dimensional scrambled Faure points and (b) one 50-dimensional scrambled Faure point.

Standard image

3.6. The sorting and mixing method

The idea of the sorting and mixing method is to break the correlation by sorting the particles along each dimension. We will give a simplified presentation of the method. For an in depth treatment of the concept we refer the reader to [13]. The foremost drawback of this method is that it requires at least a bs number of particles being simulated where b is the least prime number greater than 2s. In six dimensions this would give 136 ≈ 4 × 106 number of particles. For higher dimensions this value will explode.

The first step of the sorting and mixing method is to treat the drift and diffusion part of (3) separately,

Equation (10)

Equation (11)

The drift (10) is evaluated with the standard forward Euler method. For the diffusion (11) part we need to generate quasi-random numbers for the Wiener process. Let N = bm be the number of particles to be simulated. Here b is the least prime number greater than 2s, e.g. for s = 2⇒b = 5, and let d1,...,ds be integers greater than zero such that m = d1 + ··· + ds. For a two-dimensional diffusion process a minimum of N = 52 particles are needed.

  • Sorting. First sort the particles in ascending order of the magnitude of the first coordinate of the particles into bd1 groups. Secondly sort the particles in each group in ascending order of the magnitude of the second coordinate into bd2 subgroups. For a three-dimensional problem sort the particles in each bd2 subgroup in ascending order of the magnitude of the third coordinate of the particles into bd3 sub-subgroups. Continue sorting if dimensions are greater than three. This is a computationally demanding operation requiring $\mathcal {O}(N \log N)$ operations (with quick-sort) for each sort.
  • Mixing. Consider a vector y = (y1,...,y2s) from a (0,2s)-net in base b and define two selection functions P' and P'' by P'y: = (y1,...,ys) and P''y: = (ys+1,...,y2s). Use the first s dimension of the (0,2s)-net for mixing. Let a(y): = (⌊ bd1y1 ⌋,...,⌊ bdsys) ⌋) where the floor function, ⌊ · ⌋, returns the greatest integer less than or equal to the argument. Define the one-to-one map between array indices using the first s variables of the (0,2s)-net, j → a(P'yiN+j). The map is from [0,N) into [0,bd1× ··· × [0,bds), e.g. in two dimensions we map the vector [0,N) to the matrix [0,bd1× [0,bd2). The particle with array index a(P'yiN+j) is pushed by the quasi-random points, P''yiN+j.

The sorting and mixing procedures are illustrated in figures 4(a) and (b) for s = 2. It was proved in [13] that this construction has the following bound:

Equation (12)

where 0 ⩽ j < N and D*N(Yk) is the star discrepancy of the (0,2s)-net, Yk in base b. Since N is defined as the product of bdi, the second term converges in the limit of infinite number of particles. Note that for a finite N the bound increases with the number of time steps, i. To keep the error fixed, more particles are needed as the simulation time increases.

Figure 4.

Figure 4. Illustration of the (a) sorting and (b) mixing stages for a two-dimensional case where b = 5 and d1 = 1,d2 = 1. In (a) the particles are first sorted in ascending order according to the first coordinate value, sorted according to the particles physical position e.g. (x = 2,y = 16). Secondly bd2 subgroups are formed and the particles are sorted locally in each subgroup according to the second coordinate value. Mixing is achieved by selecting the particle (array position) with the first two random numbers y1,y2 from the quasi-random vector, y2s = (y1,y2,y3,y4). The array index of the particle is obtained by ⌊ bd1y1 ⌋ and ⌊ bd2y2 ⌋ where the floor function returns the greatest integer less than the argument. The particles are pushed with the remaining quasi-random numbers y3,y4.

Standard image

3.7. The Brownian bridge method

In this section we will turn our attention to the s × I-dimensional case. If many dimensions are used, the number of effective dimensions become important. It has been argued in [12, 21], that the midpoint Brownian bridge method effectively reduces the effective dimensions. The general Brownian bridge formula is defined as

Equation (13)

where λ = (j − i)/(k − i) for ti < tj ⩽ tk and Z is a normally distributed random number. The variables wti,wtk are previous realization of the Wiener process. The midpoint Brownian bridge formula is obtained from the special case where the time axis is populated by successive midpoint splits. First, generate the start- w0 and end-point wT . Then draw the midpoint and continue splitting, see figure 5. The generated order is: w0,wT ,wT/2,wT/4,w3T/4,...,wT−1. One interesting note is that the midpoint Brownian bridge construction reduces the variance of Wtj by a factor of 2 for each split,

The success of the Brownian bridge method can be explained by the way it utilizes low-discrepancy sequences. An important property of low-discrepancy sequences is that they are typically more uniformly distributed in lower dimensions than in higher. This is efficiently used by the Brownian bridge construction, since the first few quasi-random points describe the complete Brownian path on a coarse level. On the finer scale the reduction of variance in the Brownian bridge construction, reduces the impact of the degradation in the low-discrepancy sequence. It should be noted that the Brownian bridge method does not always perform better than the standard Monte Carlo method. For example in [24] it is shown that the performance of the Brownian bridge construction is not consistent and can actually perform worse than the standard Monte Carlo method for certain types of integrands. Thus the Brownian bridge method is problem dependent and has the potential to generate many papers.

Figure 5.

Figure 5. Illustration of the midpoint Brownian bridge.

Standard image

4. The stochastic differential equation for fast-ion thermalization

In this section we present an SDE for a simple model of fast-ion thermalization in fusion plasmas given by the Fokker–Planck equation with the Coulomb collision operator of Spitzer [25].

Equation (14)

The above equation is not of the form of (2) due to the Jacobian v2, therefore the drift A and diffusion coefficients σ are calculated from the moments of the the Coulomb operator,

The first two moments of the above operator are given by,

Equation (15)

Equation (16)

Equation (17)

Equation (18)

Equation (19)

where we have introduced the variables, following [25], $\alpha (v) = \langle \Delta v_{\Vert} \rangle + \frac {1}{2v}\langle (\Delta v_\perp )^2\rangle $ , β(v) = 〈(Δv)2〉, γ(v) = 〈(Δv)2〉. The explicit expressions for the moments are,

where vf is the thermal velocity, Cf is given in [25] and f is the index over the species of the field ions. The function G is given by,

Armed with the equations above, the covariance matrix is calculated from the following general formula,

Equation (20)

Inserting (15)–(19) in the equation above we obtain the following:

Equation (21)

Since the nonzero elements of the covariance matrix are on the matrix diagonal, the diffusion coefficient σ is immediately obtained. The elements of the drift vector are given by,

Equation (22)

The drift and diffusion coefficients are in v,ξ coordinates. Using the Itô formula for E = g(V ) = mV2/2 we obtain an SDE in energy and pitch-angle,

Equation (23)

This system is solved with the Euler–Maruyama scheme,

Equation (24)

where Z1,Z2 are normal distributed random numbers with zero mean and unit variance.

5. Application: simulation of neutral beam injection

This section concerns modeling the thermalization of fast ions from neutral beam injection in fusion devices. In practice these models often aim at determining the heat and momentum transfer from the fast ions to the thermal plasma components, along with the fast ion current drive. For this purpose it is often sufficient to solve a linearized Fokker–Planck equation for the fast ions, in which the thermal components of the plasma are considered stationary, or quasi-stationary [25]. The separation of fast and thermal ions can be done in different ways, e.g. in the test particle Monte Carlo codes NUBEAM [26] and ASCOT [27] where the test particles are dropped when they reach 3/2 of the thermal energy. In this work the separation is achieved by simulating the test particles for half a collisional slowing down time against the electrons. Thus, the energy distributions calculated here differ slightly from those obtained with NUBEAM and ASCOT in the thermal energy range, but not for the supra-thermal energies. Furthermore, by assuming the background plasma to be stationary the equation becomes autonomous. This allows us to use the technique used in [27], where the steady state distribution function is proportional to the time that a set of test particles spend in each volume element.

Four methods have been tested, the standard Monte Carlo, the sorting and mixing method, the Brownian bridge method and a naive method where the pseudo-random numbers are replaced with Faure numbers, scrambled and unscrambled without any extra modification. Measures of the run times of the simulations are not considered since the implementation of the Faure quasi-random generator has not been optimized in contrast to the pseudo-random generator, which is highly optimized in MatlabTM. However, the sorting mixing method is slower than the Brownian bridge method since it requires sorting the particles each time step. The plasma is assumed to consist of protons and electrons, both with densities 3 × 1019 1 m−3 and temperatures 4 keV. The beam ions are protons injected with 100 keV energy and a pitch angle ξ = −0.8, where ξ = v/v, v is the speed and v is the velocity component along the magnetic field. The fast protons will collide primarily with the electron above 59 keV and with the thermal protons below. The total simulation time is half a slowing down time and is split into 27 time steps. The parameters for the sorting and mixing method are, d1 = {1,2,2,2,2}, d2 = {1,1,2,3,4} and b = 5 for N = b(d1+d2). The Faure sequence was scrambled to a depth of 30-digits. The tests was conducted in MatlabTM version R2010b on a twelve core Intel machine @ 2.67 GHz with 62 GB of RAM. Since the entire particle trajectory was saved, a massive amount of memory was required with a maximum of 59 × 214 = 32 × 109 particles. The exact solution for this test case is unknown. Therefore we have measured the convergence in terms of sample variance on a batch of 20 runs, each simulation with different initial seed of the randomly scrambled Faure sequence. Two different sample-means were considered, the parallel velocity $\langle g(E, \xi )\rangle = \langle v_{\Vert}\rangle = \langle v \xi \rangle =\langle \xi \sqrt {2E/m}\rangle $ , which is related to the fast-ion current drive and the energy squared 〈g(E)〉 = 〈E2〉. The batch-mean value is calculated from,

Equation (25)

and the variance of the batch-mean is given by,

Equation (26)

From these expressions we have calculated the 99% confidence interval for the batch-mean from the Student-distribution with M − 1 degrees of freedom,

Equation (27)

where t99%,M−1 = 2.86. Simulation results for the parallel velocity for the test cases are given in table 1 and for the energy squared test case in table 2 .

Table 1. Computational results for 〈vξ〉 in (Mm s−1), with the different methods. Scrambled Faure sequences have been used for all tested methods in the table. The result 59* in the last entry of the first column was obtained with M = 180, as a reference value.

N Monte Carlo Brownian bridge Sorting and mixing Scrambled Faure
52 −2.5862±0.0801 −2.5922±0.1319 −2.6189±0.0073 −2.5853±0.1008
53 −2.5761±0.0337 −2.5875±0.0356 −2.5923±0.0042 −2.5701±0.0467
54 −2.5783±0.0192 −2.5799±0.0080 −2.5832±0.0021 −2.5748±0.0068
55 −2.5807±0.0068 −2.5797±0.0020 −2.5806±0.0007 −2.5808±0.0014
56 −2.5804±0.0033 −2.5806±0.0003 −2.5804±0.0002 −2.5819±0.0004
59 −2.5805±0.0002      
59* −2.5804±0.00008      

Table 2. Computational results for 〈E2〉 with the different methods. Scrambled Faure sequences have been used for all tested methods in the table and the values have been normalized by 103. The result 59* in the last entry of the first column was obtained with M = 180, as a reference value.

N Monte Carlo Brownian bridge Sorting and mixing Scrambled Faure
52 5.1802±0.2026 5.2769±0.4364 5.07933±0.01355 5.1869±0.3130
53 5.0672±0.0923 5.1500±0.1631 5.08620±0.00717 5.2554±0.1334
54 5.0673±0.0354 5.1096±0.0365 5.09003±0.00124 5.1792±0.0206
55 5.0929±0.0154 5.0923±0.0060 5.09075±0.00013 5.1010±0.0047
56 5.0902±0.0085 5.0913±0.0007 5.09058±0.00003 5.0912±0.0011
59 5.0908±0.0007      
59* 5.09059±0.00026      

The convergence of the normalized root mean square error, $ \mathrm {n.r.m.s} = \sqrt {\hat {\sigma }^2}/( \hat {\mu }\sqrt {M} )$ for the two test cases is illustrated in figures 6(a) and (b). In these figures, BB denotes the Brownian bridge method and SM the method of sorting and mixing. The histograms of the stationary distributions for the different methods are plotted in figures 7(a)–10(b). A reference distribution, obtained from a standard Monte Carlo simulation, is plotted in figure 7(a). In figure 7(b), the distribution of the naive method with unscrambled Faure points is plotted. The method does not converge as expected since the correlation in the Faure sequence introduces spurious drifts of the particles. The result is slightly improved if combined with the Brownian bridge method, see figure 8(a), but the distribution does not agree with the reference figure 7(a). The sorting and mixing method is the only method with a theoretical proof of the convergence for unscrambled Faure points. The distribution for the sorting and mixing method is plotted in figure 8(b) and agrees with the standard Monte Carlo distribution. The distribution from a simulation with the scrambled Faure sequence is plotted in figure 9. From this figure we can see that the scrambling reduces the error from the correlation. Even though the distribution of the naive scrambled Faure method is in agreement with the Monte Carlo reference distribution, the moments are not in agreement for the case 〈vξ〉 as seen in table 1. The distribution for the scrambled Faure with the Brownian bridge method is plotted in figure 10(a) and it is in agreement with the Monte Carlo reference. The distribution for the sorting and mixing method is plotted in figure 10(b) and agrees with the Monte Carlo reference. The sorting and mixing method and the Brownian bridge method perform much better than standard Monte Carlo for 27 time steps.

Figure 6.

Figure 6. Normalized root mean square convergence of the expected values of: (a) parallel velocity, 〈vξ〉 and (b) energy squared, 〈E2〉.

Standard image
Figure 7.

Figure 7. Histogram plot of the distribution function obtained with: (a) the standard Monte Carlo and method (b) the naive method with unscrambled Faure points.

Standard image
Figure 8.

Figure 8. Histogram plot of the distribution function obtained with: (a) the Brownian bridge method using unscrambled Faure points and (b) the sorting and mixing method using unscrambled Faure points.

Standard image
Figure 9.

Figure 9. Histogram plot of the distribution function obtained with the naive method using scrambled Faure points.

Standard image
Figure 10.

Figure 10. Histogram plot of the distribution function obtained with: (a) the Brownian bridge method using scrambled Faure points and (b) the sorting and mixing method using scrambled Faure points.

Standard image

5.1. Very long-time simulation

In order to test the applicability for long-time particle simulations we have estimated the slope of the convergence with a least-squares fit for different number of time steps. The result is plotted in figure 11. The results indicate that the sorting and mixing method performs worse than the standard Monte Carlo above 400 time steps when measuring parallel velocity and remains very low for 〈E2〉 for all tested values. The convergence of the Brownian bridge method degrades almost linearly for both measured quantities, but the convergence is better than the standard Monte Carlo even for a thousand time steps. A rough extrapolation of the slope for the Brownian bridge method indicates that it will perform worse than the standard Monte Carlo at about 1700 time steps.

Figure 11.

Figure 11. Convergence plot of the exponent β in Nβ for different number of time steps. BB is the Brownian bridge method, SM is the sorting and mixing method and the convergence rates have been estimated for the moments 〈vξ〉, 〈E2〉. Here we can see that the convergence rate decreases as the number of time steps increase.

Standard image

In summary, we have tested the methods for a very large number of time steps, I = 214 = 16384. The computational results are given in tables 3 and 4. For this value the number of dimensions of the hypercube is 215 for the Brownian bridge method. The theoretical number of particles required to uniformly cover the hypercube is astronomical, therefore we can not expect the method to perform well. The sorting and mixing method cannot perform well either because the bound of the discrepancy (12) increases with the number of time steps, the sum in (12) is of the order of bd1+⌊ d2/2 ⌋ × I × D*(Y0) ≈ 107 × D*(Y0). The Brownian bridge method only gives an acceptable estimate of 〈E2〉 for N = 56 but has the same confidence interval as the standard Monte Carlo, table 4. The performance is slightly better for 〈vξ〉 where the value is estimated correctly except for N = 52 but the confidence interval is smaller than the standard Monte Carlo at N = 56, table 3. The sorting and mixing method does not converge for the parallel velocity case, as seen in table 3 and it gives an inconsistent estimate for the energy squared case, table 4.

Table 3. Computational results for 〈v||〉 with the Brownian bridge method and the sorting and mixing method at dt = 2−15.

N Monte Carlo Brownian bridge Sorting and mixing
52 −2.617±0.096 −1.859±0.174 −2.716±0.003
53 −2.604±0.025 −2.478±0.120 −2.681±0.004
54 −2.590±0.011 −2.545±0.034 −2.635±0.005
55 −2.588±0.005 −2.577±0.008 −2.600±0.005
56 −2.588±0.003 −2.585±0.001  

Table 4. Computational results for 〈E2〉 with the Brownian bridge method and the sorting and mixing method at dt = 2−15.

N Monte Carlo Brownian bridge Sorting and mixing
52 5.171±0.184 2.112±0.208 4.958±0.014
53 5.128±0.081 4.161±0.358 5.019±0.009
54 5.123±0.040 4.882±0.077 5.082±0.006
55 5.128±0.018 5.066±0.026 5.120±0.002
56 5.123±0.007 5.113±0.007  

6. Conclusion

We have tested the applicability of the quasi-Monte Carlo method on fast-ion thermalization using a simplified neutral beam injection model. The method of sorting and mixing and the Brownian bridge method are much better than the standard Monte Carlo method with a faster convergence and more accurate estimate up to a thousand time steps for simulation of fast-ion thermalization. When very long simulations are required these methods can fail to give accurate estimates and they may not converge. For modest number of time steps (27) the value of 〈vξ〉 converges as $\mathcal {O}(N^{-1})$ for the Brownian bridge method and as $\mathcal {O}(N^{-0.6})$ for the sorting and mixing method, while for 〈E2〉 both methods converge as $\mathcal {O}(N^{-1})$ . The sorting and mixing method is more accurate than the Brownian bridge method but is also more computationally demanding due to the sorting stage and requires simultaneous simulation of an ensemble of particles. The sorting and mixing method is therefore not suitable for higher-dimensional problems s > 3 . The Brownian bridge method has a similar convergence to the sorting and mixing method and can be used for evolution of single particles, but the method can converge to the wrong value for a large number of time steps with few particles and depends strongly on the measured quantity. The Brownian bridge method and the sorting and mixing method combined with scrambled Faure are very efficient within their working domain.

Please wait… references are loading.
10.1088/1749-4699/5/1/014010