1 Introduction
1.1 The Fukushima Daiichi nuclear accident
On 11 March 2011, an earthquake under the Pacific Ocean off the coast of Japan triggered an extremely destructive tsunami that hit the Japanese coastline about 1 h later, killing 18 000 people. These events led to the automatic shutdown of four Japanese nuclear power plants. However, the submersion of emergency power generators prevented the cooling system from functioning, making the shutdown impossible at the Fukushima Daiichi nuclear power plant (NPP). This yielded a massive release of radionuclides, characterized by several episodes of varying intensity, that lasted for several weeks. The estimation of such a release is difficult and subject to significant uncertainties. Reconstructing the temporal evolution of the caesium 137 (Cs) source requires one to establish a highly variable release rate that runs over several hundreds of hours.
Since 2011, several approaches have been proposed to assess the Fukushima Daiichi radionuclide release source terms. For instance, methods based on a simple comparison between observations and simulated predictions have been investigated by , , , , , , , and , and ambitious inverse problem methods have been developed in studies such as , , and . In particular, used air concentration and deposition measurements to assess the source term of Cs by rigorously estimating the errors, and evaluated the source term by inverse modelling with ambient gamma dose measurements. Finally, inverse problem methods based on a full Bayesian formalism have also been used. applied several methods, including Markov chain Monte Carlo (MCMC) algorithms, to estimate the time evolution of the Cs release, and they accompanied this with an estimate of the associated uncertainty. refined the Cs source term with an optimization method based on a Bayesian inference from several measurement databases.
1.2 Bayesian inverse modelling and sampling of a radionuclide source
Bayesian inverse problem methods have been shown to be effective in estimating radionuclide sources . Here, we briefly describe this Bayesian framework for source term reconstruction; a more complete description is available in . The crucial elements defining the control variable vector of the source are (i) , a vector whose components correspond to the logarithm of the release rates (constant releases at a time interval, such as 1 d or 1 h), and (ii) hyperparameters such as entries of the observation error scale matrix or the time windows of the release rates (more details in Sect. ).
The posterior probability density function (pdf) of is given by Bayes' rule:
1 where is the observation vector. The pdf is the likelihood, which quantifies the fit of the source vector to the data. More precisely, the observations are compared to a set of modelled concentrations: (i.e. the predictions, constructed using a numerical simulation of the radionuclide transport from the source). is the observation operator, the matrix representing the integration in time (i.e. resolvent) of the atmospheric transport model. Therefore, the predictions of the model are considered linear in . The likelihood is often chosen as a distribution parameterized with an observation error scale matrix . The pdf represents the information available on the source before data assimilation.
Once the likelihood and the prior are defined, the posterior distribution can be estimated using a sampling algorithm. Sampling techniques include, in particular, the very popular Markov chain Monte Carlo (MCMC) methods that assess the posterior pdf of the source, or marginal pdfs thereof. This method has been applied by to estimate the Algeciras incident source location, by , and by , who evaluated Xenon-133 releases at Chalk River Laboratories. More recently, the technique has been used to assess the uncertainties in the source reconstruction by , , and .
1.3 The transdimensional analysis and objectives of this studyThe problem of finding a proper representation of (i.e. a discretization of the source term) is the problem of finding the adequate step function for , i.e. the optimal number of steps, the optimal time length of each step, and the corresponding release rates. This representation depends on the case under study: depending on the observations, more or less information is available to define a more or less temporally resolved source term. In other words, depending on the data available to inform one about the source term, the steps of the function supporting could be small (a lot of available information, allowing for a fine representation) or large (little information available, providing a coarse representation), yielding an irregular temporal discretization of the source term. Note that the choice of the discretization can be seen as a balancing issue due to the bias–variance trade-off principle : low-complexity models are prone to high bias but low variance, whereas high-complexity models are prone to low bias but high variance. The difficulty in the source term discretization is that of selecting representations that are sufficiently rich but not overly complex in order to avoid over-fitting that leads to high variance error and excessive computing time. Radionuclides releases such as those of Cs from the Fukushima Daiichi NPP are very significant over a long period of time (several weeks) and have a very high temporal variability. In such case, the choice of the discretization is a crucial and challenging task.
In this paper, we explore several ways of improving the source term assessment and its corresponding uncertainty. In the first instance, we address the issue of sampling in the case of massive and highly fluctuating releases of radionuclides to the atmosphere. In the second instance, we propose several statistical models of the likelihood scale matrix depending on the available set of observations.
The outline of the paper is as follows. In Sect. , the observational dataset used in this study and the physical model of the problem are described. We then outline the theoretical aspects of this study and first focus on the concepts of transdimensionality and model selection (Sect. ). The reversible-jump MCMC algorithm (RJ-MCMC) used to shape the release rates vector is presented in Sect. . Other ways to improve the sampling quality and reduce uncertainties are then explored via the addition of information in Sect. . In particular, methods that combine deposition and concentrations observations or that exploit the temporal and spatial information of the observations in the definition of the likelihood scale matrix are proposed. Subsequently, these methods are evaluated in Sect. . Specifically, once the statistical parametrization of the problem is chosen in Sect. , the advantages of the RJ-MCMC are explored in Sect. . The impact of using the deposition measurements and the use of the observation temporal and spatial information are described in Sect. and . We finally conclude on the contribution of each method.
This article is in line with the authors' earlier studies and is inspired by the insightful work of . In particular, the RJ-MCMC algorithm is applied to the Fukushima Daiichi case by using variable discretizations but a fixed number of steps (i.e. grid cells). Here, it is applied in a more ambitious transdimensional framework, where the number of steps of the discretization is modified through the progression of the RJ-MCMC, and using a significantly larger dataset.
2
The Fukushima Daiichi NPP Cs release and numerical modelling
2.1 Observational datasetThe dataset used in this study consists of Cs air concentration and deposition measurements over the Japanese territory. The vector of Cs air concentration measurements contains observations from stations, whose locations are shown in Fig. . This dataset is significantly larger than the one used by .
Figure 1
Maximum air concentrations of Cs (in Bq m) measured in Japan in March 2011. The light blue points correspond to concentrations below the detection limit. The purple triangle corresponds to the Fukushima Daiichi NPP.
[Figure omitted. See PDF]
This allows us to better evaluate the relevance of our methods and to accurately estimate the uncertainties in the problem. Moreover, the majority of these observations are hourly, which allows for fine sampling of the Cs source term. Additionally, deposition measurements are used. This dataset is described and made available in . The observations are located within of the Fukushima Daiichi NPP; however, measurements too close to the Fukushima Daiichi plant (typically less than 5 times the spatial resolution of the meteorological fields described in Table ) are not used to consider the short-range dilution effects of the transport model. These measurements represent the averages of each grid cell of a field of deposition observations (in Bq m) measured by aerial radiation monitoring.
2.2 Physical parametrizationThe atmospheric Cs plume dispersion is simulated with the Eulerian model ldX, which has already been validated on the Chernobyl and Fukushima Daiichi accidents as well as on the Ru releases in 2017 . The meteorological data used are the 3-hourly high-resolution forecast fields provided by the European Centre for Medium-Range Weather Forecasts (ECMWF) (see Table ).
Table 1
Main set-up parameterizations of the ldX Cs transport simulations. The spatial and temporal resolutions are those of the ECMWF meteorological fields and the transport model. The other entries correspond to parameters of the transport model only.
ECMWF | |
---|---|
Spatial resolution | |
Temporal resolution | 3 h |
Vertical resolution | 11 terrain-following layers (from 0 to 4400 m) |
Vertical mixing | K-diffusion parameterized with Louis' closure |
and under unstable planetary boundary layer conditions | |
Horizontal mixing | Constant horizontal eddy diffusion coefficient, |
Wet scavenging | (below-cloud), with ; |
(in-cloud), with ; | |
and is the rainfall intensity | |
Dry deposition | Constant deposition velocity, |
The radionuclides, in particular Cs, were mainly deposited in the Fukushima Prefecture, north-west of the Fukushima Daiichi NPP. The bulk of the releases occurred over a 2-week period beginning on 11 March. Therefore, the model simulations start on 11 March 2011 at 00:00 UTC and end on 24 March 2011, corresponding to 312 h during which radionuclides could have been released. The release is assumed to be spread over the first two vertical layers of the model (between 0 and 160 m). The height of the release has a small impact, as the modelled predictions are compared with observations sufficiently far away from the Fukushima Daiichi NPP and the release mostly remains within the atmospheric boundary layer. Finally, it is assumed that all parameters describing the source aside from the reconstructed ones are known. This is, for instance, the case for the release height.
3 Methodological aspects of the inverse problem3.1 Reversible-jump MCMC
In this section, we describe the Bayesian reversible-jump MCMC algorithm used in the following to reconstruct the vector discretized over a variable adaptive grid. A good description of the RJ-MCMC, which was introduced by , can be found in . The algorithm is also clearly explained and applied by . In particular, the method was used by in the field of inverse modelling for the assessment of substance releases as well as by to quantify Fukushima Daiichi releases, although with a fixed number of variable-sized grid cells. A key asset of the RJ-MCMC technique is its ability to sample a highly variable release by balancing bias and variance errors.
The RJ-MCMC algorithm is a natural extension of the traditional Metropolis–Hastings (MH) algorithm to transdimensional discretization grids. In a traditional MH, the distribution of the release is assessed via a succession of random walks
Indeed, as described above, the function that describes the evolution of the release rate over time is a vector or step function, where each step (i.e. a grid cell) defines a constant release rate over a time interval. When using the RJ-MCMC, these time intervals are neither fixed nor regularly spaced. Hence, an objective is to retrieve the best stepwise time partition of the source term, which defines the release rates.
To do this, we use the concept of boundary: a boundary is a time that separates two constant release rates (i.e. that separates two grid cells where the source term is defined).
Figure 2
Example of a source term partition: ) is the set of boundaries, and is the vector of release rates. The release rate defined on the time interval is constant and equal to .
[Figure omitted. See PDF]
Let be the set of boundaries. Boundaries are , , and integers in ( and are natural and fixed boundaries), i.e. the boundaries are a selection among the hours of the release. Finding the probability distribution of the time steps associated with is equivalent to finding the probability distribution of . To retrieve the best discretizations of , we can therefore sample . Figure presents an example of a source term partition.
The RJ-MCMC is basically equivalent to an MH algorithm except that some iterations of the algorithm are not intra-model random jumps (i.e. at constant ) but rather possibly inter-model random jumps (i.e. at non-constant ) or transdimensional (i.e. non-constant size).
Specifically, we use three types of random jumps that govern the positions and number of boundaries: birth, death, and move. For these three processes, we draw on the work of , , , and (importantly) . The birth process is the creation of a new boundary not yet present in , the death process is the removal of an existing boundary in , and the move process is the displacement of an existing boundary in . These processes must respect the detailed balance criterion in order to ensure the convergence of the Markov chain to the target distribution .
3.2 Handling observations and their errors3.2.1 Measurements fusion
In this subsection, we propose a method to factor both concentration and deposition measurements into a Bayesian sampling by simply taking their disparity in the observation error matrix into account. First of all, is diagonal and is, therefore, described by a collection of scale parameters. Specifically, two scale parameters, and , are used when modelling and are associated with the air concentration and deposition measurements respectively. Thus we can write the following:
2 where is the diagonal matrix of diagonal entries . Note that we use the sorting algorithm described by , which is applied to both the air concentration and deposition measurements, in all subsequent applications. This algorithm assigns a different scale parameter to certain non-pertinent observations to prevent them from artificially reducing and . With this definition of , the scale variables , and are included in the source vector and, therefore, sampled using a MCMC.
3.2.2 Spatio-temporal observation error scale matrix modellingIn this subsection, we propose modelling using the spatio-temporal information of the concentration observations. We assume that the error scale parameter corresponding to an observation–prediction pair depends mainly on the error in the model predictions. We also propose considering a prediction (i.e. an estimation of the number of radionuclides present at a definite time and place) as a point in a radionuclide plume at a specific time. The prediction modelling error varies according to its spatio-temporal position. Spatially, if two points of interest are distant by , the difference in the errors in their attached model prediction is empirically assumed to be proportional to . Temporally, let us consider two points in a plume with coordinates at times and respectively, where and are the longitude and latitude. These two points are moving with the plume, and they are spatially distant by at time , where is a reference wind speed, representing the average wind over the accidental period. Therefore, the difference in the modelling errors of these points is estimated to be proportional to . In that respect, for each concentration observation, can be associated with three coordinates:
3 where , are the coordinates (in km) of the observation, is set equal to , and and are the starting time and the duration of the observation in hours respectively. This average wind speed value has not been extensively researched and is only used to estimate the potential of the method. A clustering algorithm such as the -means algorithm can be applied to partition these three coordinates' observation characteristics into several groups. A scale parameter can then be assigned to each group. If the algorithm for sorting pertinent and non-pertinent observations is applied, this yields 4 where , and are the scale parameters associated with the spatial or spatio-temporal clusters of air concentration observations, the deposition measurements, and the non-pertinent observations respectively. The case where only the spatial qualification of the observation is used instead of both the spatial and temporal qualification is also addressed in this study.
4Application to the Fukushima Daiichi NPP Cs release
In this section, we apply the previously introduced methods to the reconstruction of the Fukushima Daiichi Cs source term between 11 and 24 March. However, we first define the key statistical assumptions and parameters of the inversion set-up.
4.1 Statistical parametrizationThe prior distributions determining the posterior distribution and the transition probabilities of the RJ-MCMC algorithm are defined in the next two sections.
4.1.1 Definition of the prior density functions
In this section, we specify the prior distributions of the source vector components: release rates and hyperparameters. The source release rate logarithms are assumed to follow a folded Gaussian prior distribution, so as to concurrently constrain the number of boundaries and the magnitude of the release rates. In the general case with , the scale matrix of the folded Gaussian prior is defined as a diagonal matrix. The entry of is associated with the th hour in , comprising hours where we consider the release of Cs possible. This yields the following:
5 where the folding position is defined as equal to the location term . We divide the 312 time intervals into two groups. The first “unconstrained” group consists of the time intervals during which we have noticed that a radionuclide emission does not trigger any noticeable observation. In other words, it is a time interval during which we have no information on a potential release. The unconstrained time intervals are the first 17 h of 11 March (hours in ), the days of 16 and 17 March (hours in ), and the last 18 h of 23 March (hours in ). The second group consists of the constrained time intervals. Two hyperparameters – for the constrained time intervals and for the unconstrained time intervals – are used to describe the matrix . If, instead of two, a single hyperparameter is used for all release rates, the unconstrained time intervals are neither regularized by the observations nor by the prior. Indeed, to account for high release rates, samples of the distribution of reach significant values. Low-magnitude release rates are then constrained very little by the corresponding prior. Release rates that are neither constrained by the observations nor by the priors can, in turn, reach very large values. This phenomenon was put forward by . The mean of the folded Gaussian prior is chosen as in order to regularize the unconstrained intervals. It is important to note that a release rate sampled by the RJ-MCMC is an aggregate of these hourly release rates, with each of them associated with either or . Hence, we define the scale of the prior associated with the release rate defined between times and as follows: 6 where is the th diagonal coefficient of , is the portion of constrained time intervals in , and and are the weighted numbers of the respective constrained and unconstrained hourly release rates included in the release rate . This definition is reflected in the prior cost: where is the number of time steps ( being the number of boundaries) considered at a certain iteration by the RJ-MCMC and which characterizes the grid on which is defined. The prior scale parameter is included in the source vector and sampled with the rest of the variables. Such a prior on the release rate logarithms also has a constraining effect on the model's complexity (Occam's razor principle) through the normalization constant of the prior pdf.
An exponential prior distribution is used for the boundaries : 9 Here, . This prior has the effect of penalizing models that are too complex.
Furthermore, because no a priori information is available, the prior distributions on the coefficients of the observation error scale matrix are assumed to be uniform, and the prior values on the regularization scale terms are assumed to follow Jeffreys' prior distribution . The lower and upper bounds of the uniform prior on the coefficients of the observation error matrix are set as zero and a large value, the latter of which is not necessarily realistic, as in .
4.1.2 Parameters of the MCMC algorithmIn the Markov chain, the variables describing the source are initialized randomly. The intra-model (i.e. with a fixed boundary partition) transitions defining the stochastic walks are set independently for each variable. Each transition probability is defined as a folded normal distribution, following . Locations of the transition probabilities are the values of the variables at the current step. Variances of the related normal distribution are initialized at and , with the values being chosen by means of experiments. However, the value of is adaptive and is updated every iterations according to the predictions' values. Furthermore, the value of is not uniform across groups of observations: there are as many variances as observation clusters.
The algorithm runs initially as a classic MH for iterations (only intra-model walks are allowed at this stage). Indeed, allowing transdimensional walks at the beginning of the run sometimes cause Markov chains to get stuck in local minima. In total, iterations are used; this large number ensures algorithm convergence (i.e. sufficient sampling of the posterior distribution of ) and corresponds to approximately 1 d of calculation for a -core computer. The burn-in is set to iterations.
The inter-model transdimensional walks (with changing boundaries) are described in Appendix , as they are technical. For this transdimensional case, is the same for birth and death processes and is equal to . There is no variance parameter for the move walk. The probability of proposing an intra-model transition, a birth transition, a death transition and a move transition is , , , and respectively.
4.2 Results
To evaluate the quality of our reconstructed source term as well as the impact of the techniques proposed in Sect. , the distributions of the Cs source are sampled and compared using several settings:
-
Sect. provides an activity-concentration-based reconstruction of the Cs source with a log-Cauchy likelihood and a threshold (to ensure it is defined for zero observations or predictions) of Bq m , the application of the observation sorting algorithm, and a spatial representation of with 10 parameters. The vector
10is sampled. A log-Cauchy likelihood is chosen because the associated reconstruction produces the best model-to-data comparisons among diverse choices of likelihoods and representations of . The sampled source is compared to the source terms of and , and all variable marginal distributions are described.
-
In Sect. , we are interested in quantifying the impact of the addition of deposition measurements in the sampling with the method presented in Sect. . The source term is sampled in two configurations, namely with or without the addition of deposition measurements. That is, the vector11is sampled with or without . A threshold of Bq m is chosen for the deposition measurements.
-
Finally, we draw a selection of violin plots in Sect. to assess the impact of the distribution of the observations in space and time clusters using the approach introduced in Sect. . More precisely, the fit to the observations of the reconstructed source is evaluated given the type of clustering (spatial or spatio-temporal) and the number of clusters.
In this section, we provide the most appropriate reconstruction of the source term, based on the activity concentration measurements, from several choices of likelihoods and representations. This best reconstruction is chosen with the help of a model-to-data comparison, i.e. FAC (factor) scores, representing the proportion of measurements for which the observed and modelled values agree. The chosen likelihood is a log-Cauchy distribution with a threshold of Bq m, and air concentration measurements are spatially clustered in nine groups using the -means method.
Figure shows the temporal evolution of the median release rate with its associated standard deviation (i.e. the evolution of the hourly release rates samples that are obtained from the RJ-MCMC samples). This temporal evolution is compared to the source term from , which was estimated from gamma dose rate measurements but from the same meteorological data and transport model, and to the more recent source term from .
Figure 3
Evolution of the release rate (in Bq s) describing the Cs source (UTC) sampled with the RJ-MCMC algorithm. The blue line corresponds to the sampled median release rate. The light blue area corresponds to the area between , the median, and , the standard deviation, of the reconstructed hourly release rates: . Our source term is compared to the source terms of and .
[Figure omitted. See PDF]
Firstly, we observe that a good match is obtained between the median release and the source term of or (on the constrained time intervals), as the differences between releases under Bq s can be neglected. Indeed, between the 11 and the 19 March 2011, the three source terms have similar peak magnitudes (close to Bq s); these peaks are observed at neighbouring time intervals. Between the 19 and the 21 March, a notable difference can be observed: the RJ-MCMC source term predicts very high 1 h peaks (close to Bq s), whereas the source term distributes the release uniformly over the whole period. Secondly, release rates are subject to large variations. Indeed, there are periods of important temporal variability in the release rate (e.g. between 19 and 21 March) and of low temporal variability (e.g. between 11 and 14 March). This justifies the use of the RJ-MCMC transdimensional algorithm, as it allows one to reconstruct certain time intervals of the release with a greater accuracy and others with less complexity. Thirdly, the temporal evolution of the release can be explored, and several prominent peaks can be observed with a magnitude approaching or surpassing Bq s: between 12 and 14 March, between 14 and 16 March, on 18 March, and between 19 and 21 March.
Figure 4
Densities or averages of variables describing the Cs source reconstructed with the RJ-MCMC technique: (a) density of relative differences in log between observations and predictions; (b) density of the total retrieved released activity (TRRA) of Cs during the Fukushima Daiichi accident (in Bq) – outliers are removed; (c) histogram of the number of boundaries; and (d) mean ( standard deviation) of the number of boundaries as a function of time (UTC).
[Figure omitted. See PDF]
Figure a describes the histogram of relative differences in log between predictions (medians of the predictions computed using the samples are employed) and air concentration observations larger than Bq m. A value of , therefore, corresponds to an observation 10 times larger than a prediction. A good fit between observations and predictions can be observed, although the observations are globally underestimated by the predictions.
Figure b shows the pdf of the total release (in Bq). It can be seen from this density that most of the total release distribution mass is sampled between and PBq and peaks at PBq. This is consistent with previous studies which estimate the release to be between and PBq . However, note that these works attempted to estimate potential releases beyond 24 March.
Figure c plots the pdf of the number of boundaries (i.e. a measure of model complexity) sampled by the algorithm. Here, the minimum of the distribution is 25 boundaries and the maximum is 45 with a peak at 35 boundaries. Recall that the maximum number of boundaries is 313.
Figure d shows the average number of sampled boundaries at each time and the associated standard deviation. The axis corresponds to the number of boundaries counted: for a given hour , we draw 12 i.e. the average over all samples of the number of boundaries counted around . In essence, the curve shows the number of boundaries necessary to model the release at a certain time (i.e. the variability in the release according to the algorithm). We observe that there is a large variability between 14 and 15 March as well as between 19 and 21 March. This correlates very well with the observable results in Fig. . This observation once again confirms the relevance of using the RJ-MCMC algorithm: periods of high variability can be sampled finely.
4.2.2 Impact of using deposition measurementsIn this section, we study the impact of adding deposition measurements to the air concentration measurements. A log-Cauchy distribution with a threshold of Bq m for the air concentration measurements and Bq m for the deposit measurements is employed. Air concentration measurements are clustered with a spatial -means algorithm in nine groups, as described in Sect. . To quantify the impact of the added measurements, we sample the following vector with an RJ-MCMC algorithm in two cases, with or without deposition measurements, i.e. with or without , as follows:
13 In the case without deposition measurements, we use the same samples as in Sect. . A posteriori marginals of these control variables are displayed in Fig. .
Figure 5
The panels correspond to probability densities or averages of variables describing the Cs source reconstructed with the RJ-MCMC technique using air concentration and deposition measurements. Panel (a) displays the medians and standard deviations of hourly release rates. The red and blue curves correspond to the sampled median release rates. The red and light blue areas correspond to the areas between , the median, and , the standard deviation, of the sampled hourly release rates for each reconstruction: . Panel (b) shows the probability density of the total release of Cs during the Fukushima Daiichi accident (in Bq). Panel (c) provides a histogram of the number of boundaries. Panel (d) shows the mean ( standard deviation) of the number of boundaries according to the hour considered.
[Figure omitted. See PDF]
In Fig. a, the medians of the two sampling datasets are similar, except for the release rates between 11 and 14 March. Indeed, when using air concentration and deposition measurements, the release rate peaks on 12 March instead of on 13 March, the latter of which is the case when using air concentrations alone.
In Fig. b, the main mode of the total retrieved released activity (TRRA) distribution retrieved using the air concentration and deposition measurements is PBq larger than the TRRA reconstructed with the air concentration measurements alone. Moreover, the reconstructed density with both types of measurements has a large variance and leads to the possibility of a larger release (up to PBq). The density of the number of boundaries with both types of measurements is shifted to higher values compared with the density with air concentrations alone, which indicates a reconstructed release of greater complexity (Fig. c). This is consistent with the fact that the second release is reconstructed with a larger number of observations.
In Fig. d, the number of boundaries was estimated to be half as large on 14 March when deposition measurements were used, but it was estimated to be larger on 15 March. These differences are rather surprising insofar as there is no obvious disparity between the medians of the release rates for the two reconstructions in Fig. a on these days.
In Table , we evaluate the FAC2, FAC5, and FAC10 scores for both cases. FAC scores are common metrics for measuring the fit of a source term with available observations. The FAC2, FAC5, and FAC10 are the proportion of measurements for which the observed and modelled values agree by a factor of 2, 5, or 10 respectively. For example, a FAC2 score of with respect to the air concentration measurements means that of the predictions (i.e. simulated measurements) are such that 14 where is the number of observations considered, and is a threshold set at Bq m. This threshold is added to limit the weight of low values, considered of lesser interest, in the indicator. Finally, to calculate the FAC scores, only observations above Bq m are considered.
Table 2Comparison of observed and simulated measurements from reconstructed predictions with air concentration measurements only or with air concentration and deposition measurements. The FAC2, FAC5, and FAC10 are the proportion of measurements for which the observed and modelled values agree by a factor of 2, 5, or 10 respectively. For example, the table reads as follows: the predictions calculated based on air concentrations only obtain a FAC5 score of 0.58 with respect to the deposition measurements.
Origin of the source term | FAC2 | FAC5 | FAC10 |
---|---|---|---|
RJ-MCMC (based on air concentrations only) | |||
Air concentration | 0.32 | 0.68 | 0.85 |
Deposition | 0.31 | 0.58 | 0.75 |
RJ-MCMC (air concentration and deposition) | |||
Air concentration | 0.31 | 0.67 | 0.85 |
Deposition | 0.49 | 0.77 | 0.86 |
Source term of | |||
Air concentration | 0.18 | 0.52 | 0.76 |
Deposition | 0.28 | 0.64 | 0.74 |
Source term of | |||
Air concentration | 0.17 | 0.52 | 0.78 |
Deposition | 0.29 | 0.64 | 0.76 |
The FAC scores show that the source term obtained with the RJ-MCMC has a better agreement with the air concentration measurements than the or source terms. This shows that the RJ-MCMC is able to sample source terms consistent with air concentration measurements but does not allow us to reach a conclusion regarding its performance: the sampling algorithm was optimized using air concentrations, whereas, for example, the source term was recovered from dose rate measurements. However, the source term found with the RJ-MCMC on air concentration measurements can be fairly evaluated using deposition measurements. It is observed that the fits with the deposition measurements of the three source terms are similar.
Furthermore, as expected, Table 2 shows that there is a significant difference in the fit with the deposition measurements when they are added to the observation dataset used to reconstruct the source term. However, the addition of the deposition measurements has no impact on the quality of the fit to the air concentration measurements. This suggests that an RJ-MCMC optimized on dose rate measurements, in addition to the deposition and air concentrations, may obtain a good agreement with the dose rate measurements, without degrading its results on deposition or air concentrations.
4.2.3Impact of the observation error matrix representation
In this section, we investigate the spatial and spatio-temporal representation of the likelihood scale matrix. Figure presents violin plots of the FAC2 and FAC5 scores for air concentration measurements against the number of scale parameters chosen to describe (i.e. the number of observations clusters). Violin plots are created out of the best scores of 44 RJ-MCMC sampling datasets with diverse likelihoods chosen after in order to make the conclusions less dependent on statistical modelling. The chosen numbers of spatial clusters (1, 2, and 9) and spatio-temporal clusters (3 and 7) of the violins are a selection for the sake of clarity.
Figure 6
Violin plots of the air concentration FAC2 and FAC5 scores against the observation error scale parameters' spatio-temporal groups number. Points composing the violin plots consist of the best FAC2 and FAC5 scores of 44 sampling datasets using diverse configurations of likelihoods and thresholds (log-normal, log-Laplace, and log-Cauchy) chosen according to . The horizontal segments within the violins represent (from bottom to top) the respective first, second, and third quartiles. The chosen numbers of spatial and spatio-temporal clusters are only a selection for the sake of clarity.
[Figure omitted. See PDF]
For both the FAC2 and FAC5 air concentration scores, it can be observed that the spatio-temporal labelling of the observations does not increase the quality of the scores, whereas the number of spatial clusters does. For example, we can observe a difference of between the second quartile of the FAC2 score for sampling datasets with two spatial groups and the second quartile of the FAC2 score for sampling datasets with nine spatial groups. One hypothesis is that the error is mainly related to the distance to the source and remains relatively homogeneous in time. However, the use of an accurate estimate of the effective wind speed as a function of time and depending on the position of the observations is necessary to draw conclusions; such a study is outside the scope of this paper.
5 Summary and conclusionsIn this paper, we investigated a transdimensional sampling method to reconstruct highly fluctuating radionuclide atmospheric sources and applied it to assess the Cs Fukushima Daiichi release. Furthermore, we proposed two more methods to add information to the model in order to reduce uncertainties attached to such a complex release.
Firstly, an RJ-MCMC algorithm was defined. This MH extension to transdimensional meshes allows one to sample both the source term and its underlying adaptive discretization in time. In the case of complex releases, an adaptive grid allows one to reconstruct the source term with high accuracy and discretization consistent with the available observations.
Secondly, we have focused on ways to add information by adapting the representation of the observation error matrix . We proposed the assimilation of deposition measurements or spatial and temporal information on the air concentration observations by increasing the complexity of .
Subsequently, the distributions of the variables defining the source of Cs were fully sampled. This enabled the estimation of the uncertainties associated with these variables as well as the evaluation and demonstration of the merits of the methods.
Firstly, we observed that the predictions reproduce the air concentration and deposition observations well. When using air concentrations only, the FAC2 and FAC5 scores corresponding to air concentration measurements are and respectively. The main part of the best total release distribution mass is estimated to be between and PBq and peaks at PBq, which is in accord with previous estimations.
Secondly, the periods of high variability in Cs releases have been reconstructed with accuracy by the RJ-MCMC algorithm. The transdimensional method also allowed the periods of low temporal variability to be sampled more coarsely, thereby avoiding variance errors and saving computing time, which, in particular, confirms the conclusions of . The priors that we chose allowed for an efficient regularization of the release rates and their adaptive mesh.
Finally, the use of combined air concentration and deposition measurements had an impact on the reconstruction of the TRRA distribution, the mean of which increased by PBq. The spatial clustering methods proved to increase the quality of the predictions by improving the model-to-data comparison.
We recommend the use of the RJ-MCMC for long-release source reconstruction. It allows one to sample highly variable source terms with great accuracy by solving the bias–variance trade-off. We also recommend the use of advanced modelling of considering the deposition measurements and the spatial information of the air concentration observations which have proven to be valuable in reducing the observation–prediction discrepancies.
Appendix A RJ-MCMC mathematical details
Here, we describe the mathematical details of the three inter-model walks: move, birth, and death. The move process is intra-dimensional. Provided that the move is a symmetric motion, the associated -transition distribution is symmetric, i.e. A1
The birth process is the creation of a new boundary, a walk from (a vector composed of boundaries) to (a vector composed of boundaries) is proposed. In other words, by generating a new boundary, one release rate is destroyed and two new release rates emerge from this destruction. We therefore assign two new release rates and , with (a Gaussian noise). The transition probability is defined as follows : A2 The probability of a birth at a certain position among positions not occupied by boundaries is A3 Furthermore, the probability of generating new release rates is Gaussian: A4 where the new boundary here is generated at hour .
The probability of death of one among the boundaries is A5 which is a uniform choice among possibilities. On the other hand, the probability of destroying the release associated with this position is .
Data availability
Each dataset used in the paper is available via a DOI on a compliant repository or directly in the cited papers. More precisely, the air concentration observation datasets used in the inversions are available in from
Author contributions
JDLB: software, methodology, conceptualization, investigation, writing – original draft, and visualization. MB: methodology, software, conceptualization, writing – review and editing, visualization, and supervision. OS: resources, methodology, conceptualization, writing – review and editing, visualization, and supervision. YR: methodology, software, and writing – review and editing.
Competing interests
The contact author has declared that none of the authors has any competing interests.
Disclaimer
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Acknowledgements
The authors would like to thank the European Centre for Medium-Range Weather Forecasts (ECMWF) for making available the meteorological fields used in this study. They would also like to thank Sophie Ricci, Lionel Soulhac, Paola Cinnella, Didier Lucor, Yann Richet, and Anne Mathieu for their suggestions and advice on this work. CEREA is a member of the Institut Pierre-Simon Laplace (IPSL).
Review statement
This paper was edited by Dan Lu and reviewed by two anonymous referees.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023. This work is published under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The accident at the Fukushima Daiichi nuclear power plant (NPP) yielded massive and rapidly varying atmospheric radionuclide releases. The assessment of these releases and of the corresponding uncertainties can be performed using inverse modelling methods that combine an atmospheric transport model with a set of observations and have proven to be very effective for this type of problem. In the case of the Fukushima Daiichi NPP, a Bayesian inversion is particularly suitable because it allows errors to be modelled rigorously and a large number of observations of different natures to be assimilated at the same time. More specifically, one of the major sources of uncertainty in the source assessment of the Fukushima Daiichi NPP releases stems from the temporal representation of the source. To obtain a well-time-resolved estimate, we implement a sampling algorithm within a Bayesian framework – the reversible-jump Markov chain Monte Carlo – in order to retrieve the distributions of the magnitude of the Fukushima Daiichi NPP caesium 137 (
These methods are applied to the reconstruction of the posterior distributions of the magnitude and temporal evolution of the
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 IRSN, PSE-SANTE, SESUC, BMCA, Fontenay-aux-Roses, France; CEREA, École des Ponts and EDF R&D, Île-de-France, France
2 CEREA, École des Ponts and EDF R&D, Île-de-France, France
3 IRSN, PSE-SANTE, SESUC, BMCA, Fontenay-aux-Roses, France