1 Introduction
The global ocean is a major sink of excess that has been emitted to the atmosphere since the beginning of the industrial revolution. In 2011, the best estimate of the ocean inventory of anthropogenic carbon () amounted to PgC or 28 % of the cumulated total emissions attributed to human activities since 1750 (Ciais et al., 2013). Between 2000 and 2009, the yearly average ocean uptake was PgC yr (Ciais et al., 2013). However, these global estimates hide substantial regional and interannual fluctuations (Rödenbeck et al., 2015), which need to be quantified in order to track the evolution of the Earth's carbon budget (e.g., Le Quéré et al., 2018).
Until recently, most estimates of interannual sea–air flux variability were based on atmospheric inversions (Peylin et al., 2005, 2013; Rödenbeck et al., 2005) or global ocean circulation models (Orr et al., 2001; Aumont and Bopp, 2006; Le Quéré et al., 2010). However, models tend to underestimate the variability of sea–air fluxes (Le Quéré et al., 2003), whereas atmospheric inversions suffer from a sparse network of atmospheric measurements (Peylin et al., 2013). These approaches are increasingly complemented by data-based techniques relying on in situ measurements of fugacity or partial pressure (e.g., Takahashi et al., 2002, 2009; Nakaoka et al., 2013; Schuster et al., 2013; Landschützer et al., 2013, 2016; Rödenbeck et al., 2014, 2015; Bitting et al., 2018; Fay et al., 2014). These techniques rely on a variety of data-interpolation approaches developed to provide estimates in time and space of surface ocean (Rödenbeck et al., 2015) such as statistical interpolation, linear and nonlinear regressions, or model-based regressions or tuning (Rödenbeck et al., 2014, 2015). These methods, and their advantages and disadvantages, are compared and discussed in Rödenbeck et al. (2015). This intercomparison did not allow for the identification of a single optimal technique but rather pleaded in favor of exploiting the ensemble of methods.
Artificial neural networks (ANNs) have been widely used to reconstruct surface ocean ; Lefèvre et al.(2005), Friedrich and Oschlies (2009b), Telszewski et al. (2009), Landschützer et al. (2013) Nakaoka et al. (2013), Zeng et al. (2014) and Bitting et al. (2018) used ANNs to study the open ocean, whereas Laruelle et al. (2017) studied the coastal region using this method. ANNs fill the spatial and temporal gaps based on calibrated nonlinear statistical relationships between and its oceanic and atmospheric drivers. The existing products usually present monthly fields with a spatial resolution and capture a large part of the temporal–spatial variability. Methods based on ANNs are able to represent the relationships between and a variety of predictor combinations, but they are sensitive to the number of data used in the training algorithm and can generate artificial variability in regions with sparse data coverage (Bishop, 2006).
This study proposes an alternative implementation of a neural network applied to the reconstruction of surface ocean over the period from 2001 to 2016. It belongs to the category of feed-forward neural networks (FFNN) and consists of a two-step approach: (1) the reconstruction of monthly climatologies of global surface ocean based on data from Takahashi et al. (2009), and (2) the reconstruction of monthly anomalies (with respect to the monthly climatologies) on a grid exploiting the Surface Ocean Atlas (SOCAT) (Bakker et al., 2016). The model is easily applied to the global ocean without any boundaries between the ocean basins or regions. However, as previously mentioned, it is still sensitive to the observational coverage. This limitation is partly overcome by the two-step approach, as the reconstruction of monthly climatologies draws on a global ocean gridded climatology (Takahashi et al., 2009); therefore, the FFNN output is kept close to realistic values. Furthermore, the reconstruction of monthly climatologies during the first step allows for a potential change in seasonal cycle in response to climate change to be taken into account when applied to time slices or to model output providing the drivers, but no carbon cycle variables.
The remainder of this paper is structured as follows: Sect. 2 introduces datasets used during this study and describes the neural network; Sect. 3 presents results for its validation and qualification, as well as its comparison to three mapping methods as part of the Surface Ocean Mapping intercomparison (SOCOM) exercise (Rödenbeck et al., 2015). Finally, results and perspectives are summarized in the last section.
2 Data and methods
2.1 Data
The standard set of variables known to represent the physical, chemical and biological drivers of surface ocean – mean state and variability – (Takahashi et al., 2009; Landschützer et al., 2013) were used as input variables (or predictors) for training the FFNN algorithm. These are sea surface salinity (SSS), sea surface temperature (SST), mixed layer depth (MLD), chlorophyll concentration (Chl ) and the atmospheric mole fraction (). Based on Rodgers et al. (2009) who reported a strong correlation between natural variations in dissolved inorganic carbon (DIC) and sea surface height (SSH), SSH was added as a new driver to this list. Initial tests suggested that the inclusion of the SSH did not significantly improve the accuracy of reconstructed at the global scale. At the basin and regional scale, however, adding the SSH improved the spatial pattern of reconstructed and the accuracy of our method.
For the first step, the reconstruction of monthly climatologies, the Takahashi et al. (2009) monthly gridded climatology () was used as the target. The original climatology was constructed by an advection-based interpolation method on a grid. It was interpolated on the SOCAT grid, which is also the resolution of the final output for the FFNN.
For the second step, the target was provided by the SOCAT v5 observational database (Bakker et al., 2016). We used a gridded version of this dataset that was derived by combining all SOCAT data collected within a box during a specific month. SOCAT v5 represents global observations of the sea surface fugacity of () over the period from 1970 to 2016. It includes data from moorings, ships and drifters. These data are irregularly distributed over the global ocean with 188 274 gridded measurements over the Northern Hemisphere and 76 065 over the Southern Hemisphere. In order to ensure a satisfying spatial and temporal data coverage, we limited the reconstruction to the period from 2001 to 2016, which represents % of the database (Fig. 1a).
Figure 1
Spatial distribution of SOCAT data (number of measurements per grid point): (a) for the 2001–2016 period; (b) for all months of January for the 2001–2016 period; (c) for all months of December–January–February for the 2001–2016 period. Please note that the color bar under panel (c) refers to both (b) and (c).
[Figure omitted. See PDF]
c The following formula is used to convert to
(Körtzinger, 1999):
1
where and are in
micro-atmospheres (), is the total pressure (Pa), J K is the gas constant and is the absolute temperature
(K). Parameter (m mol) is estimated as . The parameter is the cross-virial coefficient
(m mol): .
The total pressure is from the Jena database (6 h, ;
Monthly global reprocessed products of physical variables from ARMOR3D L4
distributed through the Copernicus Marine Environment Monitoring Service
(CMEMS; ;
MLD and CHL data were log-transformed before their use in the FFNN algorithm because of their skewed distribution. In regions with no CHL data (high latitudes in winter) log(CHL) was applied. It does not introduce discontinuities, as log(CHL) is close to zero in the adjacent region.
All data were averaged or interpolated on a grid and, depending on the resolution of the dataset, averaged over the month. It is worth noting that all datasets have to be normalized (i.e., centered to zero-mean and reduced to unit standard deviation) before their use in the FFNN algorithm, for example, Normalization ensures that all predictors fall within a comparable range and avoids giving more weight to predictors with large variability ranges (Kallache et al., 2011).
As surface ocean also varies spatially, geographical positions, latitude (lat) and longitude (long), after conversion to radians were included as predictors. In order to normalize (lat, long) the following transformation is proposed: Two functions sin and cos for longitudes are used to preserve its periodical 0 to 360 behavior and thus to consider the difference of positions before and after the 0 longitude. For step 2, data required for training were colocated at the SOCAT data positions that are used as a target for the FFNN model. Details are provided in the next section.
2.2 Methods2.2.1 Network configuration and evaluation protocol
In this work, we use Keras, a high-level neural network Python library
(“Keras: the Python Deep Learning library”, Chollet, 2015;
The number of FFNN layers and the number of neurons depends on the complexity of the problem: the more layers and neurons, the better the accuracy of the output. However, the size also depends on the number of patterns (data) used for training. The empirical rule recommends having a factor of 10 between the number of patterns (data) and the number of connections, or weights to adjust (in line with Amari et al., 1997, we use a factor of 10 that necessitates a cross-validation to avoid overfitting). This limits the size, the number of parameters and incidentally the number of neurons of the FFNN. This empirical rule was followed in this study.
-
Step 1: reconstruction of monthly climatologies. FFNN reconstructs a normalized monthly surface ocean climatology as a nonlinear function of normalized SSS, SST, SSH, Chl , MLD climatologies and geographical position (lat and long): Surface ocean from Takahashi et al. (2009) provided the target. The dataset was divided into 50 % for FFNN training and 25 % for its evaluation. This 25 % did not participate in the training. This set is used to monitor the performance of the training process and to drive its convergence. The remaining 25 % (each fourth point) of the dataset was used after training for the FFNN model validation. More details about the FFNN training process can be found in Rumelhart et al. (1986) and Bishop (1995). Validation and evaluation datasets were chosen quasi-regularly in space and time to take all regions and seasonal variability into account. In order to improve the accuracy of the reconstruction, the model was applied separately for each month. We have developed a FFNN model with five layers (three hidden layers). Twelve models with a common architecture were trained. Tests with one model for 12 months showed a slight decrease in accuracy (not presented here). About 17 500 data were available for each month to train the model, resulting in monthly FFNN models with about 1856 parameters.
-
Step 2: reconstruction of anomalies. During the second step, normalized anomalies were reconstructed as a nonlinear function of normalized SSS, SST, SSH, Chl , MLD, and their anomalies, as well as geographic position: Surface ocean anomalies were computed as the differences between colocated values based on SOCAT observations and monthly climatologies reconstructed during the first step that provided the targets:
4 The set of target data was again divided into 50 % for the training algorithm, 25 % for evaluation and 25 % for model validation. As in step (1) the model was trained separately for each climatological month. Thus, there were 12 models sharing a common architecture but trained on different data. In this step, in order to increase the amount of data during training and introduce information on the seasonal cycle, the model was trained using data from the month in question as a target as well as data from the previous and following month during the entire period from 2001 to 2016. Figure 1b and c show an example of the data distribution for the months of January over the period from 2001 to 2016 (Fig. 1b) and for the 3-month time window including December–January–February from 2001 to 2016 used in the training algorithm of the January FFNN model (Fig. 1c). In this particular example, the choice of 3 months provided a better cover of the region and doubled the number of data at high latitudes.
-fold cross-validation was used for the evaluation and the validation of the FFNN architecture. Cross-validation relied on different subsamples of the dataset to draw 25 % of independent data for validation (Fig. S1 in the Supplement). Each sampling fold was tested on five runs of the FFNN for each month. Each of these five runs was characterized by different initial values that were randomly chosen. From these five results, the best was chosen based on root-mean-square error (RMSE), the and the bias.
The final model architecture at step 2 had three layers (one hidden layer). About 10 000 samples were available for training for each month; thus, a model with 541 parameters was developed. Note that a higher number of parameters did not show a significant improvement in the accuracy.
2.2.2Reconstruction of surface ocean
The previous section presented the development of the “optimal” architecture of a FFNN model for the reconstruction of global surface ocean and the estimation of its accuracy. This FFNN model was used to provide the final product for scientific analysis and comparison with other mapping approaches. In order to provide the final output, the selected FFNN architecture was trained on all available data: 100 % of data for training, 100 % for evaluation and 100 % for validation. The network was executed five times (different initial values) and the best model was selected based on validation results considering the root-mean-square error (RMSE), the and the bias computed between the network output and the SOCAT-derived surface ocean data. The final model output is referred to as the LSCE-FFNN product.
2.3Computation of sea–air fluxes
The sea–air flux, , was calculated following Rödenbeck et
al. (2015) as
5
where is the piston velocity estimated according to Wanninkhof (1992):
6
The global scaling factor was chosen as in Rödenbeck et
al. (2014) with the global mean piston velocity equaling
16.5 cm h. Sc corresponds to the Schmidt number estimated
according to Wanninkhof (1992). The wind speed was computed from 6-hourly
NCEP wind speed data (Kalnay et al., 1996). is seawater density in
Eq. (5) and is the temperature-dependent solubility (Weiss, 1974).
corresponds to the surface ocean
output of the mapping method. was derived from
the atmospheric mixing ratio fields provided by the Jena
inversion s76_v4.1 (
3.1 Validation
The subset of data used for network validation, which was 25 % of the total, represented independent observations, as they did not participate in training during model development (see Section 2.2.1). The skill of the FFNN to reconstruct monthly climatologies of surface ocean was assessed by comparing colocated reconstructed and corresponding values from Takahashi et al. (2009). The global climatology was reconstructed with satisfying accuracy during step 1 with a RMSE of 0.17 and an of 0.93. The model output of step 2 was assessed using -fold cross-validation as previously presented: different subsets of independent data were drawn from the dataset and the network was run five times on each subset. From these 20 results the best one was chosen based on the RMSE, the and mean absolute error (MAE) (the bias is presented in Table S1 in the Supplement). A combination of the four best model output was used for the statistical analysis summarized in Table 1. Metrics were computed over the full period (2001–2016) and with reference to SOCAT observations (independent data only). At the global scale, the analysis yielded a RMSE of , whereas the MAE was 11.52 and the was 0.76. These results were comparable to those obtained by Landschützer et al. (2013) for the assessment of a surface ocean reconstruction based on an alternative neural network-based approach. The RMSE between the SOCAT data and the climatology from Takahashi et al. (2009) equalled 41.87 , which was larger than errors computed for the regional comparison between FFNN and SOCAT (Table 1). We also estimated the RMSE for the case where 100 % of the data were used for training, which equalled 14.8 and confirmed the absence of overfitting.
Table 1
Statistical validation of LSCE-FFNN. Comparison between reconstructed surface ocean and values from the SOCAT v5 database not used in the training algorithm for the period from 2001 to 2016 over the global ocean (except for regions with ice cover) and for large oceanographic regions. The number of measurements per region is given in parentheses in the first column.
Model | Latitude boundaries | RMSE | MAE | |
---|---|---|---|---|
() | () | |||
FFNN Global | 17.97 | 0.76 | 11.52 | |
Arctic (150) | 76–90 N | 22.05 | 0.54 | 17.1 |
Subpolar Atlantic (21903) | 49–76 N | 22.99 | 0.76 | 15.04 |
Subpolar Pacific (4529) | 49–76 N | 34.77 | 0.65 | 23.12 |
Subtropical Atlantic (41331) | 18–49 N | 17.28 | 0.69 | 11.27 |
Subtropical Pacific (41867) | 18–49 N | 15.86 | 0.77 | 9.9 |
Equatorial Atlantic (7300) | 18 S–18 N | 17.27 | 0.57 | 11.44 |
Equatorial Pacific (27092) | 18 S–18 N | 15.73 | 0.79 | 10.33 |
South Atlantic (3002) | 44–18 S | 17.81 | 0.63 | 12.28 |
South Pacific (12934) | 44–18 S | 13.52 | 0.63 | 9.36 |
Indian Ocean (2871) | 44 S–30 N | 17.25 | 0.62 | 11.6 |
Southern Ocean (16334) | 90–44 S | 17.4 | 0.58 | 11.92 |
Figure 2a shows the time mean difference between the estimated and from SOCAT v5 data used for validation: . Large differences occurred at high latitudes, in equatorial regions, and along the Gulf Stream and Kuroshio currents – the regions with strong horizontal gradients of . Moreover, the standard deviation of the residuals (Fig. 2b) in these regions was larger, indicating that the model fails to accurately reproduce the temporal variability. The reduced skill of the model in these regions reflects the poor data coverage along with a strong seasonal variability (e.g., the Southern Ocean) and/or high kinetic energy (e.g., the Southern Ocean and the Kuroshio and Gulf Stream currents) (Fig. 1a). At the scale of oceanic regions, (Table 1) the largest RMSE and MAE values were computed for the Pacific subpolar ocean (RMSE and MAE ), whereas the lowest correlation coefficient was obtained for the equatorial Atlantic Ocean (). These low scores directly reflect low data density and are to be contrasted with those obtained over regions with better data coverage (e.g., the subtropical North Pacific with RMSE , MAE and , or the subpolar Atlantic with RMSE , MAE and ). Despite large time mean differences computed over the eastern equatorial Pacific, scores are satisfying at the regional scale indicating error compensation by improved scores over the western basin (RMSE , MAE and ). Scores are low in the Southern Hemisphere (Table 1) and time mean differences are large (Fig. 2a) reflecting sparse data coverage (Fig. 1a).
Figure 2Time mean differences () between monthly LSCE-FFNN and SOCAT data used for evaluation of the model over the period from 2001 to 2016 (a) and its standard deviation (b).
[Figure omitted. See PDF]
3.2 QualificationThis section presents the assessment of the final time series of reconstructed surface ocean . The time series was computed using the best monthly models as described in Sect. 2.2, as well as 100 % of the data for learning, evaluation and validation.
Results of the LSCE-FFNN mapping model were compared to three published
mapping methods which participated in the “Surface Ocean pCO2 Mapping
Intercomparison” (SOCOM) exercise presented in Rödenbeck et al. (2015)
(
Figure 3
Map of biomes (following Rodenbeck et al., 2015; Fay and McKinley, 2014) used for comparison. See Table 2 for biome names.
[Figure omitted. See PDF]
We followed the protocol and diagnostics proposed in Rödenbeck et al. (2015) for the comparison of each of the respective mapping methods with one another, with respect to observations. The following diagnostics were computed: (1) the relative interannual variability (IAV) mismatch (in percent), and (2) the amplitude of interannual variations. The relative interannual variability (IAV) mismatch (in percent) is the ratio of the mismatch amplitude of the difference between the model output and the observations (its temporal standard deviation) and the mismatch amplitude of the “benchmark”. The latter was derived from the mean seasonal cycle of the corresponding model output where the trend of increasing yearly atmospheric was added (see details in Rödenbeck et al., 2015). It corresponds to a climatology corrected for increasing atmospheric , but without interannual variability. 7 where Here “mean” is a mean over the region and year, and is the seasonal cycle of from the corresponding mapping method. estimates from Jena inversion s76_v4.1 were used.
Table 2Biomes from Fay and McKinley (2014) used for time series comparison (Fig. 3).
Number | Name |
---|---|
1 | (Omitted) North Pacific ice |
2 | Subpolar seasonally stratified North Pacific |
3 | Subtropical seasonally stratified North Pacific |
4 | Subtropical permanently stratified North Pacific |
5 | Equatorial West Pacific |
6 | Equatorial East Pacific |
7 | Subtropical permanently stratified South Pacific |
8 | (Omitted) North Atlantic ice |
9 | Subpolar seasonally stratified North Atlantic |
10 | Subtropical seasonally stratified North Atlantic |
11 | Subtropical permanently stratified North Atlantic |
12 | Equatorial Atlantic |
13 | Subtropical permanently stratified South Atlantic |
14 | Subtropical permanently stratified Indian Ocean |
15 | Subtropical seasonally stratified Southern Ocean |
16 | Subpolar seasonally stratified Southern Ocean |
17 | Southern Ocean ice |
provides information on the capability of each method to reproduce the IAV compared to observations: a smaller represents a better fit compared with the reference. The amplitude of the interannual variations () of sea–air flux of (its 2-month running mean) is estimated as the temporal standard deviation over the period.
Figure 4Global oceanic : Jena (red), JMA (blue), ETH-SOMFFN (green) and LSCE-FFNN (black). (a) Global average monthly time series, (b) the global 12-month running mean average and (c) yearly mismatch (difference between mapping methods and SOCAT data).
[Figure omitted. See PDF]
3.2.1 Interannual variabilityThe time series of globally averaged surface ocean over the period from 2001 to 2016 are presented in Fig. 4 for LSCE-FFNN and the three other models. Surface ocean () varied between the four mapping methods in the range of (Fig. 4a). Modeled values were at the lower end for ETH-SOMFFN and JMA-MLR, whereas LSCE-FFNN and Jena-MLS13 computed higher values. The same behavior was found for the 12-month running mean time series (Fig. 4b). Figure 4c shows the 12-month running mean of the difference between computed and SOCAT data (model – SOCAT) over the globe. JMA-MLR mostly underestimated observed with a strong interannual variability of the misfit, especially at the end of the period, with up to . The difference between ETH-SOMFFN output and SOCAT data fluctuated in the range of , with an increase in amplitude up to from 2010 onward. Jena-MLS13 overestimated observations with the difference in the range of 0–1 . The difference between LSCE-FFNN and SOCAT varied around zero (between and 1 ).
The model was then assessed at biome scale. Results for all biomes are presented in the Supplement (Figs. S2, S3, S4). Two biomes with contrasting dynamics are discussed hereafter in greater detail: (1) the equatorial East Pacific (biome 6) characterized by a strong IAV of surface ocean and sea–air fluxes in response to ENSO, the El Niño–Southern Oscillation (Feely et al., 1999; Rödenbeck et al., 2015), and (2) the North Atlantic permanently stratified biome (biome 11) with a well-marked seasonal cycle, but little IAV (Schuster et al., 2013). Results for these biomes are presented in Fig. 5.
Figure 5
Surface ocean in the equatorial East Pacific (biome 6) (a, c, e) and the subtropical permanently stratified North Atlantic (biome 11) (b, d, f): FFNN (black), JMA (blue), Jena (red) and ETH-SOMFFN (green). (a, b) The monthly time series averaged over the biome. (c, d) The 12-month running mean averaged over the biome. (e, f) The yearly mismatch (difference between the mapping methods and the SOCAT data).
[Figure omitted. See PDF]
Biome 6 is relatively well-covered by observations and represents a key region for testing the skill of the model to reproduce the observed strong IAV linked to ENSO. El Niño events are characterized by positive SST anomalies, reduced upwelling and decreased surface ocean values. These episodes could be identified in all model time series (Fig. 5a) with reduced levels in 2004/05 and 2006/07 (weak El Niño), 2002/03 and 2009/10 (moderate El Niño), and 2015/16 (strong El Niño). JMA-MLR (blue curve in Fig. 5a) tended to underestimate during weak El Niño events; it was also underestimated during the La Niña 2011–2012 event by Jena-MLS13. LSCE-FFNN and ETH-SOMFFN, both based on a neural network approach, yielded similar results despite differences in network architecture and predictor datasets.
Data coverage is particularly high over Biome 11 (Fig. 5b, d, f). The seasonal cycle in this biome is dominantly driven by temperature. Modeled seasonal variability showed good agreement across the ensemble of methods (Fig. 5b) with an increase in spring–summer and a decrease in autumn–winter. However, the amplitude can be different by up to 10 between different models. The seasonal amplitude of computed by JMA-MLR increased from smaller values at the beginning of the time series to higher values in the middle of the 2005–2012 period. The variability of seasonal amplitude was the highest for Jena-MLS13 in line with the 12-month running mean time series (Fig. 5d). Again, similar seasonal amplitude and year-to-year variability of surface ocean were obtained with LSCE-FFNN and ETH-SOMFFN (Fig. 5b, d). The yearly mismatch (Fig. 5f) shows that observed surface ocean was underestimated by JMA-MLR at the beginning and at the end of the period by up to , and overestimated during the 2007–2011 period by up to 8 . Jena-MLS13 shows mostly positive differences in the range of 0–2 over the full period. LSCE-FFNN and ETH-SOMFFN vary around zero and between and 2 , respectively, and are close to each other.
3.2.2Sea–air flux variability
Sea–air exchange of was estimated using the same gas-exchange formulation (Eq. 5) and wind speed data (6-hourly NCEP wind speed) for each mapping data (Rödenbeck et al., 2005). It is worth noting that the sea–air flux is sensitive to the choice of the wind speed dataset (Roobaert et al., 2018).
Figure 6(a) Interannual global ocean sea–air flux (12-month running mean); (b) the amplitude of interannual flux plotted against the relative IAV mismatch amplitude. The weighted mean is given as a horizontal line.
[Figure omitted. See PDF]
Figure 6a presents the global 12-month running mean of the sea–air flux for four mapping methods. All models showed an increase in uptake in response to increasing atmospheric levels, albeit with a strong between-model variability in multi-annual trends. There is less agreement between the methods compared to reconstructions of surface ocean variability (Fig. 4b). This results from the contribution of uncertainties in sea–air flux estimations over regions with poor data coverage (mostly in the South Hemisphere: the South Pacific, the South Atlantic, the Indian Ocean and the South Ocean; see Fig. S5). Nevertheless, the relative IAV mismatch was less than 30 % for all methods (Fig. 6b), suggesting a reasonable fit to observational data. However, the relative IAV mismatch is a global score, and it is biased towards regions with good data coverage (Rödenbeck et al., 2015). The time series reconstructed in this study is too short to capture decadal variations and, in particular, the strengthening of the sink from 2000 onward (Landschützer et al., 2016). LSCE-FFNN computed a slowdown of ocean uptake between 2010 and 2013 with a flux of GtC yr compared with GtC yr for ETH-SOMFFN. A leveling-off was also found for JMA-MLR, albeit shifted in time. In general, the amplitudes of reconstructed fluxes across all four methods agreed within 0.2 to 0.36 PgC yr. The weighted mean of IAV (horizontal line in Fig. 6b) computed from the four methods included here was 0.25 PgC yr. This value is close to that reported by Rödenbeck et al. (2015) for the complete ensemble of SOCOM models (0.31 PgC yr) estimated for the period from 1992 to 2009. The largest amplitude was obtained for ETH-SOMFFN ( PgC yr). Conversely, LSCE-FFNN has the smallest amplitude with 0.21 PgC yr. Jena-MLS13 and JMA-MLR lie very close to the weighted mean value with 0.26 and 0.22 PgC yr, respectively. The weighted mean and the dispersion of individual models around it, reflect the period of analysis (2001–2015; ETH-SOMFFN output provided up to 2015) and the total number of models contributing to it (see Rödenbeck et al., 2015 for comparison). As such, it does not provide information on the skill of any particular model.
The interannual variability of reconstructed sea–air fluxes (12-month running mean) showed good agreement for biome 6 (East Pacific equatorial, Fig. 7a). A small discrepancy was found at the beginning of the period. A strong increase was computed by Jena-MLS13 for 2010–2014 that was also identified in the variability (Fig. 5a). Despite this, Jena-MLS13 had a low relative (26 %), which confirmed a tendency mentioned in Rödenbeck et al. (2015): that mapping products with a small relative IAV mismatch show larger amplitude. LSCE-FFNN and ETH-SOMFFN yielded comparable results (Fig. 7a, c) with relative IAV mismatches of 46 % and 53 %, respectively, and with amplitudes of PgC yr. Interannual variability reproduced by JMA-MLR falls within the range of the other models (Fig. 7c), but with a of %.
Figure 7Global ocean interannual sea–air flux (12-month running mean) in (a) the equatorial East Pacific (biome 6) and (b) the subtropical permanently stratified North Atlantic (biome 11). The amplitude of interannual flux plotted against the relative IAV mismatch amplitude in (c) the equatorial East Pacific (biome 6) and (d) the subtropical permanently stratified North Atlantic (biome 11). The weighted mean is given as a horizontal line.
[Figure omitted. See PDF]
Reconstructed sea–air fluxes over the North Atlantic subtropical permanently stratified region (biome 11) show large between model differences in amplitudes and variability. The two models based on a neural network again show good agreement with an of 17 % for LSCE-FFNN and 20 % for ETH-SOMFFN. Jena-MLS13 produced a strong seasonal variability (Fig. 7b) up to 0.06 PgC yr, and a small of %. Contrary to the other approaches, JMA-MLR did not reproduce a decrease in sea–air in the middle of the period by up to 0.02 PgC yr (Fig. 7b). The model is characterized by an of 46 % and an amplitude of 0.013 PgC yr.
3.2.3Sea–air flux trend
The long-term trend of sea–air fluxes is dominantly driven by the increase in atmospheric (see Fig. S7). On shorter timescales, such as for the period from 2001 to 2016, the interannual variability at regional scales reflects natural modes of climate variability and local oceanographic dynamics (Heinze et al., 2015).
Figure 8Significant (p_val ) linear trend of for the common period (2001–2015) for (a) LSCE-FFNN, (b) Jena-MLS13, (c) ETH-SOMFFN and (d) JMA-MLR.
[Figure omitted. See PDF]
Table 3Mean of sea–air flux (PgC yr) over the global ocean and per region for the common period (2001–2015). Averages over the period from 2001 to 2009 are presented in parentheses. The last column presents a comparison with the best estimates from Schuster et al. (2013) for the Atlantic Ocean (1990–2009).
Region | Latitude | LSCE-FFNN | ETH-SOMFFN | Jena-MLS13 | JMA-MLR | Schuster et al. (2013), |
---|---|---|---|---|---|---|
boundaries | 1990–2009 | |||||
Global | () | () | () | () | – | |
Arctic | 76–90 N | |||||
Subpolar Atlantic | 49–76 N | () | () | () | () | |
Subpolar Pacific | 49–76 N | () | () | () | () | – |
Subtropical Atlantic | 18–49 N | () | () | () | () | |
Subtropical Pacific | 18–49 N | () | () | () | () | – |
Equatorial Atlantic | 18 S–18 N | 0.085 (0.09) | 0.085 (0.095) | 0.08 (0.082) | 0.1 (0.11) | |
Equatorial Pacific | 18 S–18 N | 0.42 (0.41) | 0.4 (0.4) | 0.44 (0.42) | 0.38 (0.37) | – |
South Atlantic | 44–18 S | () | () | () | () | |
South Pacific | 44–18 S | () | () | () | () | – |
Indian Ocean | 44 S–30 N | () | () | () | () | – |
Southern Ocean | 90–44 S | – |
Figure 8 shows the significant linear trends (p_val ) of sea–air fluxes for LSCE-FFNN (a), Jena-MLS13 (b), ETH-SOMFFN (c) and JMA-MLR (d). A total (averaged over the globe) negative trend was computed for all models, albeit with large regional contrasts, and LSCE-FFNN falls within this range: Jena-MLS13, PgC yr yr ( PgC yr yr, total value with no significant test; Fig. S8); LSCE-FFNN, PgC yr yr ( PgC yr yr); JMA-MLR, PgC yr yr ( PgC yr yr); ETH-SOMFFN, PgC yr yr ( PgC yr yr). LSCE-FFNN computed negative trends over most of the Atlantic basin, the Indian Ocean and south of 40 S, which contrasts with decreasing fluxes over the Pacific and locally in the Antarctic Circumpolar Current. At first order, this broad regional pattern is found in all models. However, regional maxima and minima are more pronounced in Jena-MLS13 (Fig. 8b) and ETH-SOMFFN (Fig. 8c), whereas a patchy distribution at the sub-basin scale is diagnosed for JMA-MLR.
The agreement in sign of computed linear trends from the four models is presented in Fig. 9 (total linear trends with no significant test). Over most of the ocean, all four models show very close sea–air tendency. In the Indian Ocean (biome 14), in comparison, a positive trend was computed for JMA-MLR (0.0004 PgC yr yr; 0.00006 PgC yr yr with a test), whereas the other three models present a negative trend. The differences between models were also found in the Pacific Ocean, especially the southern Pacific. In the eastern equatorial Pacific region (biome 6) a total significant negative trend is presented by all models. All models reproduced a maximum in the southern part of biome 6, but they disagreed about its amplitude and spatial distribution. Almost everywhere over the Atlantic Ocean the mapping methods produced the same sign of linear trend (Fig. 9). However, in the eastern part of the subtropical North Atlantic, Jena-MLS13 gave a positive linear trend of (Fig. 8b).
Figure 9Agreement between the four mapping methods with respect to their linear trend of sea–air flux. The color bar represents the number of products that have the same sign of linear trend.
[Figure omitted. See PDF]
According to LSCE-FFNN, the global ocean took up 1.55 PgC yr on average between 2001 and 2015.This estimate is consistent with results from the other three models (Table 3; see Table S2 for estimations per biome). The spread between individual models falls in the range of the error reported in Landschützer et al. (2016), –0.6 PgC yr. Per biome, estimates of sea–air fluxes provided by LSCE-FFNN are also in good agreement with those derived from the other models.
4 Summary and conclusionWe proposed a new model for the reconstruction of monthly surface ocean . The model is applied globally and allows a seamless reconstruction without introducing boundaries between the ocean basins or biomes. Our model relies on a two-step approach based on feed-forward neural networks (LSCE-FFNN). The first step corresponds to the reconstruction of a monthly climatology. It allows for the output of the FFNN to be kept close to the observed values in regions with poor data cover. In the second step, anomalies are reconstructed with respect to the climatology from the first step. The model was applied over the period from 2001 to 2016. Validation with independent data at the global scale indicated a RMSE of 17.57 , an of and an absolute bias of 11.52 . In order to assess the model further, it was compared to three different mapping models: ETH-SOMFFN (self-organizing maps neural network), Jena-MLS13 (statistical interpolation) and JMA-MLR (linear regression) (Rödenbeck et al., 2015). Network qualification followed the protocol and diagnostics proposed in Rödenbeck et al. (2015).
Reconstructed surface ocean distributions were in good agreement with other models and observations. The seasonal variability was reproduced to a satisfying level by LSCE-FFNN, the yearly mismatch varied around zero and the relative IAV mismatch was 7 %. LSCE-FFNN proved skillful in reproducing the interannual variability of surface ocean over the eastern equatorial Pacific in response to ENSO. Reductions in surface ocean during El Niño events were well reproduced. The comparison between reconstructed and observed values yielded a RMSE of 15.73 , an of 0.79 and an absolute bias of 10.33 over the equatorial Pacific. The relative IAV misfit in this region was %. Despite the overall good agreement between models, important differences still exist at the regional scale, especially in the Southern Hemisphere and, in particular, in the southern Pacific and the Indian Ocean. These regions suffer from poor data coverage. Large regional uncertainties in reconstructed surface ocean and sea–air fluxes have a strong influence on global estimates of fluxes and trends.
Code and data availability
Python code for the climatology reconstruction (the first step of LSCE-FFNN model) and Python code for reconstruction of the anomalies (the second step of LSCE-FFNN model) are provided at the end of the Supplement.
Time series of reconstructed surface ocean and
fluxes are distributed through the Copernicus Marine Environment
Monitoring Service (CMEMS),
The supplement related to this article is available online at:
Author contributions
ADS, MG, MV and CM contributed to the development of the methodology and designed the experiments, and ADS carried them out. ADS developed the model code and performed the simulations. ADS prepared the paper with contributions from all coauthors.
Competing interests
The authors declare that they have no conflict of interest.
Acknowledgements
The authors would like to thank the two referees, Christian Rödenbeck and Luke Gregor, for their helpful comments and questions, as well as Frederic Chevallier and Gilles Reverdin for their suggestions. MV acknowledges funding from the CoCliServ and EUPHEME projects (ERA4CS program).
Financial support
This study was funded by the AtlantOS project (EU Horizon 2020 research and innovation program, grant agreement no. 2014-633211) and the CoCliServ and EUPHEME projects.
Review statement
This paper was edited by James Annan and reviewed by Christian Rödenbeck and Luke Gregor.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2019. This work is published under https://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
A new feed-forward neural network (FFNN) model is presented to reconstruct surface ocean partial pressure of carbon dioxide (
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 Laboratoire des Sciences du Climat et de l'Environnement (LSCE), Institut Pierre Simon Laplace (IPSL), CNRS/CEA/UVSQ/Univ. Paris-Saclay, Orme des Merisiers, Gif Sur Yvette, 91191, France
2 Sorbonne Université, CNRS, IRD, MNHN, Institut Pierre Simon Laplace (IPSL), Paris, 75005, France