Introduction
Chemical transport model (CTM) is a fundamental research tool for atmospheric environment and is widely used for air quality forecasting, source apportionment, and strategy design of pollution alleviation (Chuang et al., 2018; Shen et al., 2020; Zhang et al., 2014). Physical transport process is the driving force for primary pollutants, thus then affects the concentration of secondary pollutants (Byun & Schere, 2006; Tilt, 2019). It is usually numerically solved based on the Euler's mass continuity equation in CTM and occupies 57%~60% of the entire CTM's computation time (Colella & Woodward, 1984; Ying & Li, 2011; Zhang et al., 2013). Moreover, the exponential growth of computational cost is observed with finer spatial resolution due to the smaller time step and larger iteration steps required (Boffi et al., 2007; Kasim et al., 2020). The solution of physical transport process has become the critical bottleneck of the CTM's computation efficiency.
Machine learning, especially deep learning, has become a new research paradigm and acted as partial or whole replacement of complex geoscience models due to its excellent ability for non-linear fitting (Reichstein et al., 2019; Rolnick et al., 2019; Yuan et al., 2020). Specifically, for the surrogate of atmospheric transport process, Lauret et al. (2016) combined cellular automata and the artificial neural network (CA-ANN) to calculate the turbulence coefficient in horizontal two-dimensional (2-D) space. The R squared is over 0.7 for most testcases with the computation efficiency increases by ∼1.5 times. Wang and Qian (2018) extended the study to three-dimensional (3-D) space with the same CA-ANN. R squared for a single time step vary between 0.2 and 0.95, and is reduced to almost zero after 100 multiple-time-step. The computation efficiency increased is still ∼1.5 times. Vlasenko et al. (2021) emulated the entire CTM and the speed gain is 720 times of acceleration. Only 2-D horizontal distribution of daily average concentration, however, is concerned and the R squared values are only in the range of 0.38–0.67. The comprehensive performance with the aspects of efficiency promotion, long-term consistency, and spatial dimensions was still unsatisfied.
This study proposes a deep learning surrogate for the 3-D atmospheric transport process, containing both advection and diffusion mechanisms. The data set was derived from the numerical solution for the transport process of a state-of-the-art CTM, and the features input into the U-Net neural network are determined accurately according to atmospheric transport theory. Validation results indicate that the data-driven deep learning surrogate has an excellent and stable ability under both 5-min and 1-hr time step, and the advantage of Graphics Processing Unit (GPU) hardware could help accelerate the numerical computational efficiency with 164 and 14 times for the two scenarios. The deep learning surrogate is expected to provide feasible ways for not only the forward solution for concentration of atmospheric pollutants, but also the inverse problems such as emission amount estimation derived from its natural property of automatic differentiation (Huang et al., 2021).
Model Development of the Deep Learning Surrogate
Collection and Pre-Processing of Data Set
The data set consisted of meteorological field, emission inventory, and concentrations of pollutant (shown in Figure 1). Meteorological field was estimated by the Weather Research and Forecasting (WRF) model (Skamarock & Klemp, 2008) with the initial conditions and observation data set from the National Centers for Environmental Prediction (NCEP) (Grell et al., 2005; NCEP, 2000, 2004a, 2004b; Otte & Pleim, 2010). Emission inventory was derived from the Multi-resolution Emission Inventory for China (MEIC v1.3, 0.25° × 0.25°, 2016, ) (Li et al., 2017; Zheng et al., 2018) and interpolated to the required 12 km *12 km horizontal resolution. A 3-D CTM numerical model of the Community Multiscale Air Quality (CMAQ) (U.S. EPA Office of Research and Development, 2019) was selected to simulate the concentration evolution with the meteorological field and emission inventory, as well as the benchmark for computation efficiency comparison for the deep learning surrogate. The CMAQ model was regarded as one of the state-of-the-art CTM and its scientific principles and algorithms could be referred to Binkowski and Roselle (2003). In this study, carbon monoxide (CO) was selected as the target pollutant because of its high level of concentration and negligible participation in atmospheric chemical reactions in urban atmosphere (shown in Figure S1 in Supporting Information S1). The modules of “CHEM” and “AERO,” however, were both turned off during the running of the CMAQ model to assure only the transport process which included advection, diffusion and deposition are computed with the exclusion of chemistry process.
[IMAGE OMITTED. SEE PDF]
Two types of temporal resolution, that is, 1-hr and 5-min, were investigated, and the surrogate was developed and validated for each. The resolution of “1-hr” referred to the hourly average concentration and can be used for continuous simulation of inert pollutants only influenced by the transport process. Meanwhile, the resolution of “5-min” referred to the instantaneous concentration after 5 mins such that the surrogate of transport process could be coupled with the modules of chemical reaction and gas-particle partition smoothly for active pollutants. The geographical domain in both the CTM and the deep learning surrogate covered all regions of China and parts of southeast Asia with a 12 km grid resolution (512 rows × 512 columns, shown in Figure S1 in Supporting Information S1), as well as a total of 16 vertical layers with heights increasing from 0 to 19,308 m. A total of 24 days from 8 January to 1 February 2016 with the synchronization time step of 5-min was implemented for CMAQ simulation, and the hourly average concentrations and instantaneous concentrations every 5-min were then output and prepared for model training and evaluation. Therefore, there were 6,912 records for 5-min resolution and 576 records for 1-hr resolution in the data set. We selected 22 days as the training set and remaining 2 days as the validation set.
The input features for each record in the data set consisted of CO's initial concentration, emission amount of CO, wind speed in the horizontal direction and eddy diffusivity in vertical direction. The output target features were the concentration change of CO. The concentration change referred to the difference calculated by subtracting the initial concentration from output concentration after 5-min or 1-hr affected by the transport process. The concentration change was relatively low compared with the absolute concentration. Hence, the concentration alteration rather than the post-transport concentration was set as the output target feature to enable the deep learning surrogate to capture this change sensitively. Max-min normalization method was used for data preprocessing of meteorological features. The logarithmic function was used for data preprocessing of emission amount due to its extremely uneven spatial distribution (Ribeiro & Moniz, 2020).
Deep Learning Model Architecture
Since the physical transport process is mainly affected by instantaneous advection and diffusion conditions under single timestep and the integration of physical module and chemical module required to be conducted in one timestep, a deep U-shape convolutional neural network (U-Net) was selected to construct the deep learning model in this study (Dolz et al., 2018; Xing et al., 2022). U-Net used encoder, decoder and skip connections to extract image features while reducing information loss during encoding and decoding (Ronneberger et al., 2015). Figure 2 presents the specific configurations of U-Net model structures in this study. Five 3-D features with 16 vertical layers were expanded into 80 input channels, and the output was the spatial concentration change of 16 layers. The whole neural network included two sets of “down-sampling,” “deconvolution,” and “skip connection” operations. The configurations of internal model parameters were referred to the original U-Net structure (Ronneberger et al., 2015). A 1 × 1 convolution kernel was used in the last convolutional network layer before the output (Tompson et al., 2017), we excluded the activation function in the last layer as the change of concentration can be positive or negative in our case. Actually three U-Net structures with different depth were tested in this study (shown in Figure S2 in Supporting Information S1). The experimental results for three model structures indicated that the structure with two times of down-sampling obtained better performance than the other two structures according to the corresponding statistical index (shown in Figure S3 in Supporting Information S1). On the basis of U-Net-2, 3-D convolution kernel was also used for training, the R squared are 0.758 and 0.534 for 1-hr and 5-min model, and the root mean square error (RMSE) values are 6.081 × 10−3 and 0.833 × 10−4 for 1-hr and 5-min model, both worse than the case of 2-D convolution kernels (shown in Figure S4 in Supporting Information S1). The poor performance of the 3-D convolution kernel may be due to the characteristics of the convolution kernel which will weaken the edge features and the current network structure of the 3-D convolution kernel. The detailed discussion can be found in Text S1 in Supporting Information S1.
[IMAGE OMITTED. SEE PDF]
A high-level neural network API of Pytorch (version 1.7.0) was used for the implementation of the U-Net model. All the model training and experiments were run in the Python (version 3.6.10) environment. The loss function and optimizer were Relative L2 error (RL2) which could be calculated as Equation 1 (Li et al., 2020) and Adam (Kingma & Ba, 2014). The initial learning rate was 0.001, decreasing by a factor of 10 when the loss did not decrease. The batch size was set to 4, and the epoch size was set to 200. The remaining hyperparameters not described here were set to Pytorch default values. The loss trend for the surrogates and the numerator and denominator of the Relative L2 loss with the batch increases is shown in Figures S5 and S6 in Supporting Information S1.
To evaluate the acceleration ratios of the deep learning surrogate compared to the original CTM numerical model, the consuming clock time of the original CMAQ model and the U-Net model for the simulation of seven consecutive days on the CPU platform (4*40 cores) processor was compared. In addition, the computing time on the GPU platform was also collected for the deep learning surrogate. The specific hardware configurations in this study are listed in Table S1 in Supporting Information S1.
Model Validation of the Deep Learning Surrogate
Consistency Under a Single Time Step
The deep learning surrogate agreed well with the benchmark numerical model for the concentration change after a single time step. Concentration changes for the grids all over the domain were derived and compared (shown in Figure 3). R squared values of the 1-hr and 5-min models were both higher than 0.9. Meanwhile, the slope values of the fitting line for the two temporal resolutions reached as high as 0.9 with the intercepts close to zero. Several clusters of discrete points were noticed to be away from the 1:1 line for the case of 5-min, but not observed in the model of 1-hr. This phenomenon was reasonable as the concentration change was −0.3~0.3 ppm within 5 min, much lower than −2~2 ppm within 1 hr, making the deep learning surrogate more challenging to capture the tiny concentration changes of 5-min.
[IMAGE OMITTED. SEE PDF]
The consistency between the surrogate and benchmark was generally acceptable in vertical dimensional direction, especially for the height under the Planet Boundary Layer (PBL). R squared and RMSE were used for assessing consistency and plotted for each vertical layer (shown in Figure 4). Both two time-resolution models showed that R squared continues to decrease from >0.9 to <0.1 with the increase of vertical height, indicating that the surrogate has better performance in the surface and lower height. The R squared values at the twelfth layer in vertical dimension which corresponding to the top height of PBL (∼3 km), however, could still stay at 0.5–0.6 for two resolutions. It then dropped dramatically to as low as 0.1 at the highest sixteenth layers with the height of 19 km. The pollutants inside the PBL are known to be well mixed while the pollutant's concentration above the PBL can be much lower which cause imbalance in the overall data (Kim et al., 2015; Lu et al., 2019), that will tremendously challenge the approximating ability of the deep learning surrogate (Buda et al., 2018; Liu et al., 2019). Furthermore, the mechanisms of the diffusion process above PBL such as the long-range transport were also more complex and different from those in the lower layers (Hu et al., 2010; Wagstrom & Pandis, 2011). It was interesting that the continuous decrease of RMSE value with higher layer seemed a “paradox” to the decrease of R squared. The trend of RMSE should be mainly attributed to the smaller range of concentration change with height increase rather than representing better performance under a higher layer. Overall, the deep learning surrogate seemed more sensitive to the high concentration changes and the high emission amount in the lower layers. Accurate approximation under a single time step, especially for the height under PBL, provided the prerequisite for the deep learning surrogate to make reasonable continuous predictions.
[IMAGE OMITTED. SEE PDF]
Comparison with the reported index of previous studies manifested the better performance obtained in this study. Lauret et al. (2016) used the CA-ANN structure to simulate the 2-D atmospheric dispersion, with the R squared of 0.7 and slope of 0.84. Wang and Qian (2018) extend the mentioned model to 3-D application, with R squared ranging from 0.3 to 0.9 with the slope of 0.91–1.22 under different scenarios. Vlasenko et al. (2021) approximate the 2-D daily average concentration also derived from the CMAQ numerical model. Their R squared values were in the range of 0.38–0.67. The quality of the training data set and refinement of input features were regarded as the potential reasons for the better performance in this study. First, the data set was derived from the CMAQ numerical model, which is calculated according to strict theoretical equations without any noisy influence. Second, the vertical eddy diffusivity, which was a key parameter for the diffusion process (Marrouf et al., 2015), was set as the input features to help the U-Net model to learn the transport process accurately. This could be verified as we also compared the performance with the replacement of “eddy diffusivity” with more primary meteorological variables such as atmospheric pressure, temperature, humidity, etc., It was observed that the R squared values of primary meteorological variables were close or lower than those of single eddy diffusivity and the number of continuous running steps of primary meteorological variables were less than those of single eddy diffusivity (shown in Figure S7 in Supporting Information S1). At the same time, we consider the influence of boundary conditions and LULC on the model. Adding boundary conditions and adding LULC as an input channel has little effect on the overall model, while adding input channel will increase the computational burden definitely (shown in Figures S8 and S9 and Text S2 in Supporting Information S1).
Consistency Under Continuous Running
Model validation for multiple-time-step was essential for the long-term application of the deep learning surrogate. The initial concentrations at each time step are the sum of the initial concentration at the last time step and the concentration change generated by the deep learning surrogate. The continuous iterations of 400 (∼17 days) and 1,000 (∼83 hr) time step could be regarded as the lifetime for the resolution of 1-hr and 5-min, respectively, according to the comprehensive performance of two statistical indexes. Indicators of R squared and RMSE referred to the absolute ambient concentration for the grids all over the domain between the surrogate and the benchmark were estimated (shown in Figure 5). The values of R squared for the two resolutions decreased continuously due to the accumulation of errors in multiple-time-step. It reduced to 0.5 at the 400th step and 0.3 at the final 576th step for 1-hr resolution. Similarly, the R squared reduced to 0.4 at the 1,000th step and 0.1 at the final 1,200th step for 5-min resolution. Meanwhile, inflection points at the 400th step and 1,000th step for the 1-hr and 5-min resolution, respectively were observed with the evolution of RMSE values. The statistical result based on the concentration change during continuous running was also shown in Figure S10 in Supporting Information S1. The result for concentration change was generally agreed with that of the absolute concentration. The error of the 1-hr surrogate was dominantly originated from the internal bias of individual time step, while the error of the 5-min surrogate was significantly influenced by the bias accumulation of multiple steps.
[IMAGE OMITTED. SEE PDF]
The statistical performance of different vertical layers, however, presented significant diversities. Figure S11 in Supporting Information S1 selects five representative vertical layers corresponding to the height of below PBL (Layers 1, 4, and 8), PBL boundary (Layer 12) and above PBL (Layer 16). For the 1-hr deep learning surrogate, the R square values of the lower height (Layers 1 and 4) could maintain at ∼0.6 even after 576 consecutive steps. In contrast, the statistical performance of upper height (Layers 12 and 16) was much worse than that of layers under PBL. The sudden change of RMSE from Layer 12 at the 400th step should be responsible for the inflection point of Figure 5a as the RMSEs for other layers raised slowly and smoothly. The trend for the 5-min surrogate was similar to that of 1-hr, while the only difference was that the cause of RMSE inflection was attributed to all vertical layers rather than some single layer. The vertical performance of multiple-time-step was dominantly consistent with that of a single time step. However, the surface layer was what we cared most considering of health and ecosystem, and fortunately it had the best performance throughout all the vertical layers.
Spatial distribution could help to identify the deviation sources for the surrogates with two resolutions. Figure 6 shows the concentration comparison of spatial distribution in the surface (i.e., Layer 1) for 1-hr and 5-min resolutions, and the results of other vertical layers are presented in Figure S12 in Supporting Information S1. The spatial distribution of the surrogate could be observed to be highly consistent with the numerical benchmark until the 576th step for 1-hr resolution, especially for the hotspots area with high concentration. The deviation area mainly came from the offshore regions with low concentration, that is, the south-east part of the domain, which could be observed more clearly from the difference spatial distribution given by Figure S13 in Supporting Information S1. The surrogate might overestimate as high as 0.3 ppm at the sea area after the 200th step, but very limited underestimation was seen at the mainland until the final 576th step. In contrast, the surrogate for the 5-min resolution was mainly influenced by the deviation sources from the hotspots of mainland in China. Limited underestimation was seen until the 600th step. However, both significant overestimation and underestimation were existed at the hotspots of mainland in China from the 1,000th step, resulting in the overall worse performance compared with that of 1-hr resolution.
[IMAGE OMITTED. SEE PDF]
The temporal concentration evolution of vertical distribution was suitable for assessment of the 3-D surrogate in this study. It is difficult to present the vertical evolution for all grids, hence five cities all over China, that is, Beijing (north), Shanghai (east), Kunming (west), Changsha (middle) and Guangzhou (south) were selected. These five cities could also represent different pollution levels. The evolution of vertical concentrations was seen to be consistent well between the deep learning surrogate and numerical benchmark for both two resolutions (shown in Figure 7). The diurnal variation of concentration caused by the temporal change of PBL height, which was demonstrated by the numerical CMAQ model, could also be approximated almost perfectly by the deep learning surrogate for all cities. Relatively speaking, the performance of 1-hr resolution was better than that of 5-min resolution. For instance, there was a false hotspot pollution at the middle height during the 400th −800th step for Changsha city indicated by the surrogate, not seen in the benchmark numerical model.
[IMAGE OMITTED. SEE PDF]
The performance for multiple-time-step was always a bottleneck of deep learning surrogate from the reported studies. Wang and Qian (2018) tested one hundred different scenarios, and the R squared values dropped to zero at the 100th consecutive time steps for most scenarios, much smaller to the lifetime of 400 and 1,000 timesteps for the two resolutions in this study. Multiple-time-step validation has not even been investigated in the other two studies (Lauret et al., 2016; Vlasenko et al., 2021). In summary, the deep learning surrogate in this study could be generally implemented for a days-to-weeks application, especially for the area with notable anthropogenic emission and the height below PBL height such as the surface concentration.
Promotion of Computation Efficiency
Consuming clock time for the same period under different models and hardware configurations are compared and shown in Figure 8. It is worth noting that the consuming time of 1-hr was equal as that of 5-min for the CMAQ numerical benchmark because the results of 1-hr was averaged from the raw results of 5-min. Furthermore, the original CMAQ numerical model was written in Fortran framework and not suitable for the implementation on GPU platform, thus there was no computational time for the benchmark numerical model on GPU hardware.
[IMAGE OMITTED. SEE PDF]
The enlargement of the time step could notably reduce the total computation time under the same hardware. In this study, the consuming time was reduced from 460 min of the benchmark model to only 39 min of the deep learning surrogate for 1-hr resolution under the same CPU configuration (4*40 cores), achieving ∼12 times acceleration. The advantage of the deep learning surrogate over the numerical model was the end-to-end training method, which can directly predict the concentration 1 hr later without fixed iterations required by the numerical CMAQ model. However, the deep learning surrogate is not necessarily faster than the numerical benchmark under the same time step and CPU hardware. In this testcase, the deep learning surrogate with 5-min resolution consumed 456 min, almost the same as 460 min of the numerical CMAQ model using the same 4*40 CPU cores. If the deep learning architecture was more complicated or there were more inputs features, the consuming time of the surrogate could be even larger than the benchmark. This is explainable and reasonable because the large size of 3-D features input, complex architecture, and numerous neural nodes will bring large amount of big matrix operations. In contrast, the numerical operations of the traditional algorithm for physical transport process were simplified and optimized (Byun & Schere, 2006; Colella & Woodward, 1984).
Nevertheless, the deep learning architecture can efficiently accelerate the calculation relied on the GPU hardware. Compared with the time consuming (456 min for 5-min resolution and 39 min for 1-hr resolution) at the CPU platform of 4*40 cores, the same deep learning surrogate could reduce it to 32 and 3 min, respectively, at the platform of single one GPU. The accelerating ratio could reach as high as 13–14 for both two resolutions only attributed to the GPU hardware, which has the natural advantages in parallel tasks process.
In summary, the deep learning surrogate could accelerate the computation efficiency of the benchmark numerical model with the maximum speedup factor of 164 relying on the integrated benefit of end-to-end algorithm and GPU hardware. Reported acceleration work via deep learning also verified the above conclusions from this study. Only benefit from the end-to-end algorithm, Lauret et al. (2016) and Wang and Qian (2018) reported a speedup factor of 1.5 times through their neural networks. Liu et al. (2021) and Kelp et al. (2018) realized 10.6 times and 260 times acceleration, respectively via deep learning. Relying on both algorithm characteristics and GPU hardware, Vlasenko et al. (2021) reported 720 times speedup for a 1-day chemical transport simulation and Liu et al. (2021) reported 85.2 times speedup for a one-hour gas-phase chemistry solver simulation.
Limitations and Future Work
Although longer lifetime, better consistency, and significant acceleration for the deep learning surrogate of atmospheric transport process have been achieved, there are still several limitations in this study that should be investigated in the future. Currently, the study only focuses on carbon monoxide as the target pollutant, which was abundant in urban atmosphere. The applicability of the surrogate to other pollutants should be widely assessed as the emission intensity and concentration level vary in a large range for different pollutants. Additional normalization preprocessing for different species might be necessary to enhance the surrogate's applicability. Another challenge is the applicability transfer to other meteorological seasons. The training data set and validation data set now were both in the winter season. It might be uncertain to apply the winter surrogate to other seasons such as summer because of significant different meteorological conditions. Inclusion of other seasons for the training data might relieve or solve this problem. After sufficiently validated, the deep learning surrogate for atmospheric transport process will be assembled with the deep learning surrogate for chemical process developed by our group Liu et al. (2021) to obtain a real chemical transport deep learning surrogate.
Conclusions
In this paper, a deep learning surrogate based on U-Net architecture was developed toward replacing the 3-D atmospheric transport process. Validation results showed that the deep learning model could well reproduce the horizontal advection and vertical diffusion processes. The R squared value was over 0.9 for a single time step, and the lifetimes for continuous running reached 400 and 1,000 steps for 1-hr and 5-min resolutions. The largest speedup factor of 164 was achieved for the deep learning surrogate via the advantage of end-to-end algorithm and GPU architecture.
This work preliminarily proved the feasibility of data-driven approach in approximating the 3-D diffusion process of pollutants. Meanwhile, computational efficiency can be notably improved while maintaining high consistency with the numerical model. The deep learning surrogate is expected to be a real “CTM” after coupling with the remaining chemical emulator. The natural property of automatic differentiation with the surrogate will also provide a convenient solution for inverse problems such as emission amount estimation.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (T2122022, 41975152). The CTM simulations and deep learning model training were completed on the “π 2.0” cluster system of the Center for High Performance Computing in Shanghai Jiao Tong University.
Data Availability Statement
The source code of CMAQ can be retrieved from (U.S. EPA Office of Research and Development, 2019) [Software]. The source code of WRF can be retrieved from (Skamarock & Klemp, 2008) [Software]. The MEIC Data can be retrieved from (Li et al., 2017; Zheng et al., 2018) [Data set]. The NECP FNL data set is available to download from (NCEP, 2000) [Data set]. The surface observations and vertical observations data can be found and downloaded from and (NCEP, 2004a; 2004b) [Data set]. The retrieved training and validation datasets are too large (>1T) to be uploaded for share. Anyone who is interested in them is welcomed to contact the corresponding author for a point-to-point transfer. The source code of the deep learning model could be found in Zenodo: (Xu et al., 2022) [Software].
Binkowski, F. S., & Roselle, S. J. (2003). Models‐3 Community Multiscale Air Quality (CMAQ) model aerosol component 1. Model description. Journal of Geophysical Research, 108(D6), [eLocator: 4183]. [DOI: https://dx.doi.org/10.1029/2001JD001409]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2022. This work is published under http://creativecommons.org/licenses/by-nc-nd/4.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The physical transport process is the bottleneck of the computational efficiency in regional chemical transport modeling. The issue will be worse with the smaller time step due to increased iterations required with finer spatial resolution at scale. Reported surrogates of the transport process are usually unfeasible according to integrated assessment of efficiency promotion, long‐term consistency, and spatial dimensions. This study intended to approximate the three‐dimensional (3‐D) transport process (including advection and diffusion) of a state‐of‐the‐art chemical transport model, that is, Models 3/Community Multiscale Air Quality (CMAQ), via the U‐Net structure of deep learning. Two temporal resolutions of models with 1‐hr and 5‐min were developed. Validation results indicated that single‐step R squared of both models were higher than 0.9, and the lifetime for continuous running was 400 and 1,000 steps for 1‐hr and 5‐min model, respectively. Meanwhile, the computational efficiency can be promoted with the maximum of 164 times for 1‐hr and 14 times for 5‐min resolution on one GPU. The 1‐hr deep learning surrogate could still achieve 12 times acceleration on the same CPU configurations of CMAQ, mainly through the end‐to‐end direct inferring rather than time step iterations. This study preliminarily proves the feasibility of the data‐driven approach in approximating the 3‐D complex transport process of atmospheric pollutants. Furthermore, computational efficiency can be efficiently improved while maintaining consistency and accuracy. Rapid transport simulation of different pollutants with wide concentration range can be expected, which will finally benefit the acceleration of whole chemical transport modeling.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Shanghai Environmental Protection Key Lab of Environmental Big Data and Intelligent Decision‐making, School of Environmental Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
2 MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, Shanghai, China
3 Network & Information Center, Shanghai Jiao Tong University, Shanghai, China