Content area
Efficient prediction of shallow-water acoustic transmission loss (TL) is crucial for underwater detection, recognition, and communication systems. Traditional physical modeling methods require repeated calculations for each new scenario in practical waveguide environments, leading to low computational efficiency. Deep learning approaches, based on data-driven principles, enable accurate input–output approximation and batch processing of large-scale datasets, significantly reducing computation time and cost. To establish a rapid prediction model mapping sound speed profiles (SSPs) to acoustic TL through controllable generation, this study proposes a hybrid framework that integrates a variational autoencoder (VAE) and a normalizing flow (Flow) through a two-stage training strategy. The VAE network is employed to learn latent representations of TL data on a low-dimensional manifold, while the Flow network is additionally used to establish a bijective mapping between the latent variables and underwater physical parameters, thereby enhancing the controllability of the generation process. Combining the trained normalizing flow with the VAE decoder could establish an end-to-end mapping from SSPs to TL. The results demonstrated that the VAE–Flow network achieved higher computational efficiency, with a computation time of 4 s for generating 1000 acoustic TL samples, versus the over 500 s required by the KRAKEN model, while preserving accuracy, with median structural similarity index measure (SSIM) values over 0.90.
Full text
1. Introduction
Acoustic waves serve as the sole effective medium for long-distance information transmission in marine environments, playing a crucial role in modern underwater communication and sensing applications. In shallow water regions, acoustic propagation primarily manifests as a complex waveguide phenomenon, governed by the coupling between dynamic oceanographic variables and boundary interactions. Acoustic transmission loss (TL) stands out as a key quantity in underwater acoustics for characterizing acoustic propagation. Conventional physics-based approaches for underwater TL prediction rely on numerical solutions, such as KRAKEN based on normal mode [1], and BELLHOP based on ray tracing [2]. However, these methods inevitably demand intensive recalculations for each new scenario, limiting their efficiency in generating massive datasets.
In contrast, deep learning offers a data-driven alternative with strong nonlinear mapping capabilities, enabling accurate input–output approximation and batch processing of large-scale datasets, thereby eliminating the need for recalculations. Critically, this advantage has been exploited in recent advances across underwater acoustics, particularly through two paradigms: (i) deep generative augmentation methods, and (ii) supervised surrogate models. The first paradigm leverages deep generative models to learn the probability distribution of underwater acoustic data through latent variables, swiftly generating synthetic samples that statistically indistinguishable from empirical observations. Notably, variational autoencoders (VAEs) [3], renowned for their training stability and explicit low-dimensional representations, have proven effective in generating underwater acoustic channel impulse responses (CIRs) samples for data augmentation [4]. Furthermore, generative adversarial networks (GANs) [5] have demonstrated remarkable efficacy in underwater acoustic data augmentation due to their capability for producing high-fidelity synthetic samples. Liu et al. [6] employed a conditional generative adversarial network (CGAN) to synthesize cross-spectral features from image representations, which improves source ranging accuracy through enhanced feature generation. Similar extensions using GAN architectures [7,8,9] have further validated this paradigm’s effectiveness in data augmentation, where synthetic sample generation substantially enhances downstream task performance through dataset expansion. However, this approach operates as an unsupervised framework, lacking explicit input–output mapping capabilities, thus limiting controllable scenario-specific prediction.
Unlike generative models that learn data distributions, the second paradigm employs supervised deep neural networks as surrogate models to learn deterministic input-output mappings, enabling efficient controllable prediction. Varon et al. [10] trained a deep neural network as a surrogate model to predict modal wavenumbers and group speeds across diverse environments, replacing computationally intensive normal mode simulations and enabling rapid TL calculation. Subsequent developments employed specialized architectures as surrogates for TL prediction. Mallik et al. [11] used a convolutional recurrent autoencoder network (CRAN), while Sun et al. [12] developed a Multi-Scale-DUNet for TL field prediction across varied source conditions. Nevertheless, these supervised surrogate models depend on extensive labeled data and exhibit a performance tied to the training data distribution, and significant deviations in unseen test conditions may trigger accuracy degradation.
While deep generative augmentation methods learn data distributions for enhanced generalization, they inherently lack controllability. Conversely, supervised surrogate models enable scenario-specific prediction, but their generalization to unseen conditions may be constrained by training data convergence. To develop rapid prediction model mapping of sound speed profiles (SSPs) to acoustic TL through controllable generation, this study advances the first paradigm by introducing a hybrid variational autoencoder–normalizing flow (VAE–Flow) framework through a two-stage training strategy. The framework specifically employs (1) a VAE to learn latent representations of TL data on a low-dimensional manifold, and (2) a normalizing flow to establish bijective mapping between latent variables and physical parameters. This approach addresses the controllability limitation of the first paradigm by reinforcing variable mappings through the integration of a normalizing flow. Combining the trained normalizing flow with the VAE decoder can establish an end-to-end mapping from SSPs to TL, achieving prediction speeds surpassing the conventional KRAKEN model.
2. Method
This section proposes a hybrid framework for TL prediction, comprising a VAE and a normalizing flow: the VAE learns low-dimensional representations of acoustic TL data through latent variables, while the normalizing flow establishes a bijective mapping between the latent space and physical parameters, achieving end-to-end mapping from physical parameters to TL.
2.1. Problem Description
For a point source of angular frequency positioned at depth , the acoustic pressure field at depth z and horizontal distance r can be expressed as a sum of normal modes [13]:
(1)
where represents the medium density, is the mth mode wavenumber, and denotes the mth modal depth function.The outcome describes a two-dimensional spatial distribution of acoustic pressure, characterized as a complex-valued field that incorporates both magnitude and phase. In underwater acoustics research, TL is typically used for analyzing the spatial distribution of the acoustic field. Such TL can be computed from at a given receiver location with respect to the reference pressure :
(2)
where is the acoustic pressure for the source in free space.In complex ocean environments, TL is influenced by multiple factors, including the sound speed profile, seabed topography, and the source properties, among others. By parameterizing these influencing factors into a parameter vector , the transmission loss field , obtained by discretizing the receiver region into N spatial points, can be formally expressed through the forward mapping:
(3)
The physical parameter vector comprises r variables representing environmental properties and source conditions. Each component in quantifies a distinct physical factor governing acoustic propagation.
In underwater acoustics, KRAKEN, as a conventional numerical model for computing low-frequency acoustic fields in shallow-water environments, has been widely recognized for its accuracy and reliability in calculating TL and serves as an alternative model for generating datasets in the absence of experimental data. By configuring the parameter vector , KRAKEN effectively computes the acoustic field for the parameterized scenario and outputs the acoustic TL field via Equation (2). Nevertheless, when generating large-scale TL samples across high-dimensional parameter spaces, KRAKEN exhibits a fundamental limitation: any variation in components necessitates recomputing and through numerical solvers. This results in prohibitive computational costs, rendering it impractical for efficient generation of massive sample sets.
To address this computational constraint, this study aimed to develop a surrogate model mapping SSPs to acoustic TL through controllable generation and achieved the following functionality:
(1) Forward modeling: The model learns a mapping that approximates with high fidelity, while enabling swift prediction.
(2) Probabilistic sampling: By learning the TL data distribution, the model generates samples statistically indistinguishable from observations, with .
2.2. VAE–Flow Framework
To properly frame our solution, it is necessary to briefly review the manifold learning principles underlying the proposed objective. Since TL represents the spatial distribution of acoustic wave energy, the data inherently exhibit a high-dimensional structure. But it should be emphasized that the high-dimensional TL data do not uniformly populate the ambient N-dimensional space [14]. Under the manifold hypothesis, these observations reside on an r-dimensional manifold (), where r corresponds to the intrinsic dimensionality of the physical parameter space , and denotes the ground-truth probability measure within . As shown in Figure 1, a 2D physical parameter space forms a low-dimensional manifold embedded in the high-dimensional observation space.
Consequently, achieving controllable generation for rapid TL prediction, while meeting the functional requirements outlined in Section 2.1, basically necessitates
(1) Learning low-dimensional representations of TL data through encoding–decoding, where the latent space dimensionality d matches the intrinsic dimension r of the data manifold.
(2) Accurately capturing the ground-truth measure supported on the manifold and approximating a homeomorphism between the latent space and the manifold.
To address these modeling requirements, we developed a hybrid VAE–Flow framework through a two-stage training strategy, comprising both data generation and model training, as illustrated in Figure 2. To begin with, the KRAKEN model is employed to construct underwater TL datasets under varying SSPs for subsequent model training. Subsequently, during the model training, the framework was further decomposed into two distinct stages: First, a VAE was trained to learn latent representations of TL data on a low-dimensional manifold, while high-fidelity TL fields reconstruction could be achieved by sampling from the learned latent variable distributions and then passing them through the trained decoder. Second, a normalizing flow was independently optimized to establish a bijective mapping between the learned latent variables and the corresponding physical parameters of the simulated TL datasets. Ultimately, combining the trained normalizing flow with the VAE decoder established an end-to-end mapping from physical parameters to predicted TL.
Before proceeding further, it is essential to examine the theoretical foundations of this framework through the lens of VAEs and normalizing flows, with specific attention to two key aspects:
(1) Why is the model training stage divided into two separate phases?
(2) What are the functional contributions of the VAE and Flow network within each phase?
We employed VAE as our first-stage model due to its encoder–decoder architecture for effective low-dimensional representation learning. As a likelihood-based deep generative model, the VAE fundamentally performs maximum likelihood estimation (MLE) for data modeling [15]:
(4)
where the observed data x are assumed to be independently and identically distributed, and is the probability density function parameterized by . However, the integral of the marginal likelihood is intractable. The VAE addresses this by introducing a recognition model to approximate the true posterior , where can be viewed as tuning the degree of approximation. Under this framework, the log-likelihood objective of MLE can be decomposed as follows:(5)
Here, represents the Kullback–Leibler (KL) divergence measuring the discrepancy between and . Since KL divergence is non-negative, the first RHS term is called the Evidence Lower Bound (ELBO), and Equation (5) can be written as
(6)
with equality iff (i.e., when equals zero, the ELBO becomes tight). Here, represents the mathematical expectation operator, and is the prior over latent variables z. In Equation (6), functions as an encoder, while acts as a decoder, both are commonly specified as Gaussians. By introducing latent variables and the reparameterization trick, the VAE enables optimization of the following objective function:(7)
where represents a random variable sampling from and L denotes the total number of random samples.Through neural network optimization of this objective function, the VAE learns latent representations of the data. While sampling from the latent space and passing through the decoder enables data synthesis, this generation process remains uncontrolled. The core limitation stems from VAE’s construction of the density function in the full-dimensional observation space, despite the data manifold having a lower intrinsic dimensionality. This mismatch prevents the model from learning the ground-truth probability measure on the manifold, even when the objective function (ELBO) is maximized. Mathematically, this implies that maps all samples from to the correct manifold with a trivial reconstruction error for any , but the measure on has not been accurately estimated:
(8)
Here, represents the aggregated posterior in the latent space, which errantly deviates from the prior distribution . This may lead to generating samples outside the data manifold when , as the decoder operates in ambient space, while the data lie on .
To address this limitation, Daibin et al. [16] proposed a Two-Stage VAE, later generalized by Loaiza-Ganem et al. [17] as Two-Step Models. The essential of such methods lies in (i) first learning low-dimensional representations through an encoder–decoder architecture where the latent space dimensionality matches the intrinsic manifold dimension, ensuring the latent variables sampled from this space exhibit non-zero measure across the full ambient space ; (ii) training a second deep generative model on this properly aligned latent space to capture the manifold’s ground-truth probability measure [18]. While Daibin et al.’s approach employs two independent VAEs for higher quality image generation, we propose replacing the second-stage VAE with a normalizing flow to better satisfy the functional requirements of controllable generation, making it more suitable for TL prediction tasks. Consequently, a brief introduction to the principles of normalizing flows is warranted here.
Originally proposed by Rezende et al. [19], normalizing flows operate based on the core idea of exact distributional mapping via invertible variable transformations. For instance, to achieve bijective mapping between the original distribution and the latent distribution , normalizing flows construct a series of multivariable bijective functions . This allows
(9)
The probability densities are preserved through a change in the variable formula:
(10)
Here, J denotes the Jacobian matrix comprising all first-order partial derivatives. Thus, normalizing flows enable exact computation of the log-likelihood in MLE by evaluating the Jacobian determinant during variable transformation, as formalized below:
(11)
where K denotes the total number of multivariable bijective functions constituting the transformation sequence. Crucially, normalizing flows enable exact probability density estimation through invertible transformations, and precise bijective mapping between variable distributions [20]. These properties make normalizing flows better suited than VAEs for controllable generation tasks.Building upon the theoretical foundations, we now address the two previously posed research questions. Our hybrid framework achieves controllable generation by training a VAE and a normalizing flow separately: (1) Training a VAE to learn low-dimensional TL representations with latent space dimensionality aligned to the dimensionality of empirical orthogonal functions (EOF) coefficients (detailed in Section 3.1), ensuring full-dimensional coverage in , while minimizing the reconstruction error; (2) Optimizing a normalizing flow to establish a bijective mapping between latent variables and EOF coefficients, which is equivalent to minimizing , thereby learning the ground-truth probability measure on the manifold. Through two-stage training, the VAE–Flow framework essentially approximates a homeomorphic mapping between latent and physical parameter spaces, thereby enhancing the controllability of the generation process and establishing a forward mapping from SSPs to TL fields.
Notably, while existing studies [19,21] combined VAEs with normalizing flows, their objectives differed from ours. These works primarily employed normalizing flows to enhance the variational posterior approximation, replacing the standard Gaussian assumption with more flexible, complex distributions to improve image generation quality:
(12)
In Equation (12), is the initial Gaussian latent variable from the encoder, while is its transformed version through normalizing flows. However, these methods are not designed to establish an effective bijective mapping, rendering them unsuitable for controllable generation. By contrast, our framework decomposes the optimization into two dedicated stages, which ensures each model specializes in its target function, while eventual integration enables controllable generation. The complete procedure for the hybrid VAE–Flow framework can be summarized as follows:
(1) Given physical parameters drawn from a prior distribution , compute the corresponding acoustic TL fields via KRAKEN, obtaining M simulation samples .
(2) Train a VAE, with the latent space dimension set equal to the dimensionality of the physical parameters, to learn the manifold structure underlying the acoustic TL data (the reconstruction error between the decoder’s output and ground truth is minimized). Generate latent variables via , which approximately follow a distribution but are unlikely to match .
(3) Train an additional normalizing flow, take the samples as input, to learn a bijective mapping between and .
(4) Upon completion of training, end-to-end prediction from physical parameters to acoustic TL can be achieved via , , and .
Based on this theoretical framework, we further construct a VAE–Flow network in Section 3 to achieve rapid prediction of acoustic TL under different SSPs.
3. Method Implementation
This section details the implementation of the VAE–Flow framework, including data generation procedures, the VAE–Flow network architecture, training configurations, and evaluation metrics for controllable TL prediction.
3.1. Dataset Generation Procedures
In shallow-water environments, the propagation characteristics of acoustic waves are influenced by multiple environmental factors, with the sound speed profile serving as a critical parameter that characterizes the temporal and depth-dependent variations in sound speed in seawater. As a fundamental component for constructing acoustic propagation models, the SSP directly determines acoustic wave propagation paths and velocities, making it an essential environmental parameter for accurate sound field modeling. This study primarily focuses on the influence of SSP on acoustic propagation. Given the scarcity of underwater experimental data, we generated simulated acoustic TL data via the KRAKEN model under various SSPs to construct the training dataset for our network.
As illustrated in Figure 3, a typical shallow-water waveguide with a negative thermocline is parameterized by water depth, density, SSP, and seabed acoustic properties. In this case, the water depth H was 50 m, with the source depth fixed at 25 m. The water density was 1000 kg/m3, and the sound speed was characterized by a negative thermocline profile, with an upper-layer sound speed of 1520 m/s and a lower-layer sound speed of 1480 m/s, while compressional wave attenuation in the water column was neglected. The seabed was modeled as an infinite half-space with a sound speed of 1650 m/s, a density of 1700 kg/m3, and an absorption coefficient of 0.5 dB/. The sea surface was modeled as a pressure–release boundary, with water mass movements excluded during the simulations. During data generation, we employed the waveguide environment illustrated in Figure 3, with the aforementioned parameters serving as the baseline configuration for the KRAKEN simulations.
To generate a large-scale dataset of acoustic TL, this paper employed the empirical orthogonal function (EOF) method to reduce the dimensionality and reconstruct SSPs. Multiple SSPs were generated through this approach for acoustic TL computation and dataset construction, with the detailed workflow illustrated in Figure 4. Since the eigenvectors obtained from eigenvalue decomposition of a sound speed covariance matrix are mutually orthogonal, the first m dominant modes with high energy contribution rates can be selected for SSP reconstruction:
(13)
where represents the average sound speed profile, denotes the i-th order EOF mode, and is the corresponding i-th EOF coefficient.The generation process began by creating N initial sound speed profiles through random perturbations, wherein the upper-layer depth of the negative thermocline was randomly sampled 100 times across [10 m, 30 m], and the lower-layer depth was defined as (with m). Upon subjecting these 100 profiles to EOF decomposition, the variance contribution rates of the first five modes were quantified as , , , , and , respectively, revealing that the first two modes dominated of the total energy. Consequently, these two modes were selected for profile reconstruction in this section, with their spatial distributions and the average sound speed profile illustrated in Figure 5. To align with the proposed modeling framework, the EOF coefficients had to adhere to specific probability distributions during reconstruction, which was implemented by sampling from two independent priors to generate M coefficient sets. It can be established that the physical degrees of freedom for both SSPs and acoustic TL field data were determined by the dimensionality of the EOF coefficients.
This study assumed these two EOF coefficients followed Gaussian distributions, specifically and . For computational convenience, we set and defined six distinct standard deviation values, constructing the dataset as presented in Table 1. For each EOF parameter set, we sampled 6000 data points and generated 6000 corresponding SSPs using Equation (13). Subsequently, the SSPs corresponding to the six datasets were utilized as environmental parameters in KRAKEN to compute the sound field, where the source frequency was uniformly set to 200 Hz, and the simulated area covered a horizontal range of 1–10 km ( m) and a vertical range of 1–50 m ( m). This process yielded six groups of acoustic TL data matrices . Each TL sample was subsequently converted into an RGB image with pixel dimensions of , where represents the image resolution and 3 denotes the color channels (red, green, and blue). Thus, six distinct sets of TL image datasets, each determined by different EOF coefficient distributions, were further used to train the VAE–Flow network. It should be noted that TL was selected as the input for two key reasons: First, neural networks are sensitive to the numerical distribution of input data during training, and TL exhibits a relatively constrained dynamic range that mitigates the training instability caused by magnitude disparities. Second, compared to 1D features, TL representations provide a more comprehensive characterization of acoustic wave propagation in spatial domains. Furthermore, by formatting the data as RGB images, we could fully leverage the inherent advantages of convolutional neural networks in feature extraction, thereby significantly enhancing the model’s learning capacity.
3.2. VAE–Flow Network Architecture
To enable rapid prediction of shallow-water acoustic TL, a VAE–Flow network was constructed based on the modeling framework proposed in Section 2. In this architecture, the VAE network is employed for learning the latent representations of the acoustic TL fields, while achieving high-fidelity TL reconstruction, whereas the Flow network serves to establish a bijective mapping between the learned latent variables distribution and the target distribution of the EOF coefficients. As previously established in Section 2, these two training phases are mutually independent. Given that the acoustic TL datasets consisted of RGB images with pixel resolution, this section incorporated convolutional layers and residual blocks into the fundamental VAE. These enhancements were implemented to improve the network’s feature extraction capability for image data, and to mitigate potential gradient vanishing issues.
The architecture of the VAE network developed in this paper is illustrated in Figure 6. In accordance with the VAE principles, the network consists of encoder and decoder components. The encoder comprises an initial downsampling module, five residual blocks and a mapping module, while the decoder is composed of a projection module, seven residual blocks, and an output module. Considering the capability of convolutional layers in capturing the intrinsic characteristics of data, we selected them as the core component for feature extraction. The initial downsampling module of the encoder consists of a convolutional layer, followed by batch normalization and the LeakyReLU activation function. Additionally, a max pooling layer is also attached to reduce the spatial dimensions of the feature maps. It should be noted that the encoder incorporates several residual blocks, which originate from the classical ResNet [22]. Each residual block consists of two repetitive units containing a convolutional layers, batch normalization, and a LeakyReLU activation function. Crucially, a skip connection directly sums the input to the feature map before the final LeakyReLU activation function, which mitigates vanishing gradients via identity shortcut paths. Moreover, compared to the first block, the residual block introduces an additional linear transformation ( convolutional layer), as shown in Figure 6, specifically designed to address the dimensional mismatch between inputs and outputs caused by downsampling operations. Following this, the mapping module employs a global average pooling layer to reduce the spatial dimensions of the feature maps to , subsequently branching into two parallel dense layers that output the mean and standard deviation parameters of the latent distribution. Through the implementation of this architecture, the encoder learns to map high-dimensional acoustic TL images into a low-dimensional latent space during training, while outputting and of the latent variables to represent the learned underlying distribution.
In the decoder, the projection module first upsamples the latent variable z through a dense layer followed by a ReLU activation function, then reshapes the vector back into feature maps. The residual blocks in the decoder maintain the same architecture as those in the encoder, except that the convolutional layers are replaced with transposed convolutional layers to achieve feature map upsampling. Finally, the output module reconstructs the features into images, representing the generated acoustic TL fields.
Since the latent distribution learned by the VAE network exhibited discrepancies with the EOF coefficients distribution, this study used the additional incorporation of a normalizing flow network to further refine the distribution fitting. In recent years, normalizing flows have evolved significantly, giving rise to a series of influential architectures, including NICE [23], RealNVP [20], Glow [24], and TARFLOW [25]. Building upon these advancements, we adopted the fundamental architecture of RealNVP to construct the Flow network. As illustrated in Figure 7, the network comprises multiple bijectors, each containing a RealNVP module and a permute module.
The RealNVP architecture utilizes affine coupling layers to construct exact bijective mappings through a structured transformation process. In each layer, the input vector is first partitioned along its feature dimension, with one subset undergoing nonlinear transformation via a neural network module composed of dual dense layers and ReLU activation functions that generate both scaling and translation parameters (s and t), while the complementary subset remains unchanged. The subsequent affine recombination of these processed and unprocessed partitions yields two critical mathematical properties: analytically tractable invertibility, due to the explicit transformation scheme, and computational efficiency, stemming from the Jacobian determinant with low complexity. The permute module serves a critical function by systematically interchanging data dimensions during both forward and inverse propagation. This dimensional rearrangement ensures comprehensive processing through the RealNVP’s affine coupling layers, guaranteeing that all data dimensions undergo transformation rather than being limited to specific subsets. Through this mechanism, the designed Flow network directly satisfies the exact density estimation requirements essential for distribution fitting within our proposed framework.
3.3. Training Configurations
3.3.1. Network Training Configuration
Since the training processes of the two networks are mutually independent and require the completion of VAE training before extracting the corresponding latent variables for subsequent Flow network training, their configurations must be differentiated. Specifically, we first constructed the VAE and Flow networks according to Section 3.2, then employed their respective objective functions (as detailed in Section 2.2) as loss functions to quantify the training performance.
For the VAE network, the latent space dimension was set to match the order of EOF coefficients (i.e., 2). Given that the input consisted of RGB images requiring dimensional reduction to the latent space followed by reconstruction, the network contained substantial trainable parameters—specifically 4.9 million in the encoder and 7.2 million in the decoder. Notably, we represented the output data variance in the decoder as a hyperparameter , using , which also underwent optimization with the objective function during training. Previous work by Daibin et al. [16] demonstrated that proper configuration of this hyperparameter is essential for achieving high-quality reconstruction in VAE models. The batch size was specified to be 256, and the initial learning rate was set to , with the rate halved after every 1/5 of the total epochs to mitigate against training instability and avoid local optima. Upon completion of the VAE network training, the training set was projected into the latent space through the encoder to obtain corresponding latent variables, which subsequently served as inputs to the subsequent Flow network.
Since the VAE network had already performed dimensionality reduction on the original TL data, the Flow network processed low-dimensional data during training. Accordingly, the Flow network employs a series of four bijectors, each RealNVP module contains dense layers with 512 neurons, containing 3.1 million training parameters in total. Consistent with the previous training protocol, we implemented a learning rate decay mechanism with an initial rate of , halved after every 2/5 of the total epochs, and a batch size of 64. Considering the Flow network primarily consists of dense layers, an early stopping strategy was employed to mitigate overfitting, wherein training was terminated prematurely if the validation loss failed to improve over 10 consecutive epochs. The Adam optimizer was used for both the VAE and Flow networks.
3.3.2. Computational Implementation
The deep learning training performed on a server with an Intel Xeon Platinum 8370C CPU and an NVIDIA GeForce RTX 4090, using the programming environment Python 3.9 and Tensorflow 2.10.0. Furthermore, the construction of the Flow network was implemented using Tensorflow Probability (TFP), a specialized extension library of Tensorflow, which provides optimized interfaces for efficient implementation of normalizing flows, significantly enhancing both development efficiency and computational performance.
3.4. Evaluation Metrics
To evaluate the training outcomes at each stage and assess the VAE–Flow network’s generalization performance, it was necessary to introduce additional evaluation metrics beyond the loss function. This study employed the structural similarity index measure (SSIM), mean squared error (MSE), and maximum mean discrepancy (MMD) as comprehensive performance metrics. SSIM is a metric from image processing that quantifies the similarity between two images in terms of luminance, contrast, and structural characteristics. It evaluates image similarity by comparing local statistical properties of images:
(14)
where and represent the mean intensities of the ground truth x and reconstructed (or predicted) acoustic TL field , respectively. and denote the variances in x and , while is the covariance between them. and are small constants introduced to prevent division by zero.The MSE metric quantifies numerical accuracy by computing the average squared differences between corresponding elements of two matrices:
(15)
In this study, MSE was mainly employed to evaluate the accuracy of the distribution fitting achieved by the Flow network. In parallel, MMD is a kernel-based metric designed to quantify the distributional discrepancy between two sample sets in a reproducing kernel Hilbert space (RKHS). This metric can be expressed through kernel functions:
(16)
with the present study employing a Gaussian kernel for MMD computation to measure the distributional differences between the ground truth and fitting results.4. Evaluation and Analysis
Based on the established VAE–Flow network architecture and training settings, we first evaluated the performance of both the VAE and Flow networks using the D1 dataset generated in Section 3.1. The D1 dataset was divided into training, validation, and test sets at a ratio of 10:1:1. Initializing the hyperparameter to 1, we trained the VAE network for 3000 epochs. Due to the hyperparameter configuration, the total loss of the VAE tended to approach negative infinity during optimization. To better analyze the convergence behavior, we decomposed the VAE’s objective function into reconstruction loss and KL divergence loss (as discussed in Section 2.2), monitoring their respective evolution throughout the training process.
Figure 8 presents the training and validation loss curves for both components. The reconstruction loss, formulated as MSE, initially exhibited large magnitude errors, and the results after the first 100 epochs are displayed in Figure 8a. The curves demonstrate that both the training and validation losses decreased rapidly during the initial 500 epochs, followed by a gradually slowing convergence. In later training stages, the losses stabilized without apparent overfitting. Notably, the reconstruction loss here represents the summed error across all pixels ( pixels). Although the final reconstruction loss plateaued around 58, this corresponds to an average per-pixel error of merely , indicating the progressively improved reconstruction accuracy of acoustic TL fields as training proceeded. Figure 8b shows the KL divergence loss curves, which depict significant fluctuations during early training phases, reflecting the VAE’s ongoing adjustment of latent variable distributions. As training progressed, the KL divergence loss converged to approximately 7, suggesting the network increasingly prioritized reconstruction quality over strict adherence to the target latent distribution.
To evaluate the reconstruction accuracy and feasibility of the VAE network for acoustic TL fields under varying EOF coefficients, Figure 9 presents a comparison between KRAKEN-simulated TL fields and their VAE reconstructed counterparts, quantifying their similarity through SSIM and normalized pixel-wise error distributions. From an image fidelity perspective, SSIM values closer to 1.0 indicate higher similarity between the compared fields. The results demonstrate that across three randomly selected test samples, the SSIM values consistently exceeded 0.94, confirming the VAE’s capability to efficiently reconstruct TL fields with high structural fidelity. The normalized pixel error distribution reveals that regions with pronounced reconstruction errors exhibited spatial correlation with areas of high acoustic energy fluctuation. To further quantify the reconstruction performance, we computed the SSIM between all ground-truth and reconstructed acoustic TL samples in both training and test sets, with the results presented as box plots in Figure 10. The SSIM results reveal that the VAE’s reconstruction accuracy on the test set was statistically consistent with its performance on the training set. The median SSIM values reached 0.964 for both datasets, with tightly clustered distributions. The result demonstrates the VAE network’s strong generalization capability for dataset D1.
As theoretically discussed in Section 2.2, when data reside on a low-dimensional manifold within a high-dimensional space, VAEs often struggle to accurately estimate the ground-truth probability distribution. This theoretical limitation was empirically validated through visualization of the latent space distributions. Figure 11a and Figure 11b, respectively, show a comparison between the distribution of latent variables obtained by mapping the training samples of D1 through the encoder and the ground-truth EOF coefficient distribution in the space. The comparison reveals that while the learned latent distribution approximated a zero-mean Gaussian, it exhibited significant divergence from the EOF coefficient distribution. Specifically, the latent variables followed an isotropic Gaussian, with substantially larger variance than the target distribution and ; therefore, direct sampling from this latent space yielded meaningless noise, which lacked a corresponding relationship with the actual physical parameters, due to the distributional mismatch. Overall, these results indicate that although the VAE successfully projected acoustic TL data onto a 2D approximate Gaussian distribution, it failed to estimate the true probability measure governing the TL data distribution on this manifold. This fundamental limitation motivated our proposed integration of a Flow network, to achieve precise distributional alignment between the learned latent variables and the EOF coefficients.
Consequently, we first extracted the latent variables encoded from both the training and validation sets of dataset D1, maintaining the original data partitioning ratio to serve as the training and validation sets for the Flow network. The target distribution for the network was configured as , where the mean vector and covariance matrix ∑ were derived from the EOF coefficient distribution of dataset D1:
(17)
Given the low-dimensional nature of the latent data and its concentrated distribution in this stage, the Flow network was trained for 100 training epochs, with the corresponding loss curves presented in Figure 12. Figure 12 demonstrates that the early stopping mechanism was triggered at the 97th epoch during the Flow network training. As shown from the curves, both the training and validation loss had stabilized at this stage, exhibiting neither a significant decrease nor signs of overfitting. Upon completion of training, we fed the extracted latent variables into the Flow network and obtained the fitted distribution through inverse mapping. Figure 13a,b present a comparative visualization between this fitted distribution and the EOF coefficient distribution along each dimension. To further compare our work with the performance of the Two-Stage VAE, we additionally trained a simplified VAE network under identical parametric constraints. To ensure consistency, both the encoder and decoder in the simplified VAE network were constructed exclusively with fully connected layers (512 neurons), while the dimensions of the input and latent space were fixed at 2. Complementarily, an identical learning rate scheduling to the Flow network was implemented and the simplified VAE was also trained for 100 epochs. The resulting distributions in the first dimension and the second dimension are presented in Figure 13c and Figure 13d, respectively.
Figure 13 presents the probability density functions of the fitted distributions from both the Flow network and simplified VAE network against the EOF coefficient distribution in the first dimension and second dimension. The results demonstrate that the Flow network achieved precise mapping from latent variables to EOF coefficients, with the first dimension fitted distribution showing excellent consistency with , and the second dimension fitted distribution closely matching . In contrast, the simplified VAE exhibited considerable deviations in both dimensions, particularly manifesting as substantially overestimated variances compared to the true distribution. The quantitative results of sample-wise differences, as illustrated in Figure 14, demonstrate that the Flow network achieved consistently lower MSE values relative to the ground truth across both dimensions compared to the simplified VAE, and exhibited tighter distributions, with minimal outliers. The MMD metric computed simultaneously provided further confirmation, showing a remarkable improvement, from the pre-fitting MMD of between latent variables and EOF coefficients to the post-fitting MMD of for the Flow network, a performance gain two orders of magnitude greater than the simplified VAE’s result. These comprehensive metrics collectively validate that the Flow network not only enabled precise distributional mapping but also established the hybrid VAE–Flow framework’s definitive advantage over the VAE-only or Two-Stage VAE approaches. Notably, when the distribution fitting yielded poor results, it became challenging to establish a reliable mapping relationship, which consequently compromised the controllability of the generation process in the subsequent predictions.
When the training was completed, the integrated VAE–Flow network enabled rapid prediction of acoustic TL fields. First, the EOF coefficients corresponding to the test set of D1 were fed into the Flow network, where they underwent forward mapping transformation into the VAE’s latent representation. Then, these latent variables were subsequently processed though the VAE decoder to generate the corresponding TL fields. At this stage, the VAE–Flow network constituted a complete acoustic TL prediction system, capable of generating accurate TL field predictions through efficient end-to-end mapping from two EOF coefficient inputs. Figure 15 presents a comparison between the acoustic TL prediction results generated by the VAE–Flow network for randomly selected EOF coefficients from the test set and the counterparts from KRAKEN. The SSIM values consistently exceeded 0.94, which demonstrates that the trained VAE–Flow network successfully established an accurate mapping from EOF coefficients to acoustic TL fields , thereby validating the capability of the proposed framework combining a VAE and a normalizing flow for controllable generation in shallow-water acoustic propagation prediction.
To further evaluate the predictive performance of the VAE–Flow network across diverse datasets, specifically the six datasets constructed in Section 3.1, we quantified the acoustic TL fields reconstructed by the VAE network and predicted by the VAE–Flow network against KRAKEN simulations using the SSIM, with the results visualized in Figure 16a and Figure 16b, respectively. Notably, during training, only the VAE’s latent space prior distribution and the Flow network’s target distribution were adapted to each dataset’s EOF coefficient statistics, while other architectural hyperparameters were kept identical, to ensure a controlled comparison. During data generation, the variance in the Gaussian distribution governing the EOF coefficients in datasets D1-D6 was set to be progressively increased. The results revealed that as the intra-class variance of the training samples grew, the training difficulty gradually escalated, leading to a corresponding increase in error in both the reconstructed and predicted acoustic TL results. This is evidenced by the progressively declining SSIM values in Figure 16a,b. Notably, since the variance in dataset D6 was significantly higher than in other datasets, the median SSIM values between the reconstructed/predicted acoustic TL fields and the ground truth were substantially lower compared to the other datasets. The analysis revealed that the predictive performance of the VAE–Flow network was markedly inferior on datasets with larger distribution variances than small variances. This disparity can be attributed to the fact that, given the same sample size, data points drawn from a distribution with higher variance tended to be more dispersed, making it challenging to adequately capture the statistical characteristics of the distribution. As a result, the network failed to effectively learn the underlying patterns of the data during training, which ultimately led to suboptimal predictive performance.
To mitigate against this limitation, we augmented the training dataset D6 by doubling its number of samples and subsequently retrained the VAE–Flow network. A comparison between the results from the augmented dataset and the original dataset is presented in Figure 16c,d. The box plots illustrate that an increased sample size enhanced the network’s ability to learn the underlying data patterns. Specifically, when the training dataset size of dataset D6 was doubled, the reconstruction and prediction performance of acoustic TL fields showed measurable improvement. Therefore, for datasets with larger distribution variances, such an adjustment is crucial to enable the network to effectively capture the intrinsic patterns of the inputs, thereby optimizing its predictive performance.
Figure 17 illustrates a comparison of the computational time between the KRAKEN model and the VAE–Flow network for acoustic TL prediction under different sample sizes, plotted on a logarithmic scale. The trained VAE–Flow network leveraged its data-driven advantage to rapidly accomplish the task of prediction of acoustic TL fields, requiring only computation times on the order of milliseconds for small-scale sample generation. This computational efficiency advantage became more pronounced when generating large-scale samples. For instance, when producing 1000 samples, the VAE–Flow network required only 4 s, whereas the KRAKEN model necessitated 1000 repetitive computations, resulting in a total computation time exceeding 500 s. Overall, these results demonstrate that the VAE–Flow network exhibited significantly superior computational efficiency compared to the conventional KRAKEN model, thereby providing a feasible solution for rapid prediction of shallow-water acoustic transmission loss.
5. Conclusions
In this paper, a hybrid framework that integrates a variational autoencoder with a normalizing flow through a two-stage training strategy was proposed. Using a VAE to learn latent representations of TL data on a low-dimensional manifold, and a normalizing flow to establish a bijective mapping between the latent variables and EOF coefficients, this approach enables an end-to-end mapping from SSPs to TL, thereby addressing the controllability limitation of deep generative augmentation methods and enhancing prediction efficiency. Through EOF decomposition and reconstruction of SSPs, six datasets were constructed based on the KRAKEN model. After conducting evaluations and analyses for both the VAE and Flow networks, the VAE network demonstrated the capability to efficiently reconstruct acoustic TL fields with median SSIM values of 0.964 for both the training and validation datasets, thereby exhibiting a strong generalization ability. Although the VAE network successfully learned a latent distribution of the acoustic TL fields, it failed to estimate the ground-truth probability measure within the manifold, as evidenced by the significant divergence in the learned latent distribution compared to the EOF coefficient distribution. The introduction of the Flow network effectively mitigated this discrepancy, reducing the MMD from to , demonstrating its exceptional capability for distribution fitting compared to a simplified VAE and highlighting the superior ability of the VAE–Flow network for achieving controllable generation compared to the Two-Stage VAE. The VAE–Flow network thereby exhibited a higher computational efficiency than the KRAKEN model, with only 4 s for generating 1000 samples, while maintaining high accuracy.
The evaluation and analysis in this paper were based on data simulated in KRAKEN. Although we introduced measurement errors in sound speed profiles (0–1.0 m/s) during testing and verified the robustness of the proposed method against these perturbations, more complex training data and experimental data are needed in the future to verify the effectiveness of our method. To address the performance degradation under high-variance conditions, future work will incorporate regularization techniques to investigate controllable generation in datasets with significant variance. The method for rapid TL prediction presented in this paper could be further refined and extended to more complex waveguide scenarios in the future, such as sound speed profiles with seasonal variations, complex seafloor topography, and diverse source conditions.
Conceptualization, B.S. and H.W.; methodology, B.S.; software, B.S.; validation, B.S., H.W. and X.Z.; formal analysis, B.S.; investigation, H.W.; resources, B.S. and H.W.; data curation, H.W.; writing—original draft preparation, B.S.; writing—review and editing, B.S. and H.W.; visualization, H.W.; supervision, X.Z., P.S. and X.L.; project administration, X.Z., P.S. and X.L.; funding acquisition, H.W. and X.L. All authors have read and agreed to the published version of the manuscript.
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
The authors thank Shoudong Wang, Yan Lv, and Yangjin Xu for their insightful discussions and instructive guidance.
The authors declare no conflicts of interest.
The following abbreviations are used in this manuscript:
| TL | transmission loss |
| VAE | variational autoencoder |
| Flow | normalizing flow |
| SSIM | structural similarity index measure |
| CIRs | channel impulse responses |
| GANs | generative adversarial networks |
| CGAN | conditional generative adversarial network |
| CRAN | convolutional recurrent autoencoder network |
| SSPs | sound speed profiles |
| VAE–Flow | variational autoencoder-normalizing flow |
| MLE | maximum likelihood estimation |
| KL | Kullback–Leibler |
| ELBO | Evidence Lower Bound |
| EOF | empirical orthogonal functions |
| RHS | right-hand side of an equation |
| iff | if and only if |
| TFP | Tensorflow Probability |
| MMD | maximum mean discrepancy |
| RKHS | Reproducing Kernel Hilbert Space |
| MSE | mean squared error |
| MMD | maximum mean discrepancy |
| Variables for Underwater Acoustics | |
| Angular frequency of acoustic source (rad/s) | |
| Source depth (m) | |
| Receiver depth (m) | |
| r | Horizontal distance between source and receiver (m) |
| Medium density (kg/m3) | |
| Wavenumber of m-th normal mode (rad/m) | |
| Depth function of m-th normal mode (dimensionless) | |
| Acoustic pressure field at position | |
| Free-space acoustic pressure (Pa) | |
| Transmission loss at receiver location (dB) | |
| Physical parameter vector (sound speed, seabed properties, etc.) | |
| Receiver region in 2D space (m2) | |
| N | Number of spatial points in discretized TL field |
| Variables for Deep Learning | |
| Latent variables (low-dimensional representation) | |
| Low-dimensional manifold of TL data | |
| Probability measure on manifold | |
| Encoder/recognition model | |
| Decoder/generative model | |
| Evidence Lower Bound (ELBO) | |
| Kullback–Leibler divergence | |
| Random noise variable in reparameterization trick | |
| Jacobian matrix of transformations | |
| Decoder output variance hyperparameter | |
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 1 Schematic diagram of a low-dimensional manifold embedded in high-dimensional space.
Figure 2 Hybrid VAE–Flow framework through two-stage training strategy for underwater acoustic TL prediction.
Figure 3 Schematic diagram of a shallow-water waveguide with negative thermocline.
Figure 4 Flowchart for sound speed profile generation via empirical orthogonal decomposition.
Figure 5 Average sound speed profile and the first two EOF modes. (a) Average SSP. (b) Spatial patterns of the first two EOF modes.
Figure 6 Schematic diagram of VAE network architecture.
Figure 7 Schematic diagram of Flow network architecture.
Figure 8 Training and validation curves of the VAE network. (a) Reconstruction loss. (b) KL loss.
Figure 9 Comparison between the TL results of the KRAKEN and the VAE network under different test EOF coefficients.
Figure 10 SSIM of acoustic TL fields reconstructed by the VAE network against KRAKEN simulations on the training and test sets.
Figure 11 Comparison between distributions in 2D space. (a) Latent variables. (b) EOF coefficients.
Figure 12 Training and validation curves of the Flow network.
Figure 13 The probability density of the fitted distributions against the ground truth. (a) Flow fitting results in Dimension 1. (b) Flow fitting results in Dimension 2. (c) Simplified VAE fitting results in Dimension 1. (d) Simplified VAE fitting results in Dimension 2.
Figure 14 Comparison of performance between Flow network and simplified VAE network on MSE. (a) The first dimension. (b) The second dimension.
Figure 15 Comparison between the TL results of the KRAKEN and the VAE–Flow network under different EOF test coefficients.
Figure 16 SSIM of acoustic TL fields reconstructed by the VAE network against KRAKEN simulations, and predicted by the VAE–Flow network against KRAKEN simulations on distinct datasets. (a) Reconstructed SSIM results on D1–D6. (b) Predicted SSIM results on D1–D6. (c) Reconstructed SSIM results between original and augmented D6. (d) Predicted SSIM results between original and augmented D6.
Figure 17 Comparison of the computational time of the VAE–Flow network and KRAKEN model.
Sampling distribution of the six datasets.
| Dataset | | |
|---|---|---|
| D1 | | |
| D2 | | |
| D3 | | |
| D4 | | |
| D5 | | |
| D6 | | |
1. Porter, M.B. The Kraken Normal Mode Program; Technical report Naval Research Laboratory: Washington, DC, USA, 1992.
2. Porter, M.B. Bellhop: A Beam/Ray Trace Code, Version 2010-1. 2010; Available online: https://oalib-acoustics.org/website_resources/AcousticsToolbox/Bellhop-2010-1.pdf (accessed on 7 March 2025).
3. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. Proceedings of the 2nd International Conference on Learning Representations, ICLR 2014; Banff, AB, Canada, 14–16 April 2014.
4. Wei, L.; Wang, Z. A Variational Auto-Encoder Model for Underwater Acoustic Channels. Proceedings of the 15th International Conference on Underwater Networks & Systems (WUWNet ’21); Shenzhen, China, 21–24 November 2021.
5. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Proceedings of the 28th International Conference on Neural Information Processing Systems (NIPS’14); Montreal, QC, Canada, 8–13 December 2014; MIT Press: Cambridge, MA, USA, 2014; Volume 2, pp. 2672-2680.
6. Liu, J.; Zhu, G.; Yin, J. Joint color spectrum and conditional generative adversarial network processing for underwater acoustic source ranging. Appl. Acoust.; 2021; 182, 108244. [DOI: https://dx.doi.org/10.1016/j.apacoust.2021.108244]
7. C., S.C.; Kamal, S.; Mujeeb, A.; M.H., S. Generative adversarial learning for improved data efficiency in underwater target classification. Eng. Sci. Technol. Int. J.; 2022; 30, 101043. [DOI: https://dx.doi.org/10.1016/j.jestch.2021.07.006]
8. Wang, Z.; Liu, L.; Wang, C.; Deng, J.; Zhang, K.; Yang, Y.; Zhou, J. Data Enhancement of Underwater High-Speed Vehicle Echo Signals Based on Improved Generative Adversarial Networks. Electronics; 2022; 11, 2310. [DOI: https://dx.doi.org/10.3390/electronics11152310]
9. Zhou, M.; Wang, J.; Feng, X.; Sun, H.; Li, J.; Kuai, X. On Generative-Adversarial-Network-Based Underwater Acoustic Noise Modeling. IEEE Trans. Veh. Technol.; 2021; 70, pp. 9555-9559. [DOI: https://dx.doi.org/10.1109/TVT.2021.3102302]
10. Varon, A.; Mars, J.; Bonnel, J. Approximation of modal wavenumbers and group speeds in an oceanic waveguide using a neural network. JASA Express Lett.; 2023; 3, 066003. [DOI: https://dx.doi.org/10.1121/10.0019704] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/37306565]
11. Mallik, W.; Jaiman, R.K.; Jelovica, J. Predicting transmission loss in underwater acoustics using convolutional recurrent autoencoder network. J. Acoust. Soc. Am.; 2022; 152, pp. 1627-1638. [DOI: https://dx.doi.org/10.1121/10.0013894]
12. Sun, Z.; Wang, Y.; Liu, W. End-to-end underwater acoustic transmission loss prediction with adaptive multi-scale dilated network. J. Acoust. Soc. Am.; 2025; 157, pp. 382-395. [DOI: https://dx.doi.org/10.1121/10.0034857] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/39835828]
13. Jensen, F.B.; Kuperman, W.A.; Porter, M.B.; Schmidt, H. Computational Ocean Acoustics; 2nd ed. Springer Publishing Company, Incorporated: New York, NY, USA, 2011.
14. Zhang, X.; Wang, P.; Wang, N. Nonlinear dimensionality reduction for the acoustic field measured by a linear sensor array. MATEC Web Conf.; 2019; 283, 07009. [DOI: https://dx.doi.org/10.1051/matecconf/201928307009]
15. Ray, D.; Pinti, O.; Oberai, A.A. Generative Deep Learning. Deep Learning and Computational Physics; Springer Nature: Cham, Switzerland, 2024; pp. 121-146.
16. Dai, B.; Wipf, D.P. Diagnosing and Enhancing VAE Models. Proceedings of the 7th International Conference on Learning Representations, ICLR 2019; New Orleans, LA, USA, 6–9 May 2019.
17. Loaiza-Ganem, G.; Ross, B.L.; Cresswell, J.C.; Caterini, A.L. Diagnosing and Fixing Manifold Overfitting in Deep Generative Models. arXiv; 2022; arXiv: 2204.07172
18. Loaiza-Ganem, G.; Ross, B.L.; Hosseinzadeh, R.; Caterini, A.L.; Cresswell, J.C. Deep Generative Models through the Lens of the Manifold Hypothesis: A Survey and New Connections. arXiv; 2024; arXiv: 2404.02954
19. Rezende, D.J.; Mohamed, S. Variational inference with normalizing flows. Proceedings of the 32nd International Conference on International Conference on Machine Learning (ICML’15); Lille, France, 6–11 July 2015; Volume 37, pp. 1530-1538.
20. Dinh, L.; Sohl-Dickstein, J.; Bengio, S. Density estimation using Real NVP. Proceedings of the 5th International Conference on Learning Representations, ICLR 2017; Toulon, France, 24–26 April 2017.
21. Su, J.; Wu, G. f-VAEs: Improve VAEs with Conditional Flows. arXiv; 2018; arXiv: 1809.05861
22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR); Las Vegas, NV, USA, 27–30 June 2016; pp. 770-778.
23. Dinh, L.; Krueger, D.; Bengio, Y. NICE: Non-linear Independent Components Estimation. Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015; San Diego, CA, USA, 7–9 May 2015.
24. Kingma, D.P.; Dhariwal, P. Glow: Generative Flow with Invertible 1x1 Convolutions. Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems 2018, NeurIPS 2018; Montréal, QC, Canada, 3–8 December 2018; pp. 10236-10245.
25. Zhai, S.; Zhang, R.; Nakkiran, P.; Berthelot, D.; Gu, J.; Zheng, H.; Chen, T.; Bautista, M.Á.; Jaitly, N.; Susskind, J.M. Normalizing Flows are Capable Generative Models. arXiv; 2024; arXiv: 2412.06329
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.