1. Introduction
Fiber-reinforced polymer composites (FRPCs), which have outstanding material qualities such as high stiffness, strength, and strength-to-weight ratio, are favorable for applications and have gained a lot of attention from both academia and industries [1]. Manufacturing techniques of FRPCs stand for a multidisciplinary subject that requires proper design rules and topology optimizations [2].
Topology optimization is a powerful tool for helping engineers design innovative structures and products, and has already found successful applications in various energy fields [3,4]. In knowledge-driven theories, minimizing the mean compliance and elastic strain of the energy of a structure is the main objective function for the majority of structural topology optimization problems [5]. Entropy generation is often additionally used to assess topology optimization for high-performance structures in aeronautical engineering and automotive systems [6]. Greater efforts are being put into improving the robustness and effectiveness of optimization algorithms, and their applicability has been confirmed [7,8]. With regard to the topology optimization problems, numerous knowledge-driven strategies have been developed to achieve the optimal structure under given load and boundary conditions, manufacturing restraints, and performance objectives. Such models include the solid isotropic material with penalization (SIMP) [9], the level set [10], bi-directional evolutionary structural optimization (BESO) [11], and moving morphable components (MMC) [12,13].
The SIMP method predicts an optimal material distribution within a given design space, and has been considered the most widely used topology optimization scheme to explore optimal structures [14]. The SIMP method stands for one of the typical mesh-based approaches. The reference domain should first be discretized by creating a finite element mesh, followed by a finite element analysis (FEA) performed to obtain the calculation of the compliance for a topology configuration. The traditional methods devoted to topology optimization issues require considerably fine meshes to produce high-resolution designs; as a result, the FEA calculation can be difficult and time-consuming to implement. The cost of FEA computing and evaluating a compliance-based objective function under the volume constraint increases with the complexity (finer) of the finite element discretization.
A high-performance GUP solver is an efficient answer to low computing needs without sacrificing the quality of the structures, especially for large 3D design problems [15]. In addition, many researchers are increasingly interested in applying machine learning techniques to assist optimization procedures of compliance minimization. As numerical modeling naturally complements the ML technique by providing enormous amounts of relevant data, data creation is no longer a challenge in developing effective and robust data-driven models [16,17]. Ref. [18] suggested a K-means algorithm to reduce the dimension of the design variables, thus shortening the computational time. Ref. [19] established an implicit mapping model of shape function between the high resolution and a coarse resolution to reduce the FEA calculation significantly.
There have also been widespread attempts to speed the process of the convergence of the optimization by using powerful neural network (NN) models [20]. Convolutional neural networks (CNNs), a family of artificial neural networks that have gained dominance in a variety of computer vision tasks [21,22], are the most well-established deep learning algorithms in topology optimization tasks. Thus, the topology optimization problem is transformed from a design challenge to an image recognition problem. For the optimized topology configurations, researchers have proposed replacing the FEA calculation module with CNN models [23,24]. Furthermore, CNN models have been combined with generative adversarial network (GAN) algorithms [25] and long short-term memory networks (LSTM) algorithms [17,26] to accelerate the iteration procedures of the SIMP algorithm. The image-based CNN model achieves high accuracy in mapping the topology configuration to its compliance, while the training from scratch is usually time-consuming, and the black-box nature hinders its applicability [22].
In the fields of image processing and computer vision, the geometric features of an image can also be characterized using a set of measurable quantities known as moment invariants [27]. The moment invariants reflect the inherent properties of rotation, translation, and scale invariance. This calculation finds widespread use in tasks such as object recognition, tracking, and image feature extraction. Using invariants as features instead of the image itself, the selection of artificial neural networks can be flexible to replace the CNN model [24,28].
The radial basis function (RBF) has emerged as a variant of the artificial neural network, showing the principal advantages of easy design, good generalization, and faster training. GRNN [29] is one type of RBF, and it showcases the capability of quickly learning and rapidly converging to the optimal regression surface with a large number of datasets [30]. It has been demonstrated that the GRNN algorithm can more efficiently extract valuable information from images while reducing the training cost compared to the CNN models [28].
In this study, towards elucidating the correlation between compliance and optimized topology configurations, we propose a novel deep learning model based on moment invariant analysis of an efficient GRNN neural network. The paper is outlined as follows. Section 2 discusses the design problem and the existing SIMP method, and elaborates on mathematical calculations of moment invariants, along with the architecture of GRNN used in this study. To determine the suitable deep learning model, this paper compares the compliance prediction performance of the GRNN, DNN, and CNN algorithms. The SIMP simulation results are utilized for dataset generation to train the machine learning models. Section 3 evaluates the prediction model to justify the accuracy, efficiency, and reliability, with evaluations and comparison results with existing models presented. Section 4 gives concluding remarks and explains the limitations of the current framework, and future outlooks are also given.
2. Method
Figure 1 presents the flowchart of this work, which starts with the utilization of the SIMP method for data creation of structural topology configurations for the design problem. The images are processed to extract their moment invariants, which are then fed into GRNN models as features. In addition to the GRNN model, the CNN and DNN models are trained and optimized for the same design problem. In the last step, the machine learning models are assessed for the performance of accuracy, efficiency, and reliability.
2.1. The SIMP Method and Data Preparation
2.1.1. The SIMP Method
As a minimization compliance problem with combinations of variables, the SIMP procedure starts with the discretization of a given domain into a grid of finite elements. A binary value of is assigned for each element, which is written as:
(1)
The value of describes an element that is either filled with material for regions that require material or emptied of material for regions where material can be removed (representing voids).The minimization function is subjected to the equilibrium equation and material volume fraction of the optimization material to the original design volume, with consideration of the design variable constraints. The function can be defined as:
(2)
(3)
(4)
where c is a compliance value, p is the penalty factor, and are lower bounds on densities, which are introduced to refrain from the singularity. U and F stand for the global displacement and force vectors, respectively. Here, the notation K is the global stiffness matrix; and denote the displacement vector and stiffness matrix of each element. is the volume of each element, is the target material volume fraction, and N is the total number of elements in the design domain , which also denotes the resolution of the topology image.The implementation of SIMP involves multiple iterations. Each iteration produces a topology image, in accordance with four main modules performed: finite element analysis (FEA), analyzing sensitivities (Equations (2)–(4)), filtering sensitivities, and updating design variables, which produce topology images accordingly.
We analyze the time consumption in the SIMP method using the Python performance profiling tool Line profiler [31], as shown in Table 1. The higher computation effort is associated with the modules of FEA calculation and advanced filter techniques, especially when a refined mesh strategy is applied for describing the structural geometry with high resolution.
2.1.2. Data Generation via SIMP
We employ the SIMP algorithm to generate a dataset. As illustrated in Figure 2, two topology optimization tasks are considered, namely the simply supported beam in Figure 2a and the cantilever beam in Figure 2b.
The design domain has a dimension of , with load F as a concentrated force. For each task, the loading position of [, ] is defined as the distance between the load node and the center of the design domain within the Cartesian coordinates, and material volume fraction comprises the primary constraints. For the simply supported beam, the load F can be applied to the nodes of the upper and lower surfaces, while the load of a cantilever beam can act on the surface of the free end. The material volume fraction follows a normal distribution , with a minimum value of 0.176 and a maximum value of 0.802.
As displayed in Figure 2b,d, toward minimizing compliance and obtaining the optimal topology configuration, the implementation of the SIMP algorithm updates the image in response to the FEA calculation in each iteration.
It is noticed that the 40 iterations record the topology images from blurry configurations to the convergence images; thus, we collect the FEA-determined compliances and their topology images at the first 40 iterations to construct the dataset. A similar strategy has been also addressed in [32] for the CNN model training.
In accordance with the SIMP implementations, a dataset is obtained with 20,000 images for the two tasks, in which the dataset’s completeness and diversity are guaranteed.
Figure A1 illustrates the image distributions with considered constraints. Of those, and are designated for training and testing, respectively.
2.2. Moment Invariants
The fundamental of moment invariants is to extract image information using a set of measurable quantities known as invariants. The geometric moment m is the most basic type of moment invariant.
For a two-dimensional image, the order geometric moment is defined as follows:
(5)
where (x, y) represents the coordinates of each pixel in the image within a Cartesian coordinate system, and is the density distribution function. Geometric moments can reflect numerous geometric features of the image, capturing the structural characteristics of the topology image. For example, the zero-order geometric moment describes the “mass” of the image, which is the area of a binary image. Similar to the “moment of inertia” in material mechanics, and are the centroid of the image; meanwhile, the second-order moments of and describe the mass distribution of the image.Furthermore, the character of shape-preserving transformations like translation, scaling, rotation, and mirroring can be preserved by image invariants. The widely used Hu moments [27] consist of seven independent variables and are derived using only the first three-order geometric moments, which may not be sufficient for extracting enough structural information from topology images.
This study adopts the method derived from [33], which provides a set of invariant moments M of arbitrary order that is complete and independent. The calculation of invariant moments M can be derived using the complex moment , which has the following expression:
(6)
The moment invariants M of order r are derived from all the complex moments of order , where :
(7)
(8)
(9)
(10)
where B is the complex form base set of r-order moment invariants M in Equation (7). According to Equations (8) and (10), extracting the real and unconjugated imaginary parts of all elements in set B constitutes the final representation of the r-order invariant moment M set.Among them, and are indicators that satisfy the conditions for any allowed image. The base set defined by Equation (7) to Equation (10) depends on the selection of and . In practical applications, on the one hand, we hope to keep and as small as possible, because compared to higher-order moments, lower-order moments are not sensitive to noise. In addition, approaching 0 will make the value of the invariant unstable.
In this study for dealing with topology images, based on the algorithm proposed by [33], taking , , and , the set B containing seven complex elements is calculated using Equation (7). Then, according to Equation (8), we take the real part of all elements in set B, and take the imaginary part of some of them according to Equation (9), to obtain a four-order moment invariants base set M containing a total of 11 elements, as described in Equation (11):
(11)
2.3. Generalized Regression Neural Network (GRNN)
As in Figure 3, the GRNN model has four layers, which are the input layer, pattern layer, summation layer, and output layer. The fundamental principle of GRNN is nonlinear regression analysis [29], and the mathematical formula can be written as:
(12)
where is the weighted average of all sample observations , and the weight factor of each observation is the exponent of the squared Euclid distance between the corresponding sample and x. is the key parameter of the smoothing factor. When has a large magnitude, the prediction approximates the mean of all sample dependent variables, whereas approximating 0 results in a prediction that is very near to the training sample.2.4. Data Processing and Model Training
To obtain a reliable machine-learning-based prediction model, the three neural network models are designed to absorb the images to produce the corresponding compliances. The normalization technique [22] is applied first to process the input features and the predicted compliance.
For compliance prediction tasks, the GRNN model is designed in Figure 3. The topology images are processed to calculate the fourth-order moment invariants as input features using Equation (5) to Equation (10). In addition to the 11-moment invariants, the position of the loading position of and material volume fraction are also chosen as the input feature of GRNN.
The input–output pairs for the DNN and CNN model, with the model structure, are detailed in Figure A2. It can be learned that the layer structure of the DNN model is different from the GRNN model, but it is intended to use the same input information as the GRNN model. The CNN model is designed to automatically and adaptively learn topology image information and to predict the compliance, with which the layer structure, parameters, and optimization techniques are used from [23].
The platform for creating datasets is configured with topy 0.4.0 in Python2.7. The network training platform is configured with pyGRNN 0.1.2, scikit-learn 1.0.2, CUDA 11.6, and torch 1.13.1+cu116 with Nvidia driver version 512.36 (GPU: Nvidia RTX3060).
3. Results and Discussion
3.1. Moment Invariant Calculations
The moment invariant is a feature extraction technique used to extract the global features for shape recognition and identification analysis. Figure 4 gives several classic topology configurations obtained from the dataset. The calculated 11-moment invariants for each case are summarized in Table 2. The 11 independent variables can cover the structural information of translation, scaling, and rotational under a set of loading conditions and material volume constraints.
For developing a deep learning-based image recognition algorithm, the moment invariants are much more efficient compared to the traditional image-processing techniques in CNN models.
3.2. Model Performance Evaluation
3.2.1. Model Accuracy
For regression problems, the loss functions mean squared error (MSE) and coefficient of determination are used to evaluate the model’s performance.
The MSE and can be computed as:
(13)
(14)
where is the predicted values of compliances, is the ground-truth, and is the average value of all compliances.Since the model is trained on the dataset of the simply supported beam and the cantilever beam, Figure 5 first evaluates the performance of the GRNN model on the two tasks within the dataset. For the simply supported beam, the GRNN model achieves 0.981 and 0.970 on the training and testing datasets, respectively. The magnitudes are 0.992 and 0.970 on the cantilever beam.
Figure 6 offers a better insight into the learning performance of the DNN and CNN models by displaying graphs of loss and accuracy over each epoch for training and testing operations. It can be found that compared to the DNN model, the CNN model presents as a very powerful and efficient model that performs automatic feature extraction to achieve high accuracy in predicting topology compliance.
3.2.2. Model Efficiency
The finalized model parameters for GRNN are summarized in Table 3. Additionally, the parameter settings for DNN and CNN models are summarized in Table A3 and Table A4 in Appendix B, respectively.
Table 4 summarizes the prediction accuracy on training and testing sets and the computational efficiency of the three neural networks. In terms of prediction accuracy, it can be shown that all three models perform quite well with the higher than 0.96, with the GRNN model reaching a high precision of 0.998 and 0.994 on the training and testing datasets, respectively. The magnitudes are marginally lower than the CNN model, which achieves an of 0.999.
As for computational cost, the GRNN significantly overwhelms the other two neural network models. In particular, the training completion time for the GRNN algorithm is 1191.35 s, which is 30 times faster than the CNN model’s (38,662.40 s) time. Moreover, the well-trained GRNN model predicts topology compliances in 0.0019 s, which is a significant improvement over the CNN model’s 2.71 s. In comparison to the GRNN model, the DNN model achieves equivalent accuracy while having a training cost that is roughly 7 times higher and a prediction time that is 3 times slower. Overall, the GRNN model excels at obtaining high precision in predicting the compliances with the lowest computational cost.
It is evident that the well-known CNN model is a very effective tool for automatically extracting features for dealing with image recognition and segmentation problems, but it always comes with a significant training burden. As seen in Table A4, training a CNN model necessitates thousands of parameters. On the one hand, this raises the expense of training, and on the other, it turns the model into a black box with limited interpretability [22]. In contrast, the DNN model in Table A3 has a simpler layer structure and fewer parameters, which reduces training costs but makes certain accuracy tradeoffs.
Regarding the GRNN model in Table 3, it holds the simplest layer structure, with only one hyperparameter (), which is the major advantage over other types of NN models. Note that we explore a range of values before settling on the magnitude of 0.023, as in Table 3. Meanwhile, it maintains the highest accuracy attributed to the moment invariant analysis. With a large number of datasets, the GRNN model also has the intrinsic capacity to learn rapidly and quickly converge to the best parameter; therefore, it offers the lowest training cost and produces predictions quickly.
3.3. Model Generalizability
Using the SIMP algorithm, varied structural topology configurations are generated as unknown samples to examine the generalizability of the trained GRNN model.
3.3.1. Generalizability on Rotated Topology Configurations
As presented in Table 5 and Table 6, for the samples of simply supported beams and cantilever beams, the compliances predicted from the three neural network models are compared with the FEA-determined ground-truth values.
For each task, we select four configurations (specified as images []) that involve diverse constraints information of loading positions, material volume fractions, and iteration steps. Accordingly, for the simply supported beams, the FEA-determined structural compliances are 12.086, 23.871, 16.116, and 20.538, with the constraint details being found in Table A1 in the Appendix A. For the cantilever beams, the FEA-determined compliances are 213.518, 532.576, 228.132, and 168.211, with the constraints listed in Table A2 in the Appendix A.
Firstly, the three models demonstrate comparable generalization capability on the topology configurations (images []), while the GRNN model exhibits the best performance with a maximum relative error of less than 3.5%. Furthermore, when rotating those images clockwise 90, 180, and 270 degrees, it is worth noting that the time-consuming CNN model fails to learn the geometry transformations with a prediction error greater than 91%. In contrast, for the GRNN and DNN models, the rotation-invariance property based on the moment invariant analysis enables the network’s capability to predict the compliances regardless of the orientations. Thus, the predicted outcomes are the same compared to the configurations before rotation (images []).
Figure 7 further explores the model predictions along with iterations from 0 to 40. It can be seen that the developed GRNN model accurately captures the nature of minimization of compliance along the iterations. The comparison demonstrates another advantage of using moment invariants to extract structural information and then employing GRNN for predicting compliances. When the topological configurations are rotated, it appears that the CNN model entirely loses its ability to predict. However, GRNN may still sustain a comparatively high level of prediction accuracy owing to the rotational invariance of the moments.
3.3.2. Generalizability on Topology Configurations with Different Material Volume Fractions
It is known that the topology optimization follows compliance minimization criteria when subject to mechanical constraints of a certain load and material volume fraction. There are a handful of studies that explain that an increased material volume fraction produces a decreased compliance magnitude [34].
Figure 8 plots the predicted compliance in response to changed material volume fraction for the two topology optimization tasks. It can be found that the underlying mechanism of compliance variation and the input feature of material volume fraction is accurately learned by the GRNN model. The GRNN model exhibits outstanding generalization ability for predicting compliance in response to changed material volume fraction.
4. Conclusions and Outlook
In this work, a novel GRNN algorithm is developed to predict the topology compliance, together with the moment invariants as independent variables, which are obtained from the FEA-determined structural topology. The algorithm is trained using data produced by the SIMP approach, which takes into account two classic tasks—a cantilever beam and a simply supported beam—at various iteration stages. With a comparison to CNN and DNN models, the model is thoroughly evaluated using the metrics of accuracy, efficiency, and generalizability. Through the study, the following conclusions can be drawn.
(1) The prediction accuracy of the GRNN model in the training and testing sets is supported by and drastically shortens the training and prediction cost compared to the CNN and DNN models.
(2) Compared with the DNN and CNN algorithms, the GRNN model shows the best generalization ability on compliance prediction for structural topology with rotations and under different material volume fractions.
(3) The moment invariants have translation, scaling, and rotational invariance, and the evaluation demonstrates that using the moment invariants as features is important in determining a reliable deep learning model for compliance prediction in topology optimization problems.
In this study, the GRNN is considered a preferred algorithm with high accuracy, efficiency, and generalization ability in the application of exploring the underlying relationship between the structural topology and its compliance. However, we still acknowledge several limitations. The current work denotes an end-to-end paradigm to predict the structural compliances in the SIMP method, which represents a simplified classic isotropic scenario without considering the manufacturing constraints for fiber-reinforced composites. The generalizable deep learning method represents a potential strategy in aspects of replacing the FEA calculation for high-resolution problems within the traditional topology optimization framework. Future efforts can be improved by involving constraints of the placement of path planning of fibers, which is essential in customizing the mechanical anisotropy in manufacturing, with the traditional numerical implementation being more costly.
The current work is promising in that it can be extended to the revised design issues and generate real-time optimization for advanced FRPC structure design [35,36]. It can also be utilized together with fiber printing techniques for applications in additive manufacturing [37].
Y.Z.: conceptualization, methodology, software, writing—original draft, and review. Z.C.: software, data curation, visualization, writing—review and editing. Y.D.: conceptualization, methodology, writing—review and editing. All authors have read and agreed to the published version of the manuscript.
Not applicable.
The data that support the findings of this study are available upon reasonable request from the corresponding authors.
The authors declare that they have no conflict of interest to report in this paper.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 2. Illustration of topology optimization for two tasks of (a) simply supported beam and its compliance iteration plot in (b); (c) a cantilever beam and its compliance iteration plot in (d).
Figure 3. The GRNN framework uses the moment invariants of topology configuration as input features to predict the compliances.
Figure 4. Calculation of moment invariants with six classic structural topology configurations, which include the simply supported beam in (a–c) and the cantilever beam in (d–f) at iterations of 10, 20, and 40.
Figure 5. Evaluation of the GRNN performance (a) on simply supported beams with the training set and (c) the testing set; (b) on cantilever beams on the training set and (d) testing set.
Figure 6. Plots of MSE and [Forumla omitted. See PDF.] with epochs of the (a) DNN model and (b) CNN model.
Figure 7. Model generalization on a simply supported beam with the topology configuration (a) 0 degree and (b) 180 degree clockwise rotation; and on a cantilever beam with the topology configuration (c) 0 degree and (d) 180 degree clockwise rotation.
Figure 8. Investigation of the GRNN predicted compliance with different volume fractions for (a) a simply supported beam and (b) a cantilever beam.
Computing proportion for each module of the SIMP method.
Module in SIMP | Computational Cost |
---|---|
FEA | 30.3% |
Sensitivities | 0.1% |
Filter | 67.4% |
Variables update | 2.2% |
Calculations of moment invariants of different topology configurations as in
Moment Invariants | Case a | Case b | Case c | Case d | Case e | Case f |
---|---|---|---|---|---|---|
m1 | 0.011625 | 0.011821 | 0.011668 | 0.010311 | 0.010683 | 0.010505 |
m2 | 0.000767 | 0.000752 | 0.000832 | 0.000416 | 0.000458 | 0.000481 |
m3 | 0.000499 | 0.000506 | 0.000535 | −0.000240 | −0.000278 | −0.000276 |
m4 | 0.000237 | 0.000228 | 0.000266 | 0.000097 | 0.000110 | 0.000119 |
m5 | 0.000229 | 0.000222 | 0.000256 | 0.000101 | 0.000115 | 0.000124 |
m6 | −0.000106 | −0.000126 | −0.000145 | −0.000057 | −0.000065 | −0.000066 |
m7 | 0.000166 | 0.000169 | 0.000169 | 0.000125 | 0.000130 | 0.000129 |
m8 | 0.000194 | 0.000190 | 0.000211 | 0.000099 | 0.000109 | 0.000114 |
m9 | 0.000120 | 0.000123 | 0.000130 | −0.000057 | −0.000065 | −0.000065 |
m10 | 0.000179 | 0.000173 | 0.000198 | 0.000086 | 0.000096 | 0.000102 |
m11 | 0.000152 | 0.000152 | 0.000166 | −0.000071 | −0.000081 | −0.000083 |
Parameter settings for the GRNN model, with the spread
P (Input Vectors) | T (Target Class Vectors) | Spread |
---|---|---|
|
|
0.023 |
Model performance evaluation of accuracy (
Model | Prediction Accuracy | Computational Efficiency | ||
---|---|---|---|---|
Training Set | Testing Set | Training | Prediction | |
GRNN | 0.998 | 0.994 | 1191.35 s | 0.0019 s |
DNN | 0.991 | 0.964 | 7128.01 s | 0.0061 s |
CNN | 0.999 | 0.999 | 38662.40 s | 2.71 s |
Model generalization on unknown samples of simply supported beams.
Images [0°] | FEA Calculation | Compliance Predictions | |||||
---|---|---|---|---|---|---|---|
GRNN | DNN | DNN | |||||
[0°] | [90°] | [180°] | [270°] | ||||
[Image omitted. Please see PDF.] | 12.086 | 12.128 | 11.842 | 11.516 | 26.555 | 22.240 | 55.067 |
0.352% | 2.017% | 4.716% | 119.719% | 84.013% | 355.633% | ||
[Image omitted. Please see PDF.] | 23.871 | 24.684 | 24.758 | 23.297 | 30.567 | 11.785 | 33.911 |
3.408% | 3.716% | 2.405% | 28.051% | 50.630% | 42.062% | ||
[Image omitted. Please see PDF.] | 16.116 | 16.127 | 15.923 | 15.627 | 20.688 | 23.765 | 50.847 |
0.064% | 1.020% | 3.035% | 28.364% | 47.464% | 215.516% | ||
[Image omitted. Please see PDF.] | 20.538 | 20.607 | 21.134 | 19.826 | 38.330 | 14.006 | 30.903 |
0.334% | 2.753% | 3.468% | 86.626% | 31.806% | 50.465% |
Model generalization on unknown samples of the cantilever beams.
Images [0°] | FEA Calculation | Compliance Predictions | |||||
---|---|---|---|---|---|---|---|
GRNN | DNN | DNN | |||||
[0°] | [90°] | [180°] | [270°] | ||||
[Image omitted. Please see PDF.] | 213.518 | 208.998 | 206.768 | 213.867 | 39.191 | 339.148 | 39.817 |
2.117% | 3.161% | 0.163% | 81.645% | 58.838% | 81.352% | ||
[Image omitted. Please see PDF.] | 532.576 | 546.824 | 515.810 | 537.995 | 30.539 | 90.565 | 35.502 |
2.675% | 3.148% | 1.017% | 94.266% | 82.995% | 93.334% | ||
[Image omitted. Please see PDF.] | 228.132 | 228.900 | 232.161 | 225.368 | 12.440 | 110.099 | 19.820 |
0.337% | 1.766% | 1.211% | 94.547% | 51.739% | 91.312% | ||
[Image omitted. Please see PDF.] | 168.211 | 168.694 | 167.335 | 167.391 | 10.939 | 90.536 | 13.731 |
0.287% | 0.521% | 0.488% | 93.497% | 46.177% | 91.837% |
Appendix A. Topology Configurations Used for Model Generalization Evaluation
Figure A1. Demonstration of the dataset in accordance with SIMP implementations: showing the constraints and topology configurations at iterations from 0 to 40.
The loading position, material volume fraction, and iteration of the topology configurations in accordance with
Images |
Load Position | Material Volume Fraction | Iteration |
---|---|---|---|
[Image omitted. Please see PDF.] | 329 | 0.4510 | 10 |
[Image omitted. Please see PDF.] | 2297 | 0.5611 | 20 |
[Image omitted. Please see PDF.] | 3608 | 0.6206 | 30 |
[Image omitted. Please see PDF.] | 1148 | 0.3738 | 40 |
The loading position, material volume fraction, and iteration of the topology configurations in accordance with
Images |
Load Position | Material Volume Fraction | Iteration |
---|---|---|---|
[Image omitted. Please see PDF.] | 4958 | 0.5143 | 10 |
[Image omitted. Please see PDF.] | 4924 | 0.3328 | 20 |
[Image omitted. Please see PDF.] | 4960 | 0.6766 | 30 |
[Image omitted. Please see PDF.] | 4948 | 0.5483 | 40 |
Appendix B. Deep Neural Networks of DNN and CNN
Figure A2. Illustration of model architecture for neural network models of (a) DNN and (b) CNN.
The DNN model absorbs 14 features as inputs and it is finalized with three hidden layers with the batch size setting as 71, the learning rate setting as 0.0001, and the epoch setting as 400. The CNN model treats it as a topology image recognition problem and takes the image with a resolution of
Parameter settings of the DNN neural network model.
Layer (Type) | Input Shape | Output Shape |
---|---|---|
Input | 14 | 248 |
Hidden1 | 248 | 253 |
Hidden2-1 | 253 | 253 |
Hidden2-2 | 253 | 253 |
Hidden2-3 | 253 | 253 |
Output | 253 | 1 |
Parameter settings of the CNN neural network model.
Layer (Type) | Input Shape | Output Shape |
---|---|---|
CONV1 | 121,632 | 3,600,896 |
Max pooling1 | 3,600,896 | 898,560 |
CONV2 | 898,560 | 1,725,888 |
Max pooling2 | 1,725,888 | 425,600 |
Flatten | 425,600 | 425,600 |
Hidden1 | 425,600 | 400 |
Hidden2 | 400 | 200 |
Hidden3 | 200 | 1 |
References
1. Rajak, D.K.; Pagar, D.D.; Menezes, P.L.; Linul, E. Fiber-reinforced polymer composites: Manufacturing, properties, and applications. Polymers; 2019; 11, 1667. [DOI: https://dx.doi.org/10.3390/polym11101667]
2. Wong, J.; Altassan, A.; Rosen, D.W. Additive manufacturing of fiber-reinforced polymer composites: A technical review and status of design methodologies. Compos. Part B Eng.; 2023; 146, 110603. [DOI: https://dx.doi.org/10.1016/j.compositesb.2023.110603]
3. Pham, T.; Kwon, P.; Foster, S. Additive manufacturing and topology optimization of magnetic materials for electrical machines—A review. Energies; 2021; 14, 283. [DOI: https://dx.doi.org/10.3390/en14020283]
4. Zhao, Z.L.; Rong, Y.; Yan, Y.; Feng, X.Q.; Xie, Y.M. A subdomain-based parallel strategy for structural topology optimization. Acta Mech. Sin.; 2023; 39, 422357. [DOI: https://dx.doi.org/10.1007/s10409-023-22357-x]
5. Zhang, W.; Yang, J.; Xu, Y.; Gao, T. Topology optimization of thermoelastic structures: Mean compliance minimization or elastic strain energy minimization. Struct. Multidiscip. Optim.; 2014; 49, pp. 417-429. [DOI: https://dx.doi.org/10.1007/s00158-013-0991-9]
6. Charoen-amornkitt, P.; Alizadeh, M.; Suzuki, T.; Tsushima, S. Entropy generation analysis during adjoint variable-based topology optimization of porous reaction-diffusion systems under various design dimensionalities. Int. J. Heat Mass Transf.; 2023; 202, 123725. [DOI: https://dx.doi.org/10.1016/j.ijheatmasstransfer.2022.123725]
7. Jafari-Asl, J.; Seghier, M.E.A.B.; Ohadi, S.; van Gelder, P. Efficient method using Whale Optimization Algorithm for reliability-based design optimization of labyrinth spillway. Appl. Soft Comput.; 2021; 101, 107036. [DOI: https://dx.doi.org/10.1016/j.asoc.2020.107036]
8. Keshtegar, B.; Meng, D.; Ben Seghier, M.E.A.; Xiao, M.; Trung, N.T.; Bui, D.T. A hybrid sufficient performance measure approach to improve robustness and efficiency of reliability-based design optimization. Eng. Comput.; 2021; 37, pp. 1695-1708. [DOI: https://dx.doi.org/10.1007/s00366-019-00907-w]
9. Bendsøe, M.P. Optimal shape design as a material distribution problem. Struct. Multidiscip. Optim.; 1989; 1, pp. 193-202. [DOI: https://dx.doi.org/10.1007/BF01650949]
10. Allaire, Grégoire and Jouve, François and Toader, Anca-Maria. A level-set method for shape optimization. C. R. Math.; 2002; 334, pp. 1125-1130. [DOI: https://dx.doi.org/10.1016/S1631-073X(02)02412-3]
11. Xie, Y.; Steven, G. A simple evolutionary procedure for structural optimization. Comput. Struct.; 1993; 49, pp. 885-896. [DOI: https://dx.doi.org/10.1016/0045-7949(93)90035-C]
12. Guo, X.; Zhang, W.; Zhong, W. Doing topology optimization explicitly and geometrically—A new moving morphable components based framework. J. Appl. Mech.; 2014; 81, 081009. [DOI: https://dx.doi.org/10.1115/1.4027609]
13. Liu, C.; Zhu, Y.; Sun, Z.; Li, D.; Du, Z.; Zhang, W.; Guo, X. An efficient moving morphable component (MMC)-based approach for multi-resolution topology optimization. Struct. Multidiscip. Optim.; 2018; 58, pp. 2455-2479. [DOI: https://dx.doi.org/10.1007/s00158-018-2114-0]
14. Rozvany, G. The SIMP method in topology optimization-theoretical background, advantages and new applications. Proceedings of the 8th Symposium on Multidisciplinary Analysis and Optimization; Long Beach, CA, USA, 6–8 September 2000; 4738.
15. Zegard, T.; Paulino, G.H. Toward GPU accelerated topology optimization on unstructured meshes. Struct. Multidiscip. Optim.; 2013; 48, pp. 473-485. [DOI: https://dx.doi.org/10.1007/s00158-013-0920-y]
16. Alber, M.; Buganza Tepole, A.; Cannon, W.R.; De, S.; Dura-Bernal, S.; Garikipati, K.; Karniadakis, G.; Lytton, W.W.; Perdikaris, P.; Petzold, L. et al. Integrating machine learning and multiscale modeling—Perspectives, challenges, and opportunities in the biological, biomedical, and behavioral sciences. NPJ Digit. Med.; 2019; 2, 115. [DOI: https://dx.doi.org/10.1038/s41746-019-0193-y] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31799423]
17. Zhao, Y.; Chen, Z.; Dong, Y.; Tu, J. An interpretable LSTM deep learning model predicts the time-dependent swelling behavior in CERCER composite fuels. Mater. Today Commun.; 2023; 37, 106998. [DOI: https://dx.doi.org/10.1016/j.mtcomm.2023.106998]
18. Chen, Z.; Liu, W. An efficient parameter adaptive support vector regression using K-means clustering and chaotic slime mould algorithm. IEEE Access; 2020; 8, pp. 156851-156862. [DOI: https://dx.doi.org/10.1109/ACCESS.2020.3018866]
19. Huang, M.; Du, Z.; Liu, C.; Zheng, Y.; Cui, T.; Mei, Y.; Li, X.; Zhang, X.; Guo, X. Problem-independent machine learning (PIML)-based topology optimization—A universal approach. Extrem. Mech. Lett.; 2022; 56, 101887. [DOI: https://dx.doi.org/10.1016/j.eml.2022.101887]
20. Kallioras, N.A.; Kazakis, G.; Lagaros, N.D. Accelerated topology optimization by means of deep learning. Struct. Multidiscip. Optim.; 2020; 62, pp. 1185-1212. [DOI: https://dx.doi.org/10.1007/s00158-020-02545-z]
21. Dong, Y.; Tao, J.; Zhang, Y.; Lin, W.; Ai, J. Deep learning in aircraft design, dynamics, and control: Review and prospects. IEEE Trans. Aerosp. Electron. Syst.; 2021; 57, pp. 2346-2368. [DOI: https://dx.doi.org/10.1109/TAES.2021.3056086]
22. Li, Z.; Zhang, Y.; Ai, J.; Zhao, Y.; Yu, Y.; Dong, Y. A Lightweight and Explainable Data-driven Scheme for Fault Detection of Aerospace Sensors. IEEE Trans. Aerosp. Electron. Syst.; 2023; early access [DOI: https://dx.doi.org/10.1109/TAES.2023.3303855]
23. Lee, S.; Kim, H.; Lieu, Q.X.; Lee, J. CNN-based image recognition for topology optimization. Knowl.-Based Syst.; 2020; 198, 105887. [DOI: https://dx.doi.org/10.1016/j.knosys.2020.105887]
24. Yan, J.; Zhang, Q.; Xu, Q.; Fan, Z.; Li, H.; Sun, W.; Wang, G. Deep learning driven real time topology optimisation based on initial stress learning. Adv. Eng. Inform.; 2022; 51, 101472. [DOI: https://dx.doi.org/10.1016/j.aei.2021.101472]
25. Li, B.; Huang, C.; Li, X.; Zheng, S.; Hong, J. Non-iterative structural topology optimization using deep learning. Comput.-Aided Des.; 2019; 115, pp. 172-180. [DOI: https://dx.doi.org/10.1016/j.cad.2019.05.038]
26. Rade, J.; Balu, A.; Herron, E.; Pathak, J.; Ranade, R.; Sarkar, S.; Krishnamurthy, A. Algorithmically-consistent deep learning frameworks for structural topology optimization. Eng. Appl. Artif. Intell.; 2021; 106, 104483. [DOI: https://dx.doi.org/10.1016/j.engappai.2021.104483]
27. Hu, M.K. Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory; 1962; 8, pp. 179-187.
28. Dong, Y.; Huang, J.; Ai, J. Visual perception-based target aircraft movement prediction for autonomous air combat. J. Aircr.; 2015; 52, pp. 538-552. [DOI: https://dx.doi.org/10.2514/1.C032764]
29. Specht, D.F. A general regression neural network. IEEE Trans. Neural Netw.; 1991; 2, pp. 568-576. [DOI: https://dx.doi.org/10.1109/72.97934]
30. Zheng, X.; Yang, R.; Wang, Q.; Yan, Y.; Zhang, Y.; Fu, J.; Liu, Z. Comparison of GRNN and RF algorithms for predicting heat transfer coefficient in heat exchange channels with bulges. Appl. Therm. Eng.; 2022; 217, 119263. [DOI: https://dx.doi.org/10.1016/j.applthermaleng.2022.119263]
31. Siqueira, A.S.; da Silva, R.C.; Santos, L.R. Perprof-py: A python package for performance profile of mathematical optimization software. J. Open Res. Softw.; 2016; 4, e12. [DOI: https://dx.doi.org/10.5334/jors.81]
32. Sosnovik, I.; Oseledets, I. Neural networks for topology optimization. Russ. J. Numer. Anal. Math. Model.; 2019; 34, pp. 215-223. [DOI: https://dx.doi.org/10.1515/rnam-2019-0018]
33. Flusser, J.; Zitova, B.; Suk, T. Moments and Moment Invariants in Pattern Recognition; John Wiley & Sons: Hoboken, NJ, USA, 2009.
34. Guest, J.K. Imposing maximum length scale in topology optimization. Struct. Multidiscip. Optim.; 2009; 37, pp. 463-473. [DOI: https://dx.doi.org/10.1007/s00158-008-0250-7]
35. Sattari, K.; Xie, Y.; Lin, J. Data-driven algorithms for inverse design of polymers. Soft Matter; 2021; 17, pp. 7607-7622. [DOI: https://dx.doi.org/10.1039/D1SM00725D] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34397078]
36. Hao, X.; Zhang, G.; Deng, T. Improved Optimization of a Coextrusion Die with a Complex Geometry Using the Coupling Inverse Design Method. Polymers; 2023; 15, 3310. [DOI: https://dx.doi.org/10.3390/polym15153310] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/37571203]
37. Liu, J.; Gaynor, A.T.; Chen, S.; Kang, Z.; Suresh, K.; Takezawa, A.; Li, L.; Kato, J.; Tang, J.; Wang, C.C. et al. Current and future trends in topology optimization for additive manufacturing. Struct. Multidiscip. Optim.; 2018; 57, pp. 2457-2483. [DOI: https://dx.doi.org/10.1007/s00158-018-1994-3]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Topology optimization techniques are essential for manufacturing industries, such as designing fiber-reinforced polymer composites (FRPCs) and structures with outstanding strength-to-weight ratios and light weights. In the SIMP approach, artificial intelligence algorithms are commonly utilized to enhance traditional FEM-based compliance minimization procedures. Based on an effective generalized regression neural network (GRNN), a new deep learning algorithm of compliance prediction for structural topology optimization is proposed. The algorithm learns the structural information using a fourth-order moment invariant analysis of the structural topology obtained from FEA at different iterations of classical topology optimization. A cantilever and a simply supported beam problem are used as ground-truth datasets, and the moment invariants are used as independent variables for input features. By comparing it with the well-known convolutional neural network (CNN) and deep neural network (DNN) models, the proposed GRNN model achieves a high prediction accuracy (
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 School of Aerospace Engineering and Applied Mechanics, Tongji University, Shanghai 200092, China;
2 Department of Aeronautics and Astronautics, Fudan University, Shanghai 200433, China