This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
1. Introduction
Quality detection is an important task in the industrial production process, which is of great significance to protect the personal safety of users and avoid economic losses. In industrial quality detection, big data analysis technology, the Internet of Things, edge computing, and other technologies are widely used. Studying an industrial detection algorithm that can be organically combined with the Internet of Things and edge computing is imminent. Early machine vision algorithms used manual feature selection and trained classifiers to identify defect features. This approach relies too much on the robustness of the extracted features, resulting in much time-consuming development.
In recent years, convolutional neural networks (CNNs) have rapidly developed in image classification, object detection, and image segmentation. However, due to the particularity of industrial defect detection scenarios, deep learning has not been applied widely. As shown in Figure 1, the brightness, colour, object size, and background discrimination of industrial pictures differ from natural scenes. Specifically, the industrial surface defect detection datasets scale is relatively small, unlike large-scale general datasets such as ImageNet, PASCAL VOC2007/2012, and COCO in classical computer vision tasks. Moreover, the industrial datasets differ significantly from the public datasets in the number of samples, sample scale, and proportion of positive and negative samples. This difference leads to less application of general-purpose object detectors in industrial product surface defect detection. In addition, most industrial detection scenarios are offline and require a light model, while existing general object detection models are challenging to meet actual needs.
[figure(s) omitted; refer to PDF]
This study proposes a novel surface defect detection model for industrial products. The model consists of two submodels, T-model and S-model. T-model has a large depth and high detection accuracy and is suitable for scenarios with no obvious requirements for speed. S-model has a small depth and high speed and is more suitable for edge computing scenarios with high-speed requirements. Both models consist of a backbone, neck, and detector. In the backbone, this study combines CNN and swin-transformer. CNN extracts local information in shallow features, swin-transformer extracts global information in deep layers, and each layer outputs weighted features through a unified residual path. In neck, this study designs a dual attention structure to focus on the features of object regions. In detector, this study designs multiple predictions heads on T-model and S-model.
In order to improve the detection effect of the lightweight S-model, a knowledge distiller is presented in this study. It contains the foreground attention-guided distillation of objects and global distillation. The knowledge distiller can significantly improve the detection accuracy of the S-model without extra overhead.
Overall, this study organically combines some existing techniques to create two detection models. An attention module is independently designed inside the model, and a knowledge distiller is designed to transfer the teacher network’s knowledge for the lightweight network’s guided training. The detection model designed in this study can not only be used for defect detection but also has a wide range of application scenarios in the direction of processing a large amount of image data in the cloud platform combined with the Internet of Things.
The main contributions of this study are listed as follows:
(1) This study proposes a novel model for surface defect detection of industrial products. Compared with the traditional CNN-based target detection methods, the model combines CNN and swin-transformer, which significantly improves the accuracy and can be used for application deployment in many image data analyses, edge computing, and other scenarios.
(2) This study designs two isomorphic object detection models. In backbone, this study extracts local and global important features by combining CNN and swin-transformer and designs a unified residual path to fuse features at different levels. In the neck, the attention module is designed to improve the effect of feature fusion. At the same time, this study designs the attention module to the decoupling detection head to improve detection accuracy.
(3) This study designs a novel attention-guided distillation strategy. The distiller uses the dual attention module to guide the generation of attention region features. The distiller transfers the knowledge of the T-model into the S-model, instructing the S-model to learn the T-model. The distiller in this study can obtain a lightweight, high-accuracy, and fast industrial quality detection model.
2. Related Work
2.1. Object Detection
In recent years, CNN-based object detection algorithms have been used in many domains. CNN-based object detection algorithms are generally divided into two categories: one stage and two stage. One-stage algorithms include YOLO [1–6], SSD [7], RetinaNet [8], and DSSD[9]. The methods are to directly divide the input image into multiple 1 × 1 grids, where each cell is responsible for detecting objects whose centre points fall within the grid. The methods significantly improve the detection speed, but the accuracy is slow. The two-stage algorithms include Rcnn [10], FastRcnn [11], FasterRcnn [12], and MaskRcnn [13]. These detection algorithms generate boxes via RPN, and the second-level detector uses boxes to conclude. The methods are better at detecting, but the speed is very low.
The one-stage object detector usually contains three parts: backbone, neck, and detector. The backbone is generally composed of multiple groups of convolutions for feature extraction. The well-known backbones include ResNet [14], ResNext [15], VGG [16], DenseNet [17], MobileNet [18], CSPDarkNet [19], and EfficentNet [20], etc. The neck fuses feature maps at different levels in the backbone to enhance the semantic and fine-grained features. Typical structures of neck mainly include FPN [21], PANet [22], Bi-FPN [20], etc. Two types of detectors commonly used in one-stage object detection are as follows: coupled and decoupled.
2.2. Attention
The attention mechanism is to make the network pay more attention to the area of the object and ignore the unimportant areas. Its essence is to use the relevant features to learn the weight distribution and then apply the learned weight to the original features. Attention makes the network pay more attention to the target object. Classical attention networks include SENet [23], SKNet [24], ResNext [15], CBAM [25], and self-attention [26].
2.3. Vision-Transformer
The transformer [26] is an attention-based encoder-decoder architecture in deep learning. Compared with CNN, vision transformer (Vit) [27] can obtain more refined global attention features and achieve good performance on multiple benchmarks such as ImageNet, COCO, and ADE20k. However, it also has some drawbacks. First, the range of object scales for visual detection varies greatly, and the performance of the Vit [27] may not be optimal in different scenarios. Second, if the image resolution is high, transformer [26] requires much computation. Swin-transformer [28] solves this problem by shifting window partitions to calculate self-attention.
2.4. Knowledge Distillation
Knowledge distillation is a widely used method for model compression. Knowledge distillation is to transfer the knowledge of the T-model into S-model to improve the accuracy of lightweight models without adding extra computation. Knowledge distillation methods can be roughly categorized into response based [29–32], feature based [33, 34], and relation based [35]. Response-based methods use the output of the last layer of the teacher network to imitate the teacher’s final prediction. This method is simple and efficient, but it relies on the output of the last layer and cannot make the student model obtain the supervision of the middle layer of the teacher model. Feature-based methods use the features in the middle layer of the teacher model to guide the student network to perform feature selection. However, two drawbacks need to be solved urgently. One is choosing the intermediate layer, and the other is matching the feature representation between the intermediate layer and the guiding layer if the layers’ sizes are different. Relation-based methods take advantage of the inner product of the features between the two layers and employ the teacher structure as knowledge to guide the student model.
2.5. Application in Big Data, Industrial Internet of Things, Defect Detection, and Other Scenarios
Wang et al. proposed a deep learning model [36] combining GRU and LSTM and modeled the crack width of the dam, which can effectively predict the change of dam defects. Chen et al. proposed a training method for CNN and proposed a two-layer parallel training (BPT-CNN) architecture [37] in a distributed computing environment. BPT-CNN effectively improves the training performance of CNN, saves training time while maintaining accuracy, and has wide application fields.
In addition, there are many excellent survey proposed in the fields of big data, industrial Internet of things, etc. Pu et al. proposed an automatic fetal ultrasound standard plane recognition (FUSPR) [38] based on deep learning in an Industrial Internet of Things (IIoT) Environment. Cao et al. proposed a novel BERT-based deep space network (BDSTN) [39] to learn the demand pattern of taxis. Chen et al. combined CNN and LSTM to propose multiple closed spatiotemporal CNNs (MGSTC) [40] for traffic flow prediction; Wang et al. [41] proposed the application of big data technology to data mining, data analysis, and data sharing in large amounts of data, and to create huge economic benefits by using the potential value of data. Zhang et al. [41] summarized the existing blockchain-based systems and applications, which have broad application prospects in different data processing and transmission scenarios. Zhang et al. proposed a combined method of Weber local descriptor (WLD) and local binary pattern (LBP) for seam carving forgery detection [42]. To sum up, some detection methods can be combined with existing technologies such as CNN, not only in defect detection scenarios, but also in different application scenarios such as IoT and big data analysis.
3. Method
The flowchart of the proposed model is shown in Figure 2. The model consists of two submodels: T-model and S-model. In order to improve the detection effect of the lightweight S-model, this study designs a knowledge distiller and a dual attention module. By generating attention features from the pretrained T-model, the S-model training process is guided to learn the features from the T-model. In this way, the S-model reaches or even exceeds the detection effect of the T-model.
[figure(s) omitted; refer to PDF]
The structure of the T-model and S-model is shown in Figure 3. Both models propose a novel backbone combining convolutional layers with swin-transformer [28] to extract finer-grained image features. This study also adds weighted residual connection paths in the backbone. In the neck, a double-tower structure and dual attention modules are designed to improve the effect of feature fusion. Finally, multiple decoupled detectors are used to detect objects of different scales.
[figure(s) omitted; refer to PDF]
The main formula symbol table used in the rest of the method of this study is annotated as Table 1.
Table 1
The main formula symbol used in the rest of the method.
Formula symbol annotation | Location |
Formula (1) | |
Formulas (2)–(4) | |
Formula (5) | |
Formula (6) | |
Formulas (9) and (10) | |
Formula (11) | |
Formula (13) | |
Formula (17) | |
Formula (18) | |
3.1. S-Model and T-Model
3.1.1. Backbone
Most general-purpose object detectors are based on CNN and employ large-scale convolution kernels in the first layer of the backbone to increase the receptive field. However, more detailed information will be lost with the increase of convolutional layers. Swin-transformer [28] uses window self-attention to significantly reduce computation load and extract small-scaled features. Therefore, this study designs a novel backbone based on swin-transformer [28].
As shown in Figure 3, our backbone includes block-B, block-S, SPP, and weighted residual fusion paths. The first layer selects a 3
[figure(s) omitted; refer to PDF]
In order to further improve the detection of small objects, this study designs a unified weighted residual path (Figure 3) and performs fusion of features from the backbone. The fusion formula is as follows:
3.1.2. Neck
The neck is designed to use better the features extracted by the backbone. In this study, the SPP structure is designed to enhance the invariance and robustness of image features. Figure 5 shows the structure of block-N in the top-down path and the bottom-up path in the neck. In block-N, this study designs a dual attention module, which will be introduced in detail in Section 3.2.
[figure(s) omitted; refer to PDF]
3.1.3. Decoupled Detector
In object detection, the role of the head in CNN is to regress the generated features to the bounding box and classified into some categories. Most methods use one head for classification and regression. This solution has limitations because classification relies more on fine-grained features, while regression relies more on semantic information to locate the object.
To stress the issue, YOLOX proposes the decoupled detector, which divides classification and regression into two types of problems. In this study, the decoupled detector is innovated (Figure 6), and a dual attention module is added to the classification branch. In our proposal, 6 decoupled detectors are used to detect features of the T-model, and four decoupled detectors to detect features of the S-model. The dense stacking of multiple decoupled detectors helps the models to detect objects of different scales.
[figure(s) omitted; refer to PDF]
3.2. Attention Module
Attention is usually embedded in CNNs and used to generate attention matrices to optimize features. CBAM [25] is a classic attention module that combines channel and spatial attention. Specifically, a 1 × 1 × C feature map will be obtained by pooling in channel attention, and a H × W × 1 feature map will be obtained by pooling in spatial attention. However, pooling leads to much information being lost. A mutual mapping between the three dimensions in a C × H × W image is crucial to extracting attention.
The attention structure proposed in this study (Figure 7) includes channel and spatial attention. The resulting attention features are multiplied by input features as their features are combined. In terms of channel attention, this study first uses 1 × 1 convolution instead of pooling to achieve channel interaction and information integration. It then inputs the features into a two-layer neural network (MLP) to amplify the relationship between channels and spatial dimensions. The number of neurons in the first layer in MLP is C/r (r is the reduction ratio, r = 16), the activation function is Relu, and the number of neurons in the second layer is C. In this study, the attention structure introduces BN to reduce the gradient dispersion and speed up the convergence speed. This study uses two dilated convolutions (d = 4, r = 16) identical to BAM [43] for spatial information integration to focus on spatial information.
[figure(s) omitted; refer to PDF]
The attention extraction formula in this study is as follows:
3.3. Knowledge Distillation Module
The lightweight model has speed and memory consumption advantages, but the detection effect is challenging to meet the requirements. Knowledge distillation is an effective method to improve the detection accuracy of small models. The general knowledge distiller focuses on the extraction of the overall features. The formula is as follows:
This study uses knowledge distillation to improve the S-model’s performance. The image’s foreground and background are distilled separately. The positive and negative samples are separated by separating the foreground and background, which solves the imbalance of positive and negative samples in the image. This study also designs an attention module to generate attention regions, forcing the student model to learn the vital features of the image.
In the distillation method of foreground and background, this study first sets the binary mask of foreground segmentation (the real position of the object frame) and sets objects within the ground truth box to 1 and objects outside the ground truth box to 0. It is expressed as follows:
Due to the large-scale variation of objects in industrial detection datasets and the uneven distribution of positive and negative samples of objects, these will adversely affect the distillation effect. For this reason, this study uses the scaling mask to balance the object scale with reference to FGD [44] to solve the problem of an unbalanced object scale. The formula is as follows:
After separating the foreground and background, this study uses the designed dual attention module to generate the attention mask. The attention mask formula is as follows:
In the distillation process of foreground and background, binary mask, scale mask, and attention mask are used for attention-guided distillation. The loss function is as follows:
In addition, this study uses the attention loss function to let the student model learn the attention mask of the teacher model. The loss formula is as follows:
In the feature distillation of the foreground, this study uses the designed attention module to distil the entire feature map. The loss formula is as follows:
3.4. Loss Function
In order to solve the unbalanced positive and negative samples in the PCB dataset, this study uses QFocal loss as the classification and confidence losses of the T-model and S-model. The formulas are as follows:
In summary, the total loss function proposed in this study is as follows:
4. Experiments and Analysis
In this section, experiments are carried out on the PCB defect dataset, the NEU surface defect dataset, and the aluminium defect dataset.
4.1. Experimental Environment and Parameter Settings
This article implements the code in the PyTorch framework, version 1.9.0. CUDA version 11.4, cuDNN version 8.0. This study’s model training and inference are performed on NVIDIA RTX 6000 × 1 and Intel i9-9900k@5 GHz × 1. Experimental platforms are GPU memory of 24 GB and CPU memory of 32 GB. The IDE used in the experiment is Pycharm 2019 Professional Edition.
In the training process, this study uses YOLOv5 as the baseline to build the T-model, uses the pretraining weights of YOLOv5 in the initial training of the T-model, and saves T-model weights after training. This study uses T-model for pretraining, loads the pretraining weights of the T-model to S-model for training, and uses knowledge distillation during the training process. Unless otherwise specified, the network in this study is trained with Adam for 300 iterations with an initial learning rate of 0.001, and the learning rate is adjusted using cosine annealing. This article uses a weight decay of 0.0001 and a momentum of 0.9. Also, the input image is resized to 640
4.2. Object Detection and Evaluation Indicators
FPS [45]: in this study, the model inference is carried out under the same equipment conditions. The same size image is used to calculate FPS and evaluate the model’s processing speed.
IOU [45]: object detection uses the IOU to calculate the degree of coincidence between the predicted box and the ground-truth box, which further measures the accuracy of detecting the corresponding object in a specific dataset.
mAP [45]: mAP is the sum of the average precision of all categories divided by the number of all categories. [email protected] is the model accuracy index when the IOU is 0.5. [email protected]:.95 is obtained by calculating an mAP every 0.5 from IOU from 0.5 to 0.95 and finally averaging these maps.
4.3. Experiments on the Aluminium Defect Dataset
The aluminium defect dataset is the images of aluminium surface defects published by Baidu AI (Figure 8). It contains 412 images in total. There are four types of defects, i.e., Zhen_kong, ca_shang, zang_wu, and zhe_zhou, and one image may contain different types of defects. In this study, the dataset is processed with photometric and geometric distortion. Specifically, this study brightens the picture and then flips and pans the picture to expand the dataset, increasing the number of pictures in the dataset to 1236. Then, the extended images are divided into a training set, validation set, and test set with a ratio of 8 : 1:1. It can be seen from Figure 8 that the surface defect scale of aluminium material varies greatly, which brings difficulties to the detection.
[figure(s) omitted; refer to PDF]
4.3.1. Comparative Experiments
Table 2 lists the comparison experiment results of aluminium defect datasets. It can be seen that the T-model of this study surpasses most classic object detection models, indicating that the model is more suitable for the field of industrial detection. T-model size is smaller than YOLOv5X, YOLOv4, YOLOR-P6, Faster-R-CNN, etc. The F1-score and mAP of the T-model are all in the leading position. T-model surpassed the newly proposed YOLOR 1.13% on [email protected], and F1-score exceeded 0.05. YOLOX has better detection performance than T-model, but T-model accuracy is close to YOLOX. However, due to the swin-transformer, the FPS advantage of the T-model is not apparent. In contrast, the S-model is lighter, the reasoning speed is fast, and the accuracy after knowledge distillation is close to the T-model.
Table 2
Comparative experiment of the aluminium defect dataset.
Approach | Backbone | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 | Model size (M) | FPS |
SSD | VGG-16 | 86.58 | 35.2 | 99.8 | 62.5 | 0.77 | 99.76 | 67 |
Faster-rcnn | ResNet-50 | 92.35 | 42.3 | 73.88 | 97.5 | 0.84 | 523 | 16 |
YOLOv3 | DarkNet-53 | 95.9 | 54.4 | 94.7 | 95.7 | 0.95 | 234 | 19 |
YOLOv5S | CSPDarkNet | 97.2 | 56.6 | 96.4 | 96.8 | 0.97 | 26 | 50 |
Efficientdet-d3 | EfficientNet-B3 | 93.52 | 50.2 | 91.85 | 90.91 | 0.91 | 14.78 | 20 |
YOLOv5X | CSPDarkNet | 94.3 | 54.9 | 95.1 | 94.7 | 0.95 | 167 | 21 |
Centrenet | Hourglass-104 | 60.3 | 38.6 | 51.09 | 60.61 | 0.55 | 124.61 | 48 |
Retinanet | ResNet-101 | 96.52 | 46.7 | 94.79 | 93.32 | 0.94 | 144.84 | 20 |
YOLOv4 | CSPDarkNet-53 | 94.13 | 59.1 | 83.38 | 96.01 | 0.89 | 245.53 | 29 |
YOLOR-P6 | CSPDarkNet | 97 | 53.1 | 88.3 | 97.8 | 0.93 | 140.88 | 61 |
YOLOX-s | Darknet-53 | 99.2 | 58.2 | 98.91 | 99.3 | 0.99 | 34.2 | 41 |
T-model | Our backbone | 98.53 | 58.8 | 98.85 | 98.38 | 0.98 | 81.6 | 19 |
S-model | Our backbone | 97.82 | 55.86 | 97.83 | 96.87 | 0.97 | 25.1 | 36 |
In order to more intuitively show the detection effect of the T-model, Figure 9 shows the detection results of 16 pictures.
[figure(s) omitted; refer to PDF]
4.3.2. Ablation Experiments
Table 3 lists the ablation experiments performed with YOLOv5S as the baseline. In the ablation experiment of the backbone, this study uses the B4, S3, and SPP outputs in the backbone as the input of YOLOv5S neck for experimenting. The results in the second row show that when using the backbone of the T-model, [email protected] is 0.63% higher than that of YOLOv5S, and the other indicators are also slightly improved. On the neck ablation experiment, this study experiments with three C3s in the YOLOv5S backbone with one SPP output as the neck input. The third row shows that the detection effect is improved when introducing the neck in this study.
Table 3
Ablation experiment with YOLOv5S.
Approach and improvement | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
YOLOv5S | 97.2 | 56.6 | 96.4 | 96.8 | 0.97 |
+Our backbone | 97.83 | 56.94 | 97.84 | 96.87 | 0.97 |
+Our neck | 97.71 | 57.13 | 97.1 | 97.38 | 0.98 |
This paper also conducts ablation experiments with T-model as the baseline to verify the residual fusion path in the backbone and the down sampling path in neck. The ablation results are listed in Table 4. The first row shows the results without weighted fusion paths in the backbone and down sampling paths in neck. The second row shows that when adding the residual fusion path to the backbone, [email protected] increases by 0.58%, and the F1-score increases by 0.02. The third rows show that when introducing the down sampling path in neck, [email protected] is improved by 0.52%, and the F1-score is improved by 0.02. In general, the residual fusion path and the down sampling path are beneficial to industrial detection scenarios with drastic changes in scale.
Table 4
Ablation experiment with the T-model.
Approach and improvement | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
T-model | 97.91 | 57.43 | 97.24 | 97.87 | 0.97 |
+Residual fusion path | 98.02 | 57.57 | 98.16 | 98.47 | 0.99 |
+Residual fusion path + down sampling path | 98.53 | 58.8 | 98.85 | 98.38 | 0.98 |
4.3.3. Experiments on the Detector
This study examines several different detectors, including coupled detector, decoupled detector, and decoupled detector with added dual attention module. This study conducts experiments with T-model as the baseline. Specifically, this study experiments with these detectors on T-model, and the results are listed in Table 5. The experimental results in the second row show that the [email protected] and F1-score of decoupled detector are improved by 0.78% and 0.05, respectively, compared with coupled detector. The third line shows that the detection accuracy is further improved when adding the dual attention module to the classification branch of decoupled detector, indicating that the dual attention module improves the classification accuracy.
Table 5
Experiments on three detectors.
Detector | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
Coupled detector | 97.24 | 57.05 | 86.4 | 97.31 | 0.93 |
Decoupled detector | 98.02 | 57.7 | 98.16 | 98.47 | 0.98 |
Decoupled detector + attention | 98.53 | 58.8 | 98.85 | 98.38 | 0.98 |
4.3.4. Experiments on the Dual Attention Module
This study analyses the features of CBAM. As shown in Figure 10, it is a heat map comparison between CBAM and the dual attention module. It can be seen that the dual attention module pays more attention to object areas such as wrinkles.
[figure(s) omitted; refer to PDF]
This study uses T-model as the baseline for experiments on the dual attention module (DA). The experimental results are listed in Table 6. It can be seen that the dual attention module effect surpasses CBAM.
Table 6
Comparison experiment between CBAM and dual attention module.
Approach | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
T-model without DA | 97.52 | 57.2 | 96.75 | 97.87 | 0.97 |
+DA | 98.53 | 58.8 | 98.85 | 98.38 | 0.98 |
+CBAM | 96.68 | 53.15 | 95.75 | 96.83 | 0.96 |
4.3.5. Knowledge Distillation
This study conducts experiments related to knowledge distillation on lightweight models. The experiments are performed on isomorphism object detectors and heterogeneous object detectors, respectively. On isomorphic object detectors, this study experiments T-model and S-model, Efficientdet-d7, and Efficientdet-d3. This study experiments with faster-R-CNN-ResNet50, faster-R-CNN-VGG16, YOLOv4, and YOLOv4-tiny on heterogeneous object detectors. In addition, this study also experimented with output layer knowledge distillation on T-model and S-model as a comparative experiment. Table 7 lists the experimental results of various classic object detectors. It can be seen that knowledge distillation improves the detection performance of the S-model, especially when T-model and S-model are isomorphic. It proves that the knowledge distiller can improve the detection accuracy of small models without increasing the number of parameters.
Table 7
Knowledge distiller experiments.
Models | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) |
S-model | 95.76 | 54.23 | 95.82 | 95.25 |
S-model with KD | 97.82 | 55.86 | 97.83 | 96.87 |
Faster-rcnn-VGG16 | 87.22 | 36.18 | 77.42 | 88.84 |
Faster-rcnn-VGG16 with KD | 87.87 | 36.26 | 87.46 | 86.67 |
Efficientdet-d3 | 93.52 | 50.2 | 91.85 | 90.91 |
Efficientdet-d3 with KD | 95.8 | 51.34 | 91.88 | 96.78 |
YOLOv4-tiny | 84.43 | 35.79 | 80.46 | 86.01 |
YOLOv4-tiny with KD | 84.94 | 35.6 | 80.44 | 86.9 |
4.4. Experiments on the PCB Defect Dataset
The PCB defect dataset has 1386 images, and its annotation files contain the object location information and classification. It contains six defects: missing hole, mouse bite, open circuit, spur, short, and spurious copper. Each image may contain multiple defective objects of the same type (missing holes in Figure 11). In the experiments in this study, 900 images of different defect types are selected for training (mainly minor defects that are difficult to identify). In our experiment, images and annotations are divided into the training set, validation set, and experiment set according to the ratio of 6 : 2 : 2.
[figure(s) omitted; refer to PDF]
This study conducts a comparative experiment on the PCB surface defect dataset, omitting FPS since the input size is still 640. It should be noted that the effect of the S-model is the result of distillation through the knowledge distiller. The comparison experiments are listed in Table 8. It can be seen that most of the classic object detection models are not effective for small object detection, and the T-model in this study shows a better performance.
Table 8
Comparative experiment of PCB surface defects.
Approach | Backbone | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
SSD | VGG-16 | 17.1 | 10 | 19.93 | 54.64 | 0.29 |
Faster-rcnn | ResNet-50 | 78.33 | 50.24 | 67.74 | 74.51 | 0.7 |
YOLOv3 | DarkNet-53 | 73.32 | 47.50 | 90.85 | 52.66 | 0.67 |
YOLOv5S | CSPDarkNet-53 | 91.2 | 51.8 | 81.82 | 92.12 | 0.86 |
Efficientdet-d3 | EfficicientNet-B3 | 71.8 | 35 | 99.44 | 49.51 | 0.66 |
YOLOv5X | CSPDarkNet-53 | 91.8 | 50.7 | 95.2 | 91.3 | 0.93 |
Centrenet | Hourglass-104 | 43.4 | 14.53 | 43.79 | 52.2 | 0.48 |
Retinanet | ResNet-101 | 13.18 | 5 | 66.67 | 4.31 | 0.08 |
YOLOv4 | CSPDarkNet-53 | 81.2 | 50.06 | 89.37 | 94.78 | 0.92 |
YOLOR-P6 | CSPDarkNet | 94.7 | 54.5 | 93.1 | 93.1 | 0.94 |
YOLOX-s | Darknet-53 | 95.84 | 62.03 | 87.16 | 93.73 | 0.91 |
T-model | Our backbone | 94.79 | 58.02 | 88.14 | 95.65 | 0.91 |
S-model | Our backbone | 92.51 | 56.98 | 83.79 | 92.33 | 0.88 |
In order to more intuitively show the detection effect of the T-model, Figure 12 shows the detection results of 16 pictures.
[figure(s) omitted; refer to PDF]
4.4.1. Ablation Experiment
Table 9 lists the ablation experiments performed in this study with YOLOv5S as the baseline. The experimental operations are consistent with the experiments on the aluminium defect dataset. The results in the second row show that after using the backbone of the T-model in this study, [email protected] is 3.68% higher than that of YOLOv5S indicating that the backbone in this study is beneficial for small object detection. The third line shows that the detection effect is improved when introducing the neck of the T-model. The dual attention module makes the network pay more attention to the object and improves the small object detection effect.
Table 9
Ablation experiment with YOLOv5S.
Approach and improvement | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
YOLOv5S | 91.2 | 51.8 | 81.82 | 92.12 | 0.86 |
+Our backbone | 91.98 | 52.48 | 96.05 | 92.46 | 0.94 |
+Our neck | 92.05 | 53.09 | 96.62 | 93.6 | 0.95 |
This study also uses T-model as the baseline to conduct ablation experiments to verify the residual fusion path in the backbone and the down sampling path in the neck. The experimental operations are consistent with the experiments on the aluminium defect dataset. The ablation test results are listed in Table 10. The first row shows the results without weighted fusion paths in the backbone and down sampling paths in neck. The second row shows that when adding the residual fusion path to the backbone, the effect is significantly improved, indicating that the residual path can better fuse the fine-grained features in the backbone. The third row shows that by introducing the down sampling path in neck, [email protected] is improved by 2.4%, and the F1-score is improved by 0.03. It shows that the down sampling path is beneficial for small object detection because more fine-grained feature information is fused into neck. Overall, the residual fusion and the down sampling paths can improve the detection effect.
Table 10
Ablation trials with T-model.
Approach and improvement | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
T-model | 92.39 | 54.21 | 91.71 | 89.84 | 0.88 |
+Residual fusion path | 93.65 | 56.44 | 89.94 | 94.43 | 0.92 |
+Residual fusion path + down sampling path | 94.79 | 58.02 | 88.14 | 95.65 | 0.91 |
4.5. Experiments on the NEU Surface Defect Dataset
North-eastern university releases the NEU surface defect dataset. This dataset collects six typical defects on the surface of the hot-rolled strip: rolled-in scale, patches, crazing, pitted surface, inclusion, and scratches. Each image has several defects of the same type. The label file marks the category and specific location of the defective object.
Figure 13 shows different kinds of steel defects. The categories’ defects have significant differences in appearance. Such scratches (last column) could be horizontal, vertical, slanted, etc. At the same time, each category has similarities in defects, such as rolled-in scale, crazing, and pitted surface. In addition, due to the influence of lighting and specific materials, the grey value of each category of defect images will also change. Object detection in the NEU surface defect dataset contains three difficulties: intraclass defects have significant appearance differences, interclass defects have similar aspects, and defect images are affected by changes in lighting and specific materials. According to the observation of the dataset, some cracks are concentrated in one direction because the steel is slender, and scratches are in any direction. Cracks and scratches are very similar, so scratches in specific directions are easily identified as cracks. Therefore, this study increases the data enhancement of cracks, rotates, splices crack pictures, and reduces the recognition error rate.
[figure(s) omitted; refer to PDF]
In order to show the generalization ability of this model in industrial detection, this study conducts a comparative experiment on the NEU surface defect database, where the S-model is the result of knowledge distillation. The comparative test results are listed in Table 11. It can be seen that the T-model and S-model still have better performance, indicating that the object detection model of this study has a particular generalization ability in multiple industrial detection datasets. Figure 14 shows the different steel defect recognition results; it can be seen that all the defects are accurately identified.
Table 11
Comparative experiments on NEU defect datasets.
Approach | Backbone | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
SSD | VGG-16 | 66.73 | 26.6 | 87.27 | 34.52 | 0.5 |
Faster-rcnn | ResNet-50 | 76.95 | 39.9 | 44.14 | 87.86 | 0.59 |
YOLOv3 | DarkNet-53 | 68.9 | 33.5 | 72.2 | 65.9 | 0.7 |
YOLOv5S | CSPDarkNet-53 | 70.9 | 35.6 | 78.4 | 64.3 | 0.35 |
Efficientdet-d3 | EfficicientNet-B3 | 65.65 | 34 | 84.76 | 48.44 | 0.77 |
YOLOv5X | CSPDarkNet-53 | 72 | 37.7 | 66.8 | 72.5 | 0.69 |
Centrenet | Hourglass-104 | 39.24 | 13.5 | 55.5 | 18.2 | 0.29 |
Retinanet | ResNet-101 | 66.54 | 31.9 | 81.73 | 44.62 | 0.57 |
YOLOv4 | CSPDarkNet-53 | 68.73 | 32.9 | 95.6 | 40.54 | 0.56 |
YOLOR-P6 | CSPDarkNet | 76.52 | 38.3 | 50.86 | 83.37 | 0.63 |
YOLOX-s | Darknet-53 | 79.38 | 40.23 | 53.73 | 84.88 | 0.66 |
T-model | Our backbone | 75.56 | 39.68 | 53.45 | 83.16 | 0.65 |
S-model | Our backbone | 74.41 | 39.2 | 49.05 | 82.28 | 0.62 |
[figure(s) omitted; refer to PDF]
4.6. Experiments on Glass Bottle Bottom Mould Point Dataset
This is a dataset for recognising mould point sequences at the bottom of glass bottles, with different permutations representing different product lot numbers. The mould point identification on the bottom of glass bottles is mainly used to locate the batch of glass products online to trace the product. The dataset contains 900 glass bottle bottom mould point images, each with a native resolution of 800 × 780 and 18 types. Each image has a corresponding label file. In this study, the label files correspond to the images one by one, and the training set, test set, and validation set are divided according to the ratio of 6 : 2 : 2.
Figure 15 shows the images of different glass bottle bottom mould points, and the high similarity between objects makes detection difficult. In order to demonstrate the generalization ability of the model in industrial detection, this study conducts a comparative experiment on the glass bottle bottom mould point database, where the S-model results from knowledge distillation. The comparison test results are listed in Table 12. It can be seen that T-model and S-model still have better performance, indicating that the object detection model in this study has a specific generalization ability on multiple industrial detection datasets. Figure 16 shows the recognition effect of 16 different model point images. It can be seen that all model point objects are accurately recognized and positioned.
[figure(s) omitted; refer to PDF]
Table 12
Comparative experiments on NEU defect datasets.
Approach | Backbone | [email protected] (%) | [email protected] (%) | Precision (%) | Recall (%) | F1 |
SSD | VGG-16 | 70.44 | 32.52 | 78.94 | 49.48 | 0.5 |
Faster-rcnn | ResNet-50 | 83.62 | 38.73 | 74.89 | 75.4 | 0.59 |
YOLOv3 | DarkNet-53 | 91.37 | 40.66 | 84.39 | 92.33 | 0.7 |
YOLOv5S | CSPDarkNet-53 | 93.21 | 62.1 | 85.19 | 91.44 | 0.35 |
Efficientdet-D3 | EfficicientNet-B3 | 86.39 | 55.67 | 67.49 | 89.9 | 0.77 |
YOLOv5X | CSPDarkNet-53 | 98.4 | 64.8 | 91.47 | 99.04 | 0.69 |
Centrenet | Hourglass-104 | 68.51 | 43.18 | 84.14 | 48.74 | 0.29 |
Retinanet | ResNet-101 | 80.79 | 49.89 | 66.28 | 77.56 | 0.57 |
YOLOv4 | CSPDarkNet-53 | 96.1 | 50.06 | 89.37 | 94.78 | 0.56 |
YOLOR-P6 | CSPDarkNet | 99.3 | 59.9 | 53.5 | 99 | 0.63 |
YOLOX-s | Darknet-53 | 99.55 | 70.68 | 80.32 | 99.9 | 0.66 |
T-model | Our backbone | 99.36 | 69.53 | 81.1 | 99.5 | 0.89 |
S-model | Our backbone | 98.1 | 65.2 | 79.44 | 98.91 | 0.89 |
[figure(s) omitted; refer to PDF]
5. Conclusions
This study proposes a deep learning model for industrial quality detection. The model consists of T-model and S-model, which aims to meet detection tasks under different conditions. The model uses a combination of swin-transformer and convolution to extract the global information of the image. A dual attention module is designed to improve the neck’s attention to important areas of the image, thereby improving the detection effect of the model. This study also designs a knowledge distiller using a dual attention module to improve the detection effect of the S-model. Finally, this study designs an adaptive loss weight updater to adjust the loss weights automatically. The experimental results show that the T-model in this study has high accuracy and is suitable for online data processing in scenarios such as IoT intelligent computing and big data analysis. The S-model in this study is fast and suitable for use in scenarios such as edge computing. In general, the model in this study can meet the needs of different scenarios and achieve a balance between accuracy and speed.
Disclosure
Zekai Zhang, Mingle Zhou, and Rui Shao are the co-first authors.
Acknowledgments
This work was completed with the technical support of Key R & D Plan of Shandong Province (Soft Science Project) (2021RZA01016) and Plan of Youth Innovation Team Development of Colleges and Universities in Shandong Province (SD2019-161), Jinan.
[1] J. Redmon, S. K. Divvala, R. B. Girshick, A. Farhadi, "You only look once: unified, real-time object detection," Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 779-788, .
[2] J. Redmon, A. Farhadi, "YOLO9000: better, faster, stronger," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6517-6525, .
[3] J. Redmon, A. Farhadi, "YOLOv3: An Incremental Improvement," 2018. http://arxiv.org/abs/1804.02767
[4] A. Bochkovskiy, C. Y. Wang, H. Y. M. Liao, "YOLOv4: Optimal Speed and Accuracy of Object Detection," 2020. http://arxiv.org/abs/2004.10934
[5] Z. Ge, S. Liu, F. Wang, Z. Li, J. Sun, "YOLOX: Exceeding YOLO Series in 2021," 2021. http://arxiv.org/abs/2107.08430
[6] Q. Chen, Y. Wang, T. Yang, X. Zhang, J. Cheng, J. Sun, "You only look one-level feature," Proceedings of the 2021 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 13039-13048, .
[7] W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C. Yang Fu, A. C. Berg, "SSD: single shot MultiBox detector," Proceedings of the 2016 Computer Vision 14th European Conference (ECCV), pp. 21-37, .
[8] T. Y. Lin, P. Goyal, R. B. Girshick, K. He, P. Dollár, "Focal loss for dense object detection," Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), pp. 2999-3007, .
[9] C. Y. Fu, W. Liu, A. Ranga, A. Tyagi, A. C. Berg, "DSSD: Deconvolutional Single Shot Detector," 2017. http://arxiv.org/abs/1701.06659
[10] R. B. Girshick, J. Donahue, T. Darrell, J. Malik, "Rich feature hierarchies for accurate object detection and semantic segmentation," Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 580-587, .
[11] R. B. Girshick, "Fast R-CNN," Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1440-1448, .
[12] S. Ren, K. He, R. B. Girshick, J. Sun, R.-C. N. N. Faster, "Faster R-CNN: towards real-time object detection with region proposal networks," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 39 no. 6, pp. 1137-1149, DOI: 10.1109/tpami.2016.2577031, 2017.
[13] K. He, G. Gkioxari, P. Dollár, R. B. Girshick, R.-C. N. N. Mask, "Mask R-CNN," Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2980-2988, .
[14] K. He, X. Zhang, S. Ren, J. Sun, "Deep residual learning for image recognition," Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770-778, .
[15] S. Xie, R. B. Girshick, P. Dollár, Z. Tu, K. He, "Aggregated residual transformations for deep neural networks," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5987-5995, .
[16] K. Simonyan, A. Zisserman, "Very deep convolutional networks for large-scale image recognition," Proceedings of 2015 the 3rd International Conference on Learning Representations, .
[17] G. Huang, Z. Liu, L. van der Maaten, K. Q. Weinberger, "Densely connected convolutional networks," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2261-2269, .
[18] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, H. Adam, "MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications," 2017. http://arxiv.org/abs/1704.04861
[19] C. Y. Wang, H. Y. M. Liao, Y. H. Wu, P. Y. Chen, J. W. Hsieh, I. H. Yeh, "CSPNet: a new backbone that can enhance learning capability of CNN," Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1571-1580, .
[20] M. Tan, Q. V. Le, "EfficientNet: rethinking model scaling for convolutional neural networks," Proceedings of the Proceedings of 2019 the 36th International Conference on Machine Learning (ICML), . 6114
[21] T. Y. Lin, P. Dollár, R. B. Girshick, K. He, B. Hariharan, S. J. Belongie, "Feature pyramid networks for object detection," Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 936-944, .
[22] S. Liu, L. Qi, H. Qin, J. Shi, J. Jia, "Path aggregation network for instance segmentation," Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 8759-8768, .
[23] J. Hu, L. Shen, S. Albanie, G. Sun, E. Wu, "Squeeze-and-Excitation networks," IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42 no. 8, pp. 2011-2023, DOI: 10.1109/tpami.2019.2913372, 2020.
[24] X. Li, W. Wang, X. Hu, J. Yang, "Selective kernel networks," Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 510-519, .
[25] S. Woo, J. Park, J. Y. Lee, I. S. Kweon, "CBAM: convolutional Block Attention module," Proceedings of the 15th European Conference Computer Vision (ECCV), .
[26] A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N-Gomez, Ł. Kaiser, I. Polosukhin, "Attention is all you need," Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), pp. 5998-6008, .
[27] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, N. Houlsby, "An image is worth 16x16 words: transformers for image recognition at scale," Proceedings of the 9th International Conference on Learning Representations (ICLR 2021), .
[28] Z. Liu, Y. Lin, Y. Cao, H. Hu, Y. Wei, Z. Zhang, S. Lin, B. Guo, "Swin transformer: hierarchical vision transformer using shifted windows," Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 9992-10002, .
[29] G. E. Hinton, O. Vinyals, J. Dean, "Distilling the Knowledge in a Neural Network," 2015. http://arxiv.org/abs/1503.02531
[30] J. H. Cho, B. Hariharan, "On the efficacy of knowledge distillation," Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 4793-4801, .
[31] C. Yang, L. Xie, S. Qiao, A. L. Yuille, "Training deep neural networks in generations: a more tolerant teacher educates better students," Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence (AAAI), pp. 5628-5635, .
[32] M. Phuong, C. Lampert, "Distillation-based training for multi-exit architectures," Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1355-1364, .
[33] J. He, C. Li, J. Ye, Y. Qiao, L. Gu, "Self-speculation of clinical features based on knowledge distillation for accurate ocular disease classification," Biomedical Signal Processing and Control, vol. 67,DOI: 10.1016/j.bspc.2021.102491, .
[34] M. Sepahvand, F. Abdali-Mohammadi, "Overcoming limitation of dissociation between MD and MI classifications of breast cancer histopathological images through a novel decomposed feature-based knowledge distillation method," Computers in Biology and Medicine, vol. 145,DOI: 10.1016/j.compbiomed.2022.105413, .
[35] F. Tung, G. Mori, "Similarity-preserving knowledge distillation," Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 1365-1374, .
[36] J. Wang, Y. Zou, P. Lei, S. Sherratt, L. Wang, "Research on recurrent neural network based crack opening prediction of concrete dam," Journal of Internet Technology, vol. 21, pp. 1151-1160, 2020.
[37] J. Chen, K. Li, K. Bilal, X. Zhou, K. Li, P. S. Yu, "A Bi-layered parallel training architecture for large-scale convolutional neural networks," IEEE Transactions on Parallel and Distributed Systems, vol. 30 no. 5, pp. 965-976, DOI: 10.1109/tpds.2018.2877359, 2019.
[38] B. Pu, K. Li, S. Li, N. Zhu, "Automatic fetal ultrasound standard plane recognition based on deep learning and IIoT," IEEE Transactions on Industrial Informatics, vol. 17 no. 11, pp. 7771-7780, DOI: 10.1109/tii.2021.3069470, 2021.
[39] D. Cao, K. Zeng, J. Wang, P. K. Sharma, X. Ma, Y. Liu, S. Zhou, "BERT-based deep spatial-temporal network for taxi demand prediction," IEEE Transactions on Intelligent Transportation Systems, vol. 23 no. 7, pp. 9442-9454, DOI: 10.1109/tits.2021.3122114, 2022.
[40] C. Chen, K. Li, S. G. Teo, X. Zou, K. Li, Z. Zeng, "Citywide traffic flow prediction based on multiple gated spatio-temporal convolutional neural networks," ACM Transactions on Knowledge Discovery from Data, vol. 14 no. 4,DOI: 10.1145/3385414, 2020.
[41] J. Zhang, S. Zhong, T. Wang, H. C. Chao, J. Wang, "Blockchain-based systems and applications: a survey," Journal of Internet Technology, vol. 21 no. 1, 2020.
[42] D. Zhang, Q. Li, G. Yang, L. Li, X. Sun, "Detection of image seam carving by using weber local descriptor and local binary patterns," Journal of Information Security and Applications, vol. 36, pp. 135-144, DOI: 10.1016/j.jisa.2017.09.003, 2017.
[43] J. Park, S. Woo, J. Y. Lee, I. S. Kweon, "BAM: bottleneck attention module," Proceedings of the British Machine Vision Conference 2018 (BMVC), .
[44] Z. Yang, Z. Li, X. Jiang, Y. Gong, Z. Yuan, D. Zhao, C. Yuan, "Focal and Global Knowledge Distillation for Detectors," 2021. https://arxiv.org/abs/2111.11837
[45] R. Padilla, W. L. Passos, T. L. B. Dias, S. L. Netto, E. A. B. da Silva, "A comparative analysis of object detection metrics with a companion open-source toolkit," Electronics Now, vol. 10 no. 3,DOI: 10.3390/electronics10030279, 2021.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Copyright © 2022 Ze-Kai Zhang et al. This is an open access article distributed under the Creative Commons Attribution License (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. https://creativecommons.org/licenses/by/4.0/
Abstract
Industrial quality detection is one of the important fields in machine vision. Big data analysis, the Internet of Things, edge computing, and other technologies are widely used in industrial quality detection. Studying an industrial detection algorithm that can be organically combined with the Internet of Things and edge computing is imminent. Deep learning methods in industrial quality detection have been widely proposed recently. However, due to the particularity of industrial scenarios, the existing deep learning-based general object detection methods have shortcomings in industrial applications. This study designs two isomorphic industrial detection models to solve these problems: T-model and S-model. Both proposed models combine swin-transformer with convolution in the backbone and design a residual fusion path. In the neck, this study designs a dual attention module to improve feature fusion. Second, this study presents a knowledge distiller based on the dual attention module to improve the detection accuracy of the lightweight S-model. According to the analysis of the experimental results on four public industrial defect detection datasets, the model in this study is more advantageous in industrial defect detection.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer