1. Introduction
Walnut shell–kernel separation is a critical procedure in the deep processing of walnuts and is a vital link to advance the industrial chain and added value of walnuts [1,2]. Following the mechanical breaking of the walnut, it is impossible to completely separate the shell and kernel during cleaning due to the complicated structure of mixtures and diversified forms [3]. Manual intervention is needed in the walnut selection. It incurs high costs and low efficiency [4], thereby restricting the development of the walnut industry. Shell–kernel recognition is one of the crucial technologies to realise intelligent shell–kernel selection. Analysing walnut shell–kernel detection algorithms based on deep learning is critical for achieving intelligent and mechanical walnut shell–kernel separation.
Walnut shell–kernel separation is a process that separates shell–kernel mixtures acquired after shell breaking by the shell-breaking device and collects walnut kernels. Many enterprises still adopt the manual selection technique, which has high labour intensity and separation costs [5]. Several Chinese and foreign scholars have conducted abundant investigations on the shell–kernel separation method of nuts. For instance, Krishnan et al. [6] proposed the magnetic shell–kernel separation technique according to their differences in magnetic conductance. However, fresh impurities were introduced into the separation process, necessitating further separation. Moreover, it was easy to contaminate kernels. Wang et al. [7] invented an electrostatic shell–kernel separation machine for nuts and successfully realised the shell–kernel separation of nuts based on their different dielectric properties. Moreover, there was no mechanical force on the shell and kernel during separation, without the restriction of particle size, shape and proportion. They provided a novel solution to the shell–kernel separation of nuts. Nevertheless, these methods presented various challenges, such as complicated separation processes, high costs, low efficiency, and poor guarantee of food quality and safety, among others. Hence, their promotion was challenging. The airflow selection method has a simple device structure and good separation effect and solves existing problems to a large extent. It is a commonly used method in walnut shell–kernel separation. Liu et al. [8], Cao et al. [9], and Li et al. [10] investigated walnut shell–kernel separation based on airflow separation and achieved a relatively good effect. Due to the impact of diversified sizes, shapes and windward areas, the walnut shells and kernels have complicated and changing motions in the airflow field, thus making it arduous to achieve thorough separation. In particular, some overlapping aerodynamic characteristics exist between walnut shells and kernels, thereby increasing the separation difficulty and restricting the improvement of the shell–kernel separation effect. Recently, the development of computer technology and machine vision technology provided novel solutions to walnut shell–kernel separation. Wang et al. [11] designed a pecan shell–kernel separation system based on a fuzzy clustering algorithm and separated shells and kernels after breaking using the fuzzy clustering algorithm. Their results demonstrated that the walnut kernel recognition rate of the system was higher than 83%. Jiang et al. [4] classified shells and kernels of black walnuts based on hyperspectral fluorescence imaging technology; the recognition rate could reach 90.3%. Jin et al. [12] designed an intelligent black walnut shell–kernel recognition system that acquired relatively high separation precision and calculation speed. It had some application potential in the field of walnut shell–kernel separation. However, it is hard to create high-efficiency applications of conventional image recognition technology in complicated actual production due to its detailed processing, low recognition precision, and requests for the manual extraction of target features [13,14]. With quick recognition, small model volume and stronger timeliness, the target detection algorithm based on YOLO can achieve end-to-end detection. Moreover, it has good feature extraction and generalisation abilities [15,16]. This algorithm is becoming an essential tool for the online detection and target recognition of agricultural products. For instance, Yao et al. [17] and Xiao et al. [18] used YOLOv5 and YOLOv8 target recognition algorithms, respectively, to conduct a real-time detection of fruit maturity. The results demonstrated that these models exhibited fast and accurate recognition. Wang et al. [19] used the YOLOv5 algorithm for the real-time recognition of apple stem and calyx, laying the foundation for the automation of fruit loading and packaging systems. Wu et al. [20] applied the deep convolutional neural network algorithm to detect walnut shell kernels, achieving good recognition results when the shell kernels were dispersed and classified. Meng et al. [21] detected tea buds amidst complex backgrounds using the YOLOv7 algorithm, providing a theoretical basis for intelligent tea picking. In addition, Zhang et al. [22] used YOLOX to achieve high-precision detection and counting of winter jujube, testing the method’s accuracy and effectiveness under different scales and scenarios. The results indicated that the algorithm had strong robustness under scenarios with shadows, coverage, and incomplete contours. These studies collectively demonstrate the success of YOLO-based target detection in recognition tasks, providing a reference for the development of walnut shell kernel separation technology.
This study proposed a fast walnut shell–kernel detection algorithm based on YOLOX to address the failure of complete shell–kernel separation after airflow selection. To verify the detection effect of the proposed algorithm in walnut shells-kernels, the images of shell–kernel mixtures that have not been separated and have disordered distribution after breaking walnuts were labelled. It achieved fast and accurate walnut shell–kernel detection based on the YOLOX target detection network, aiming for a fast, high-efficiency, lightweight walnut shell–kernel detection model. Moreover, the same dataset was selected for comparison with YOLOv3, Faster Region-based Convolutional Neural Network (Faster R-CNN), and the Single Shot MultiBox Detector algorithms (SSD) network models. This study provides references to investigate the walnut shell–kernel separation technology and research and development of online separation devices.
2. Materials and Methods
2.1. Materials
2.1.1. Data Acquisition and Processing of Walnut Shells and Kernels
Different species of walnuts have significant differences in appearance. For example, Wen 185 Walnut has thin pericarps and smooth shells. Yunnan Juglans sigillata has an obovate kernel with flat surfaces at two ends and wrinkles on the surface (Figure 1). These factors may influence the walnut shell–kernel recognition effect. Two major species (Yunnan Juglans sigillata and Wen 185 Walnut) in the primary producing areas of walnut in China were selected to obtain target detection models applicable to different walnut species. After breaking, the samples of walnut-shell–kernel mixtures were prepared. Using the Nikon D3500 camera, a total of 2753 photos were captured at the same height and background. The pictures were taken at a resolution of 3904 pixels × 2928 pixels in March 2023. The quality of the photos was primarily evaluated based on traditional personal perception [23,24], which generally met the practical criteria for shell and kernel separation conditions.
During walnut shell–kernel detection, numerous factors, such as walnut species, illumination, and shell–kernel distribution, may generate differences in walnut shell–kernel pictures and thereby influence their recognition effect (Figure 2). Walnut shell–kernel images under different scenes were collected. To obtain the optimal model for walnut shell–kernel target detection, the images of shells and kernels of Yunnan Juglans sigillata and Wen 185 Walnut under natural light and supplementary light, with and without mutual shielding, were collected in this study.
2.1.2. Dataset Preparation
A total of 2753 images were collected in this study. The walnut shell and kernel datasets were manually labelled using Labelme-master (Labelme-master 5.1.1, USA), during which the label frames selected the minimum enclosing rectangles of walnut shells and kernels. Label documents in the JSON format were produced. Because this study focused on walnut shell–kernel recognition, the images were only divided into shells, kernels and background while making labels. Only the shell and kernel must be labelled, whereas other parts of the image were labelled as background automatically using Labelme. Later, the JSON data were transformed into data in the COCO format.
2.1.3. Test Conditions
The training and test of research models of Yunnan Juglans sigillata and Wen 185 were conducted under the same conditions. The hardware setting of the device was 12th Gen Intel(R) Core(TM) i7-12700H 2.70 GHz, 11 GB NVIDIA GeForce RTX 2050Ti as CPU, 16 GB memory, and 64-bit Windows 11 system. The PyTorch deep-learning framework was applied, and the PyCharm programming platform was used.
2.2. Methods
2.2.1. Walnut Shell–Kernel Recognition Algorithm Based on YOLOX
YOLO (You Only Look Once) is an object recognition and positioning algorithm based on deep neural networks. It is an end-to-end real-time object detection system that treats the object detection task as a regression problem. YOLO series algorithms are favoured by engineering researchers for their fast response, high accuracy, simple structure, and easy deployment. The YOLOX model has improved and optimised based on the YOLO v5 network model from three aspects: decoupled head layer, data enhancement and anchor-free. Additionally, the model has been widely used due to its high accuracy and efficiency [25,26]. Unlike Faster R-CNN, YOLO can predict several candidate frames simultaneously. Based on the idea of regression, it directly detects target positions and classifies target objects via the first-order network [27,28]. Compared with a series of target detection algorithms like R-FCN and Faster R-CNN, YOLO has an outstanding characteristic of high operational speed and can meet end-to-end training and real-time detection [29]. In this test, walnut shell–kernel detection was realised based on the YOLOX deep convolutional neural network model, and the optimal model was achieved via 110 epochs of training. The validity and adaptation of the model were verified by comparing evaluation metrics.
2.2.2. YOLOX Network Structure
YOLOX uses the overall layout of the YOLO series, and its network structure mainly comprises Input, Backbone network, Neck network and Prediction [25,26]. Figure 3 shows the structure of YOLOX. Similar to YOLOv4 and YOLOv5, the Input applies the Mosaic data enhancement mode. Images were spliced after stacking, scaling, tailoring, and arranging different images [30]. The optimal anchoring frame value of the dataset was calculated automatically. The Backbone primarily consists of Focus and CSP. Based on YOLOv4-tiny, it realised a cross-stage local fusion network. In YOLOX, the input images are first cut into pieces in Focus, and the original input is divided into two branches using the GSPDarknet structure for convolutional operation and N residual block operation. Later, the branches were joined [31]. This action effectively relieves the vanishing gradient problem and decreases the number of network parameters. Moreover, it not only shortens the calculation time but also increases precision. The spatial pyramid pooling structure was applied at the position of the Backbone output, and the maximum pooling feature extraction was conducted by pooling nuclei with different sizes. This enhances the reception field of convolutional kernels effectively and is conducive to extracting richer local feature information [32]. The Neck comprises feature pyramid networks (FPN) + path aggregation network (PAN). From top to bottom, FPN integrates the deep layer features with superficial layer features by upsampling. It is mainly used to transfer semantic features. From bottom to top, PAN transfers superficial layer features to deep layers and integrates them. It is primarily used to transfer positioning information [33]. The prediction of YOLOX changed Yolohead into Decoupled Head [25] and realised regression and classification in two parts. Regression and classification are integrated during prediction. Moreover, the convergence rate and precision of the algorithm are improved.
2.2.3. Evaluation Metrics
To verify practicability and detection effect of the YOLOX model, the same dataset was selected for comparison with YOLOv3, Faster R-CNN and SSD network model. Average precisions (AP50, AP75, and APs) at the intersection over union (IOU) = 0.50, IOU = 0.75 and IOU = 0.50:0.95 and average recall (AR) at IOU = 0.50:0.95 were selected as the evaluation metrics of performances of the walnut shell–kernel detection training model.
(1)
(2)
(3)
where P refers to precision and R is recall. TP means the number of accurately predicted positive samples. FP is the quantity of wrongly predicted positive samples. FN is the quantity of wrongly predicted negative real samples.3. Results and Analysis
3.1. Model Training
In this study, 110 training epochs were performed on the dataset. Figure 4 shows the network training results. From the beginning to the 25th epoch of training, the learning efficiency of the model was high, and the loss curve demonstrated a high convergence rate. At about the 750th iteration, the learning efficiency of the model reached saturation gradually, and the loss fluctuated at about 1.4. The AP (IOU = 0.5) was 96.3% for the final training model, and the AR was 84.7%. The AP (IOU = 0.75) was 92.4%, and the AP (IOU = 0.50:0.95) was 80.6%.
3.2. Detection Results
In total, 550 walnut shell–kernel images in the test set were selected to verify the validity of the YOLOX network under different walnut shell–kernel detection scenes. According to test results, the AP50 and AR of YOLOX were 97.2% and 84.7% with respect to different walnut shell–kernel separations, respectively. The floating point operations per second (FLOPs) were 351.9, and the model size was 99 MB. The model presented high detection precision and speed and was robust under mutual shell–kernel shielding and illumination. Figure 5 shows the detection results. The confidence of detection results is presented above the detection frame. In Figure 5, all species of walnut shells and kernels could be recognised accurately, and the confidence is higher than 0.90. This outcome demonstrates that the proposed algorithm can effectively distinguish walnut shells and kernels in the mixture under different scenes with relatively high detection confidence.
3.3. Performance Comparison of Several Target Detection Algorithms
To evaluate the walnut shell–kernel detection effect of the YOLOX network, the walnut shell and kernel training sets were trained based on YOLOv3, SSD [27], and Faster R-CNN target detection algorithms under the same conditions. The number of training epochs was set at 110. Moreover, the performance of the above four detection algorithms was evaluated using the test set. The visual confusion matrix of performance comparison and evaluation metrics of the detection model at IOU = 0.5 are shown in Figure 6 and Table 1, respectively.
Table 2 shows that the YOLOX model side is basically equivalent to the Faster R-CNN target detection algorithm and is larger than YOLOv3 and SSD target detection algorithms. Compared with YOLOv3, Faster R-CNN, and SSD target detection algorithms, the AP of YOLOX at IOU = 0.50 is increased by 2.1%, 1.3%, and 3.4%, and the AR is increased by 10%, 2.3%, and 9%, respectively. The AP at IOU = 0.50:0.95 is increased by 12.1%, 3.9%, and 9.8%, respectively.
The 550 images in the test set were classified according to walnut species, illumination, and shielding conditions. Given the same test condition, the walnut shell–kernel image test sets under different scenes were tested using the above four target detection algorithms. Table 2 presents the test results of different algorithms.
Table 2 shows that the YOLOX detection model achieves relatively high AP and AR values under different walnut species, illumination, and shielding conditions. The AP and AR values of the YOLOX detection model are equivalent to those of Faster R-CNN and are evidently superior to those of YOLOv3 and SSD models. According to the results, YOLOX can quickly detect walnut shell–kernel in the target detection model. The YOLOX algorithm is a relatively good selection for online walnut shell–kernel separation with considerations to both recognition precision and detection speed.
3.4. Walnut Shell–Kernel Detection Effect Analysis under Different Scenes
3.4.1. Detection Analysis of Different Walnut Species Based on the Network Model
Shells and kernels of Wen 185 and Yunnan Juglans sigillata were detected using the YOLOX network (Figure 7). According to detection, the AP50 values were 96.8% and 95.7%, the AR values were 84.8% and 80.5%, and the APs values were 80.9% and 76.3%, respectively. The results demonstrated that the YOLOX network could still accurately distinguish the shells and kernels of different walnut species with relatively high detection confidence.
3.4.2. Walnut Shell–Kernel Detection Effect Analysis under Different Illumination Intensities
To verify the walnut shell–kernel recognition effect of the model under different illumination conditions, the shells and kernels of Wen 185 and Yunnan Juglans sigillata were tested using the YOLOX algorithm in this study. The AP50 of detection results were 95.7% and 95.9%. The AR values were 79.8% and 85.2%. The APs values were 75.5% and 81.3%. Figure 8 depicts the detection effects. To summarise, the YOLOX network can accurately recognise the walnut shells and kernels under different illumination conditions, with relatively high detection confidence.
3.4.3. Walnut Shell–Kernel Detection Effect under Mutual Shielding
To evaluate the walnut shell–kernel recognition effect under complicated conditions, the shells and kernels of Wen 185 and Yunnan Juglans sigillata with mutual shielding were tested using the YOLOX algorithm. The AP50 values were 95.8% and 96.4%. The AR values were 82% and 83.7%. The APs values were 78.1% and 79.8%. Figure 9 shows the detection effects. The YOLOX network could accurately recognise walnut shells and kernels accurately under mutual shielding with relatively high detection confidence.
4. Discussions
Table 2 demonstrates the superiority of the YOLOX detection model over YOLOV3, Faster-RCNN, and SSD. To achieve this, YOLOX uses the Mosaic data enhancement strategy for input and replaces the traditional YOLO baseline header with a decoupling baseline, significantly improving the convergence speed. In addition, SimOTA is used, reducing training time and avoiding extra solver hyperparameters in the Sinkhorn Knopp algorithm, thereby improving detection accuracy and efficiency. In this paper, YOLOX obtained good recognition accuracy for walnut shell kernel detection in different scenarios, including complex cases such as occluded shell kernels. This performance meets the requirements for online real-time detection of walnut shell kernels. Yu et al. [34] used an improved YOLOv5 series algorithm to detect mixed materials in walnut kernels, achieving an mAP value of 88.9%. Meanwhile, Pham et al. [35] used the YOLOv7 series algorithm to detect and identify the good, bad, or incomplete cashews on a packaging production line, with an average identification accuracy (mAP) of about 90%, demonstrating good results. While near-infrared spectroscopy is also an important tool in nut detection [36], it demands higher hardware requirements. YOLO-based detection technology has higher detection efficiency, more affordable detection equipment, and a wider range of application objects. Moreover, it has certain advantages in detection accuracy, making it suitable for current online detection and identification requirements for walnut shell kernels.
However, it is important to note that the YOLOX model does have limitations. As shown in Figure 10, after the walnut shell breaks, the shell kernels exhibit different shapes, and their appearance closely resembles that of walnut shells. When shell kernels densely block each other and have similar appearances, the detection performance of walnut shells and kernels is slightly poor, resulting in lower confidence scores. Sometimes, the model may mistakenly predict two closely contacted and shielded walnut shells or kernels within a box, resulting in target positioning errors or reduced confidence. In addition, when exposed to strong light, the detection accuracy is limited due to overexposure in the central region of images caused by manual photo acquisition, resulting in increased similarity between shell kernels and unclear details [37]. The YOLOX detection model struggles to extract features effectively when processing walnut shell kernels with similar appearances, resulting in reduced confidence in some images during detection.
In future research work, the utilisation of scientific image acquisition can be employed to further improve the photo accuracy of walnut shells and kernels [38,39]. In addition, the future should aim to create a more lightweight detection model, exploring the replacement of the backbone network with other lightweight alternatives and reducing the number of model parameters.
5. Conclusions
The YOLOX algorithm is applied in this study to realise the fast recognition and accurate separation of walnut shells and kernels after breaking. Some significant conclusions could be drawn as follows:
(1). For walnut shell–kernel detection, the AP50, APs, and AR of the YOLOX algorithm are 96.3%, 80.6%, and 84.7%, respectively. The model size was 99 MB, and the FLOPs were 351.9. The AR of the YOLOX target detection algorithm is increased by 10%, 2.3%, and 9% than those of YOLOv3, Faster R-CNN, and SSD target detection algorithms. Moreover, APs increased by 12.1%, 3.9%, and 9.8%, respectively. Moreover, YOLOX has apparent advantages in the model size and detection speed. It can decrease the consumption of memory to a great extent during model training, which is beneficial for the migration application of the model.
(2). Under different walnut species, supplementary light, and shielding conditions, AP50 of the YOLOX algorithm is higher than 95%, and AR is higher than 79%. The YOLOX algorithm can realise accurate walnut shell–kernel recognition and has good robustness. Research conclusions can provide technological support to walnut shell–kernel separation.
Resources, Y.Z.; data curation, X.W.; writing—original draft preparation, Y.Z. and J.M.; writing—review and editing, X.W. and Y.L.; visualization, H.L. supervision, Z.L.; project administration, H.L. and Z.Z. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
The data presented in this study are available upon request from the corresponding author.
The authors thank Hong Zhang from Tarim University for thesis supervision. The authors are grateful to the anonymous reviewers for their comments.
The authors declare no conflict of interest.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 2. Images of walnut shell–kernels under different scenes: (a) natural light, (b) supplementary light, (c) with mutual shielding, and (d) without shielding.
Figure 5. Detection results of the YOLOX target detection algorithm. (a) Detection results of Yunnan Juglans sigillata. (b) Detection results of Wen 185.
Figure 6. Normalized confusion matrices of detection results. (a) YOLOX. (b) YOLOv3. (c) Faster-RCNN. (d) SSD.
Figure 7. Detection effects of different walnut species. (a) Wen 185. (b) Yunnan Juglans sigillata.
Figure 8. Detection effects under natural light and supplementary light. (a) Natural light. (b) Supplementary light.
Figure 9. Detection effect with and without shielding conditions. (a) With shielding. (b) Without shielding.
Figure 10. Detection effects under complicated targets and strong illumination. (a) Detection effect of complicated targets. (b) Detection effect under strong illumination. (c) Detection effect under similar appearances.
Performance comparison of several target detection algorithms.
Algorithm | Model Size/MB | FLOPs | AP50(%) | AP75(%) | APs(%) | AR(%) |
---|---|---|---|---|---|---|
YOLOX | 99 | 351.9 | 96.3 | 92.4 | 80.6 | 84.7 |
YOLOv3 | 61.53 | 193.87 | 94.2 | 83.7 | 68.5 | 74.7 |
Faster-RCNN | 98.85 | 427.07 | 95 | 89.1 | 76.7 | 82.4 |
SSD | 3.04 | 7.02 | 92.9 | 82.6 | 70.9 | 75.7 |
Detection results of several algorithms under different conditions.
Algorithm | Considered Factors | Sample Condition | AP50 (%) | APs (%) | AR (%) |
---|---|---|---|---|---|
YOLOX | Walnut species | Wen 185 | 96.8 | 80.9 | 84.8 |
Yunnan Juglans sigillata | 95.7 | 76.3 | 80.5 | ||
Light source | Supplementary light | 95.7 | 75.5 | 79.8 | |
Natural light | 95.9 | 81.3 | 85.2 | ||
Shielding condition | With mutual shielding | 95.8 | 78.1 | 82.0 | |
Without shielding | 96.4 | 79.8 | 83.7 | ||
YOLOv3 | walnut species | Wen 185 | 95.2 | 70.0 | 75.7 |
Yunnan Juglans sigillata | 94.1 | 65.2 | 71.8 | ||
Light source | Supplementary light | 94.2 | 64.6 | 71.4 | |
Natural light | 95.0 | 70.4 | 75.9 | ||
Shielding condition | With mutual shielding | 94.5 | 65.6 | 71.6 | |
Without shielding | 95.8 | 71.2 | 76.6 | ||
Faster |
Walnut species | Wen 185 | 95.4 | 79.3 | 84.3 |
Yunnan Juglans sigillata | 95.6 | 74.3 | 80.1 | ||
Light source | Supplementary light | 95.3 | 73.5 | 79.6 | |
Natural light | 95.7 | 79.7 | 84.6 | ||
Shielding condition | With mutual shielding | 95.7 | 75.7 | 81.2 | |
Without shielding | 96.4 | 78.8 | 83.8 | ||
SSD | walnut species | Wen 185 | 94.2 | 72.0 | 76.7 |
Yunnan Juglans sigillata | 92.9 | 67.3 | 72.7 | ||
Light source | Supplementary light | 93.0 | 67.0 | 72.4 | |
Natural light | 93.6 | 72.0 | 76.5 | ||
Shielding condition | With mutual shielding | 92.8 | 67.1 | 71.9 | |
Without shielding | 96.4 | 78.8 | 83.8 |
References
1. An, M.; Cao, C.; Wu, Z.; Luo, K. Detection Method for Walnut Shell-Kernel Separation Accuracy Based on Near-Infrared Spectroscopy. Sensors; 2022; 22, 8301. [DOI: https://dx.doi.org/10.3390/s22218301] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36365998]
2. Liu, M.; Li, C.; Cao, C.; Tu, D.; Li, X.; Che, J.; Yang, H.; Zhang, X.; Shi, M.; Zhao, H. et al. Research progress of key technology and device for size-grading shell-breaking and shell-kernel separation of walnut. Trans. Chin. Soc. Agric. Eng.; 2020; 36, pp. 294-310.
3. Niu, H. Experimental Study and Design of Separation Device of Walnut Shell and Kernel. Master’s Thesis; Tarim University: Alaer, China, 2017.
4. Jiang, L.; Zhu, B.; Rao, X.; Berney, G.; Tao, Y. Discrimination of black walnut shell and pulp in hyperspectral fluorescence imagery using Gaussian kernel function approach. J. Food Eng.; 2006; 81, pp. 108-117. [DOI: https://dx.doi.org/10.1016/j.jfoodeng.2006.10.023]
5. Nahal, A.M.; Arabhosseini, A.; Kianmehr, M.H. Separation of shelled walnut particles using pneumatic method. Int. J. Agric. Biol. Eng.; 2013; 6, pp. 88-93.
6. Krishnan, P.; Berlage, A.G. Separation of shells from walnut meats using magnetic methods. Trans. ASAE; 1984; 27, pp. 1990-1992. [DOI: https://dx.doi.org/10.13031/2013.33083]
7. Wang, Z.; Xiao, W. Electrostatic Fruit Shell Kernel Separator. CN2041490, 26 July 1989. Available online: https://kns.cnki.net/kcms2/article/abstract?v=kxaUMs6x7-4I2jr5WTdXtkOSbVhUnTwo_UJJAd4NGDqpeyCQzr6nn8GZTDCXBoMWz67kms7HBWUKA5AE-8ynjg%3d%3d&uniplatform=NZKPT (accessed on 6 June 2023).
8. Liu, M.; Li, C.; Cao, C.; Wang, L.; Li, X.; Che, J.; Yang, H.; Zhang, X.; Zhao, H.; He, G. et al. Walnut Fruit Processing Equipment: Academic Insights and Perspectives. Food Eng. Rev.; 2021; 13, pp. 822-857. [DOI: https://dx.doi.org/10.1007/s12393-020-09273-6]
9. Cao, C.; Li, Z.; Luo, K.; Mei, P.; Wang, T.; Wu, Z.; Xie, C. Experiment on Winnowing Mechanism and Winnowing Performance of Hickory Material. Trans. Chin. Soc. Agric. Mach.; 2019; 50, pp. 105-112.
10. Li, H.; Tang, Y.; Zhang, H.; Liu, Y.; Zeng, Y.; Niu, H. Technological parameter optimization for walnut shell-kernel winnowing device based on neural network. Front. Bioeng. Biotechnol.; 2023; 11, 1107836. [DOI: https://dx.doi.org/10.3389/fbioe.2023.1107836]
11. Wang, T.; Cao, C.; Xie, C.; Li, Z. Design of hickory nut’ shell and kernel sorting system based on fuzzy clustering algorithm. Food Mach.; 2018; 34, pp. 110-114.
12. Jin, F.; Qin, L.; Jiang, L.; Zhu, B.; Tao, Y. Novel separation method of black walnut meat from shell using invariant features and a supervised self-organizing map. J. Food Eng.; 2008; 88, pp. 75-85. [DOI: https://dx.doi.org/10.1016/j.jfoodeng.2008.01.022]
13. Fan, X.; Xu, Y.; Zhou, J.; Li, Z.; Peng, X.; Wang, X. Detection system for grape leaf diseases based on transfer learning and updated CNN. Trans. Chin. Soc. Agric. Eng.; 2021; 37, pp. 75-85.
14. Lin, J.; Wu, X.; Chai, Y.; Yin, H. Structure optimization of convolutional neural networks: A survey. Acta Autom. Sin.; 2020; 46, pp. 24-37.
15. Huu-Thiet, N.; Chien, C. Analytic Deep Neural Network-Based Robot Control. IEEE/ASME Trans. Mechatron.; 2022; 27, pp. 2176-2184.
16. Agu, S.; Eze, C.; Omeje, U. Separation of oil palm kernel and shell mixture using soil and palm ash slurries. Niger. J. Technol.; 2017; 36, pp. 621-627. [DOI: https://dx.doi.org/10.4314/njt.v36i2.38]
17. Yao, J.; Qi, J.; Zhang, J.; Shao, H.; Yang, J.; Li, X. Real-Time Detection Algorithm for Kiwifruit Defects Based on YOLOv5. Electronics; 2021; 10, 1711. [DOI: https://dx.doi.org/10.3390/electronics10141711]
18. Xiao, B.; Nguyen, M.; Yan, Q. Fruit ripeness identification using YOLOv8 model. Multimed. Tools Appl.; 2023; 9, [DOI: https://dx.doi.org/10.1007/s11042-023-16570-9]
19. Wang, Z.; Jin, L.; Wang, S.; Xu, H. Apple stem/calyx real-time recognition using YOLO-v5 algorithm for fruit automatic loading system. Postharvest Biol. Technol.; 2022; 185, 111808. [DOI: https://dx.doi.org/10.1016/j.postharvbio.2021.111808]
20. Wu, Z.; Luo, K.; Cao, C.; Liu, G.; Wang, E.; Li, W. Fast location and classification of small targets using region segmentation and a convolutional neural network. Comput. Electron. Agric.; 2020; 169, 105207. [DOI: https://dx.doi.org/10.1016/j.compag.2019.105207]
21. Meng, J.; Kang, F.; Wang, Y.; Tong, S.; Zhang, C.; Chen, C. Tea Buds Detection in Complex Background Based on Improved YOLOv7. IEEE Access; 2023; 11, pp. 88295-88304. [DOI: https://dx.doi.org/10.1109/ACCESS.2023.3305405]
22. Zhang, Y.; Zhang, W.; Yu, J.; Hei, L.; Chen, J.; He, Y. Complete and accurate holly fruits counting using YOLOX object detection. Comput. Electron. Agric.; 2022; 198, 107062. [DOI: https://dx.doi.org/10.1016/j.compag.2022.107062]
23. Zhang, H.; Ji, S.; Shao, M.; Pu, H.; Zhang, L. Non-destructive Internal Defect Detection of In-Shell Walnuts by X-ray Technology Based on Improved Faster R-CNN. Appl. Sci.; 2023; 13, 7311. [DOI: https://dx.doi.org/10.3390/app13127311]
24. Wang, D.; Dai, D.; Zheng, J.; Li, L.; Kang, H.; Zheng, X. WT-YOLOM: An Improved Target Detection Model Based on YOLOv4 for Endogenous Impurity in Walnuts. Agronomy; 2023; 13, 1462. [DOI: https://dx.doi.org/10.3390/agronomy13061462]
25. Zhang, G.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv; 2021; arXiv: 2107.08430
26. Shen, C.; Ma, C.; Gao, W. Multiple Attention Mechanism Enhanced YOLOX for Remote Sensing Object Detection. Sensors; 2023; 23, 1261. [DOI: https://dx.doi.org/10.3390/s23031261]
27. Shang, Y.; Zhang, Q.; Song, H. Application of deep learning based on YOLOv5s to apple flower detection in natural scenes. Trans. Chin. Soc. Agric. Eng.; 2022; 38, pp. 222-229.
28. Yang, Z.; Zhao, C.; Maeda, H.; Sekimoto, Y. Development of a Large-Scale Roadside Facility Detection Model Based on the Mapillary Dataset. Sensors; 2022; 22, 9992. [DOI: https://dx.doi.org/10.3390/s22249992] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36560361]
29. Chang, Y.; Zhang, Y. Deep Learning for Clothing Style Recognition Using YOLOv5. Micromachines; 2022; 13, 1678. [DOI: https://dx.doi.org/10.3390/mi13101678]
30. Huang, G. A Comparative study of underwater marine products detection based on YOLOv5 and underwater image enhancement. Int. J. Eng.; 2021; 7, pp. 213-221.
31. Song, X.; Wu, Y.; Liu, B.; Zhang, Q. Improved YOLOv5s Algorithm for Helmet Wearing Detection. Comput. Eng. Appl.; 2023; 59, pp. 176-183.
32. Liu, J. Research and Software Design of Transmission Line Foreign Objects Detection Algorithm Based on YOLOX. Master’s Thesis; China University of Mining and Technology: Xuzhou, China, 2022.
33. Zhao, Y.; Han, R.; Rao, Y. A New Feature Pyramid Network for Object Detection. Proceedings of the 2019 International Conference on Virtual Reality and Intelligent Systems; Hunan, China, 14 September 2019.
34. Yu, L.; Qian, M.; Chen, Q.; Sun, F.; Pan, J. An Improved YOLOv5 Model: Application to Mixed Impurities Detection for Walnut Kernels. Foods; 2023; 12, 624. [DOI: https://dx.doi.org/10.3390/foods12030624]
35. Pham, V.; Nguyen, N.; Pham, V. A Novel Approach to Cashew Nut Detection in Packaging and Quality Inspection Lines. Int. J. Adv. Comput. Sci. Appl. (IJACSA); 2022; 13, pp. 356-361. [DOI: https://dx.doi.org/10.14569/IJACSA.2022.0131243]
36. Ma, L.; Ma, J.; Han, J.; Li, Y. Research on target detection algorithm based on YOLOv5s. Comput. Knowl. Technol.; 2021; 17, pp. 100-103.
37. Rajevenceltha, J.; Gaidhane, V. An efficient approach for no-reference image quality assessment based on statistical texture and structural features. Eng. Sci. Technol.; 2022; 30, 101039. [DOI: https://dx.doi.org/10.1016/j.jestch.2021.07.002]
38. Rajevenceltha, J.; Gaidhane, H. A novel approach for image focus measure. Signal Image Video Process.; 2020; 15, pp. 547-555. [DOI: https://dx.doi.org/10.1007/s11760-020-01775-4]
39. Al, N.; Gaidhane, V.H.; Rajevenceltha, J. Image Focus Measure Based on Polynomial Coefficients and Reduced Gerschgorin Circle Approach. IETE Tech. Rev.; 2023; [DOI: https://dx.doi.org/10.1080/02564602.2023.2227145]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Walnut shell–kernel separation is an essential step in the deep processing of walnut. It is a crucial factor that prevents the increase in the added value and industrial development of walnuts. This study proposes a walnut shell–kernel detection method based on YOLOX deep learning using machine vision and deep-learning technology to address common issues, such as incomplete shell–kernel separation in the current airflow screening, high costs and the low efficiency of manually assisted screening. A dataset was produced using Labelme by acquiring walnut shell and kernel images following shellshock. This dataset was transformed into the COCO dataset format. Next, 110 epochs of training were performed on the network. When the intersection over the union threshold was 0.5, the average precision (AP), the average recall rate (AR), the model size, and floating point operations per second were 96.3%, 84.7%, 99 MB, and 351.9, respectively. Compared with YOLOv3, Faster Region-based Convolutional Neural Network (Faster R-CNN), and Single Shot MultiBox Detector algorithms (SSD), the AP value of the proposed algorithm was increased by 2.1%, 1.3%, and 3.4%, respectively. Similarly, the AR was increased by 10%, 2.3%, and 9%, respectively. Meanwhile, walnut shell–kernel detection was performed under different situations, such as distinct species, supplementary lighting, or shielding conditions. This model exhibits high recognition and positioning precision under different walnut species, supplementary lighting, and shielding conditions. It has high robustness. Moreover, the small size of this model is beneficial for migration applications. This study’s results can provide some technological references to develop faster walnut shell–kernel separation methods.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details


1 Modern Agricultural Engineering Key Laboratory at Universities of Education Department of Xinjiang Uygur Autonomous Region, Tarim University, Alaer 843300, China;
2 Modern Agricultural Engineering Key Laboratory at Universities of Education Department of Xinjiang Uygur Autonomous Region, Tarim University, Alaer 843300, China;
3 Modern Agricultural Engineering Key Laboratory at Universities of Education Department of Xinjiang Uygur Autonomous Region, Tarim University, Alaer 843300, China;