1. Introduction
Fruits, essential in diets, are rich in vitamins, dietary fibers, polyphenols, and essential minerals, and are widely recognized for their health benefits [1]. However, it is crucial to consider that the quality of fruit is dictated by the ripening process which may take place during long transportation periods [2]. Fruit ripening results in changes in the color, texture, aroma, and nutritional quality of fresh fruit. It is also accompanied by the release of other volatile organic compounds (VOCs), which comprise diverse classes of chemicals [3], including water, esters, alcohols, aldehydes, ketones, lactones, and terpenoids. This diverse range of compounds can usually act together to reinforce or interfere with the sensor signal. Therefore, in such cases, it is more convenient to develop global selectivity devices, i.e., non-selective sensors, which are capable of detecting the differences in the breathing patterns of each class of fruit [4]. The ripening of a unique fruit may affect the others and, by extension, impact the overall product quality to final consumers. Therefore, effective fruit ripening monitoring is an attractive approach to ensure standard quality control and predict shelf life.
Sensor devices, such as artificial olfactory systems, are becoming increasingly popular for the detection of environmental pollutants [5], explosive compounds [6], food safety [7], and disease [8] due to their wide detection range, cost-effectiveness, portability, and non-destructiveness. Consequently, these systems offer a new method for monitoring fruit quality as a replacement for human specialists who take advantage of visual features in the grading process. The various gases emitted by fruit can interact with the dyes present on the colorimetric sensor array, resulting in the formation of a distinctive fingerprint pattern [9] that can be analyzed using pattern recognition techniques to predict the degree of fruit ripeness based on the observed changes in color. The development of naked-eye-detected artificial olfactory systems in smart food packaging has garnered considerable attention due to their ability to provide on-packaging indicators of fruit maturity and ripeness to consumers. However, the commercialization of artificial olfactory systems is currently hindered by the challenge of accurately distinguishing complex patterns. While colorimetric sensors (or opto-noses) are highly cross-reactive, they produce nonlinear and multidimensional datasets [10,11,12]. In addition, variable lighting conditions can significantly impair pattern discernment in image-based recognition systems, thereby reducing identification accuracy. Traditional statistical techniques such as principal component analysis (PCA) [13,14,15], hierarchical clustering analysis (HCA) [16,17,18,19], and discriminant analysis (LDA) [20,21,22], are unable to meet the needs for the accurate extraction and analysis of complex image datasets.
To tackle these challenges, the deep convolutional neural network (DCNN) offers a powerful method compared with traditional statistical techniques. PCA and HCA primarily focus on descriptive analysis, limiting their predictive capabilities, whereas DCNNs are more flexible and tailored for classification, making them more suitable for these predictive tasks. Traditional classification methods like LDA are fundamentally linear and often fall short when applied to nonlinear data like images. In contrast, DCNNs excel by extracting complex features and learning intricate patterns, thereby effectively processing multidimensional image data and reducing environmental variability [23,24]. Guo et al. utilized ResNet101 for the colorimetric detection of meat freshness [25], achieving an impressive accuracy of 98.5%. They used 20 unique types of porous nanocomposites made of chitosan, dye, and cellulose acetate and integrated a DCNN into a smartphone application to rapidly identify food freshness. Huang et al. combined DCNN, PAN-NSS, and an app to develop a rapid, long-lasting nitrite sensor system in food products [26]. They extended sensor life from 7 to over 30 days and enhanced the accuracy of food classification to 91.33–100.00%. Ma et al. developed a colorimetric sensor array using six chromogenic indicators loaded with UiO-66-OH to assess chicken freshness [27]. They used the WISeR-50 algorithm and achieved a 98.95% accuracy rate. With the advancement of deep learning, the densely connected convolutional network (DenseNet) [28] has emerged as a novel algorithm with distinct advantages in complex image processing. The efficient use of dense connections not only enhances feature extraction and reuse but also minimizes overfitting. Thus, this study integrated DenseNet with colorimetric sensor arrays to evaluate fruit ripeness.
Therefore, the original objectives of the present work were to simulate the human olfactory system (Figure S1a), and establish a low-cost, non-destructive, high accuracy fruit ripeness prediction system, employing colorimetric sensor techniques integrated with DenseNet (Figure S1b). The GC−MS method was first used to determine the characteristic VOCs of mango, peach, and banana at different ripening stages. Then, colorimetric sensing combinatorics with multiple dyes were used to enable the sensitive detection of various characteristic VOCs and formed scent fingerprints. Finally, the DenseNet model was used to extract fingerprint information and categorize fruit ripeness.
2. Materials and Methods
The overview of the methodological framework is presented in Figure 1. Initially, a fruit firmness test and sensory evaluation were employed to determine the stage of fruit ripeness. Subsequently, gas chromatography–mass spectrometry (GC-MS) analysis was utilized to determine the characteristic VOCs emitted by the fruits at different stages of ripening. Thereafter, the prepared characteristic VOCs with different concentrations were analyzed using the colorimetric sensor arrays and classified using the HCA method. The images of the colorimetric sensor arrays before and after their reactions with the VOCs released from fruits were captured using a camera (Canon EOS 70D, Canon Inc., Tokyo, Japan) and coupled with the DenseNet model to predict the fruit’s ripeness.
2.1. Chemicals and Materials
All the dyes (Table S1) were purchased from Sigma-Aldrich (Merck KGaA, St. Louis, MI, USA). Anhydrous ethanol (≥99.5%) and ethyl acetate were purchased from Sinopharm. Trans-2-hexenal and benzaldehyde were purchased from Sigma-Aldrich. Hexyl acetate, (+)-limonene, β-myrcene, 3-carene, and isoamyl acetate were purchased from Aladdin Biochemistry Technology Co., Ltd (Shanghai, China).
2.2. Fabrication of the Colorimetric Sensor Arrays
For the fabrication of the colorimetric sensor arrays, initially, 5 mg of each dye was dissolved in 5 mL of anhydrous ethanol using ultrasonication for 10 min to obtain a clear solution. Subsequently, a 3 cm × 3 cm Polyvinylidene fluoride (PVDF) membrane with pore diameters of 0.22 µm was used as a substrate. Then, a volume of 2.5 µL of each dye solution in 25 species were successively deposited onto the membrane’s surface in a 5 × 5 grid pattern, ensuring a 6 mm distance between adjacent dye spots. Following this, the prepared colorimetric arrays were heated at 75 °C on a heating plate for 30 min to guarantee the complete evaporation of ethanol, then cooled to room temperature for 30 min, and finally stored in a brown vacuum container filled with N2 for at least three days before measurements.
2.3. Fruit Firmness Test
A texture analyzer (TA−XT2i, Stable Micro Systems, Godalming, UK) equipped with a P/2 cylindrical probe was used to conduct puncture tests on the fruit with a testing speed of 1 mm s–1 and a penetration depth of 8 mm. Firmness was measured at three equidistant locations around the equator of the fruit. The mean firmness value of the fruit for each day was calculated from three replicates.
2.4. Sensory Evaluation of the Fruits
In this study, five experienced fruit farmers were invited to assess and score the ripeness of three fruits (mango, banana, and peach) using sensory analysis. Participants ranged in age from 40 to 50 years, and gender distribution was balanced (3 men and 2 women).
Following the ISO 8587 (2007) [29] standard, the assessment criteria were based on sensory characteristics such as smell, vision, taste, and touch. The assessment comprehensively considered the color change, firmness, texture, and smell of the fruit during ripening. Mangoes transition from a green, hard state with a weak aroma to a yellow, soft, and juicy one with a strong aroma and eventually become dark with spots or mold. Peaches shift from green with a mild aroma to pink, soft, and juicy with a sweeter aroma and end up dark, possibly with mold or bruises. Bananas change from green with a subdued scent to yellow, soft, and creamy with a distinct aroma and later turn brown with potential black spots.
A scale of 1–10 was used for scoring, with higher scores indicating higher fruit ripeness. A score below 4 indicates that the fruit is unripe, a score of 4 to 7 indicates that the fruit is in the ripening stage, while a score above 7 means that the fruit is overripe. All scoring was performed in a sensory laboratory equipped with separate compartments according to ISO 8589 (2007) [30], and it was ensured that the evaluation process was carried out under standardized white light conditions to minimize interference from environmental factors.
2.5. GC-MS Analysis of the Characteristic VOCs Emitted by the Fruits during Different Ripening Stages
The characteristic VOCs were extracted using a solid-phase microextraction (SPME) method [31,32,33,34]. For each sample, 5 g of fruit was ground and added to a headspace vial. The non-polar divinylbenzene/carboxen/polydimethylsiloxane (DVB/CAR/PDMS) fibers then penetrated the septa to extract the volatiles for 40 min at 45 °C. GC-MS analysis was carried out on a 7890B GC System with a 7000C GC-MS Triple Quad mass detector (Agilent Technologies, Santa Clara, USA). The SPME fiber was quickly inserted into the GC-MS system and thermally desorbed at 250 °C for 3 min. Helium of high purity (99.999%) was used as the carrier gas, and the flow rate was maintained at 1.0 mL min–1. The column temperature was increased following a heating program, with an initial temperature of 50 °C maintained for 3 min, followed by a 10 °C min–1 ramp to 250 °C and a 5 min holding period. The mass spectra were analyzed using the NIST17.L library.
2.6. Preparation of the Characteristic VOCs
Gas mixtures were prepared according to previous methods [35], with the setup illustrated in Figure S2. Three mass flow controllers (CSC200-C, Sevenstar, Beijing, China) were employed to control the flow rate of mixture gas, including VOCs, dry N2, and wet N2, achieving different concentrations of the characteristic volatiles (Figure S2). The VOCs stream was maintained for 30 min in the pipelines to stabilize the analyte concentration. Subsequently, the colorimetric sensor arrays were placed within a gas chamber under exposure to the VOCs vapor.
2.7. Raw Data Acquisition and Process for the Characteristic VOCs
The sensor arrays were placed in the reaction chamber and reacted to the characteristic volatiles for 30 min, and the images of the sensor arrays before and after exposure were captured by a standard flatbed scanner (Epson Perfection V600, Seiko Epson Corporation, Suwa, Nagano, Japan). The RGB differential values (ΔR, ΔG, ΔB) of each spot in pre-exposure and post-exposure images were extracted and calculated using Python. The color differential maps were then generated by expanding the RGB range of 3–10 to 0–255. Euclidean Distance (ED = Euclidean Distance) was utilized as a quantitative measure to describe the detection response, obtained by subtracting the ΔR, ΔG, and ΔB values (ED = ) [36,37,38]. HCA based on ED coupled with Ward’s minimum variance method was processed using the Origin software (OriginPro 2021, OriginLab Corporation, Northampton, MA, USA). All experimental data are the average of a minimum of three replicates.
2.8. Images Collection for Real Samples
In this study, each unripe fruit was placed in a transparent box with colorimetric sensor arrays attached to the interior surface of a transparent container (Figure S3). They were placed in a climatic chamber at a constant temperature of 25 °C and a humidity level of 95% which are common fruit storage conditions. Every 12 h, one image of each colorimetric sensor array was taken with a camera under arbitrarily transformed light and angles which can enhance the robustness of deep learning models by providing a comprehensive dataset. Finally, a dataset of 14,778 colorimetric sensor array images, to be used as a train and validation set, was collected to classify ripeness into ten classes: mango_unripe, mango_ripe, mango_overripe, peach_unripe, peach_ripe, peach_overripe, banana_unripe, banana_ripe, banana_overripe, and blank. The train and validation set was randomly selected from the dataset in a ratio of 17:3.
Separate test data, derived from different batches of fruits, was used to verify the robustness of the model. One image of each colorimetric sensor array was captured using a camera under arbitrarily transformed light and angles every 12 h. Finally, a dataset of 1370 colorimetric sensor images documenting the progression of fruits from their unripe to overripe stages was collected as the test set.
2.9. DenseNet Model Architecture
The DenseNet model (Figure S4) consists of an input layer, multiple convolutional layers, a spatial pyramid pooling (SPP) structure, a fully connected layer, and an output layer. This model utilizes five identical dense blocks, each of which contain three 3 × 3 convolutional layers, batch normalization, and a ReLU activation function, with a growth rate of 32. A transition module is used between every two dense blocks for downsampling transformation, consisting of a batch normalization layer, 1 × 1 convolutional layer, and 2 × 2 avgpooling.
2.10. Data Process for Deep Learning
For image processing, all the images were resized to 255 × 255 pixels as input. Four DCNN models, google inception net (GoogleNet) [39], inception version 3 (Inception_v3) [40], residual network 18 (ResNet18) [41], and DenseNet models were implemented in Pytorch (1.8.1 + cu111). All model hyperparameters were set equally. The models implemented cross-entropy loss as the loss function, stochastic gradient descent (SGD) as the optimizer, momentum as 0.9, and weight decay as 0.0001. The training process consisted of 500 epochs with a learning rate scheduler for stable performance. The learning rate was initially set to 0.01 and was gradually reduced to 0.005 between the 300th and 450th epochs, and then to 0.001 for the remaining epochs. A batch size of 256 was used for the train set. DCNN models were implemented using Python on a computer equipped with RTX 3090 GPUs.
3. Results and Discussion
3.1. GC-MS Analysis of Fruit Characteristic Volatiles at Different Ripening Stages
The degree of banana ripeness was determined by the integration of firmness test and sensory evaluation. It was observed that banana firmness (Figure S5c) decreased progressively with increased storage time, with the rate showing significant changes at 2 days and 6 days. The color transformation of bananas was documented (Figure S6), transitioning from light green (unripe) to yellow (ripe), and ultimately to brown (overripe). Average sensory evaluations indicated that bananas were generally unripe between 0 ≤ days < 2, ripe between 2 ≤ days < 6, and overripe between 6 ≤ days ≤ 8 according to sensory score ratings, as detailed in Table S2. These sensory stages correspond closely with the observed changes in banana firmness. Likewise, mangoes were classified into three ripeness stages based on the firmness test (Figure S5a), and sensory evaluation (Table S3, Figure S7), with 0 ≤ days < 3 for unripe, 3 ≤ days < 8 for ripe, and 8 ≤ days ≤ 10 for overripe. Peaches were classified into three ripeness stages based on the firmness test (Figure S5b), and sensory evaluation (Table S4, Figure S8), with 0 ≤ days < 2 for unripe, 2 ≤ days < 7 for ripe, and 7 ≤ days ≤ 10 for overripe.
The VOCs released from the fruit (mango, peach, and banana) during different ripening stages were monitored using the GC-MS method. According to the relative abundances and variation patterns at various ripening stages, the representative characteristic VOCs were identified (Table S5) [42,43,44,45,46,47,48,49]. (+)-limonene, β-myrcene, and 3-carene were identified as the characteristic VOCs for mango. Benzaldehyde, ethyl acetate, and hexyl acetate were identified as the characteristic VOCs for peach. Ethanol, trans-2-hexenal, and isoamyl acetate were identified as the characteristic VOCs for banana.
3.2. Sensor Response to the Individual Gas Analyte
The gas-sensitive dyes for fabricating the colorimetric sensor arrays were selected according to the characteristic VOCs released from the fruit. The identified characteristic VOCs are classified as hydrocarbons, esters, alcohols, and aldehydes. Thus, a diverse range of dye types was selected, including two aldehyde/ketone-sensitive dyes, three solvatochromic dyes, three redox dyes, three Lewis acidic dyes, and fourteen pH indicators, which are sensitive to the aforementioned characteristic VOCs.
The fabricated colorimetric sensor arrays were employed to detect the various concentrations of fruit characteristic VOCs. Under exposure to trans-2-hexenal with concentrations ranging from 3–1000 ppm (Figure S9), the images of the colorimetric sensor arrays exhibited color change. The RGB differential maps of trans-2-hexenal at concentrations of 20 ppm, 100 ppm, 250 ppm, 500 ppm, 1000 ppm, and 1500 ppm were obtained by subtracting the pre-exposure image from each post-exposure image (Figure 2a). As the trans-2-hexenal concentrations increased, the RGB differential map showed deeper color, implying concentration-dependent response characteristics of the colorimetric sensor arrays.
An ED heatmap analysis was performed to show the intensity variation (Figure 2b). ED was utilized as a quantitative measure to describe the signal intensity obtained by extracting the ΔR, ΔG, and ΔB values (ED = ). Based on the heatmap analysis, dyes of pararosaniline, merocyanine 540, o-tolidine, o-dianisidine, bromophenol blue, cresol red, 3,3,5,5-tetraiodophenolsulfonphthalein, and leuco malachite green were identified as the most sensitive dyes to trans-2-hexenal. The ED values of these sensitive dyes to trans-2-hexenal followed a similar trend to the total ED values of the colorimetric sensing combinatorics, both exhibiting a noticeable increase with the concentration of trans-2-hexenal (Figure 2c). The limit of detection (LOD) was the minimum concentration whose corresponding ED value is above the mean of the blank control (EDblank = 47.75) plus three times its standard deviation (3σ = 6). The LOD for trans-2-hexenal was found to be 10 ppm (Figure 2c).
3.3. Classification Performance for the Multiple VOCs
To evaluate the capacity of the sensor arrays to distinguish multiple VOCs, another eight characteristic VOCs including isoamyl acetate, benzaldehyde, hexyl acetate, ethanol, ethyl acetate, 3-carene, β-myrcene, and (+)-limonene were also tested. As shown in ED heatmaps, the ED value increased with the enhanced VOC concentrations and all of the VOCs showed different intensity variations (Figure S10).
As the concentration of the characteristic VOCs increases, the total ED value of the colorimetric sensing combinatorics rises, with the LOD established using ED for isoamyl acetate (50 ppm), benzaldehyde (20 ppm), hexyl acetate (25 ppm), ethanol (50 ppm), ethyl acetate (250 ppm), 3-carene (25 ppm), β-myrcene (100 ppm), and (+)-limonene (25 ppm) (Figure 3a). Distinguishable patterns for all nine characteristic VOCs at 500 ppm were observed (Figure 3b), demonstrating the capacity of the colorimetric sensor arrays to distinguish between multiple VOCs.
HCA, as a statistical method for grouping data points into clusters based on their similarity, was used to evaluate the distinguishing performance of the colorimetric sensor arrays for the nine characteristic VOCs with different concentrations. Ward’s minimum variance method was used to determine gas-induced variations of the 27 × 25 (9 VOCs × 3 concentrations × 25 ED values) dimensional matrix. The resulting cluster-tree showed that the nine VOCs with different concentrations formed clearly separate clusters with a 100% success rate (Figure 3c). Together, these findings suggest that it is possible to use colorimetric sensor arrays to monitor fruit ripeness.
3.4. Deep Learning-Enabled Fruit Ripeness Recognition
As a proof of concept, the colorimetric sensor arrays were employed to monitor the fruit ripeness of real samples. The colorimetric sensor arrays were attached to a transparent container containing fruit (Figure S3), maintained at a steady temperature of 25 °C and a humidity level of 95%. Images of the sensor arrays were captured at 12 h intervals using a camera. However, in a real fruit ripening detection environment, variations in light and shooting angles present challenges for accurately classifying fruit ripeness levels.
The traditional classified method (ED) cannot solve the problem. To establish a reference range, this work measured the firmness values and sensory evaluations, and calculated the ED values of ten samples for each fruit variety stored at different ripeness levels [25]. Before calculating the ED values, all images were calibrated for color balancing to ensure accurate color extraction under arbitrary lighting conditions according to the marked white and black labels [50] using Photoshop software (Adobe Photoshop Version 22.4.3, Adobe Inc., San Jose, CA, USA). A comparative analysis of pre- and post-image calibration is presented in Figure S11. After color calibration, the dye’s RGB values exhibit closer similarity under varied lighting conditions, showing the effectiveness of color calibration. Based on sensory evaluation and firmness data, fruits were categorized into ten distinct classes for classification tasks: mango_unripe, mango_ripe, mango_overripe, peach_unripe, peach_ripe, peach_overripe, banana_unripe, banana_ripe, banana_overripe, and blank. The ED values’ reference range (Table S6) for each category of fruit ripeness was established accordingly, which was utilized to forecast the ripeness of unknown fruit samples. For each category, twenty colorimetric sensor images were randomly selected to evaluate the prediction accuracy of the ED method based on the ED values’ reference range. However, ED analysis showed an overall accuracy rate of 74.50% (Figure S12), indicating great challenges in distinguishing fruit ripeness. The low accuracy can be attributed to the fact that the values of ED can be affected by photographic conditions such as lighting, zoom, and angle.
Therefore, DCNN models, as a powerful method, were used to identify multidimensional image data and reduce the impact of varied environmental conditions. The dataset consists of 12,484 images for training, 2294 images for validation, and 1370 images for testing (Figure 4a). A ten-category image classification network was designed with various DCNN backbones, each consisting of an input layer, multiple convolution (conv.) layers, fully connected layers (FC), and an output layer. The trained DCNNs efficiently extracted features from colorimetric sensor images. After training, the accuracy of the models’ classification of the array images was evaluated using the validation and test set. When an image was input into the classification network, the system provided the most likely ripeness category for the fruit (blue circles in Figure 4a).
Four models, including DenseNet, GoogleNet, Inception_v3, and ResNet18, were implemented. The GoogleNet model uses “Inception” modules for efficient image feature processing. Inception_v3, an evolution of GoogleNet, utilizes more layers and diverse kernel sizes to enhance feature extraction and efficiency. ResNet18 utilizes residual blocks with cross-layer connections to improve information flow. DenseNet, detailed in our methodology, utilizes dense connections to promote feature reuse and address the vanishing-gradient problem, thereby improving efficiency and performance significantly. With all model hyperparameters set equally, the performance of the four models is compared in terms of training and validation accuracy, F1_score, and test accuracy (Table 1).
The accuracy of the training and validation set was first compared. Examination of the accuracy-loss graphs revealed that, with an increasing number of training epochs, training accuracy improved, and training loss decreased, approaching zero as illustrated in (Figure 4b and Figure S13). All models converged within 200 epochs, achieving peak ripeness prediction accuracy, and indicating effective training. The confusion matrices for the four DCNN models were compared, whose diagonals show the correctly classified sample ratios (Figure 4c and Figures S14–S16). The DenseNet model outperformed others on the validation set with a 97.39% accuracy, followed by GoogleNet, Inception_v3, and ResNet18 with 97.17%, 96.03%, and 95.29% (Figure 4d and Table 1), respectively.
F1_score as an indicator of accuracy was also used to evaluate and compare the performance of the DCNN models, using the following equation:, where “p” and “r” denote precision and recall, respectively. F1_score is an ideal comprehensive metric that considers both precision and recall simultaneously, making it a better choice when seeking a balance between precision and recall. The DenseNet model exhibited the highest F1_score with 0.9712 among the compared models followed by GoogleNet, Inception_v3, and ResNet18 with 0.9683, 0.9560, and 0.9498, respectively, (Figure 4d and Table 1).
The accuracy of the separate test set for the four models was then compared. Test data derived from different batches of fruit evaluates the robustness of the model. The DenseNet model outperformed others on the test set in predicting fruit ripeness with an 82.20% accuracy, followed by GoogleNet, Inception_v3, and ResNet18 with 78.85%, 78.63%, and 76.73% (Table 1), respectively, demonstrating the superior robustness of the DenseNet model. Compared to previous studies, the test accuracy in our work is better or comparable with that of the previous literature (Table S7) [51,52,53,54,55,56].
Due to the DenseNet model’s outstanding performance, Gradient-weighted Class Activation Mapping (Grad-CAM) was applied to provide a more intuitive understanding of this model’s decision-making process. The visual class heatmap, acquired based on the attention mechanism [57], can indicate the features that have a more important impact on the results through darker colors. Heatmaps of the ripeness stages of mangoes (unripe, ripe, and overripe) were analyzed and obtained using Grad-CAM (Figure S17). These maps showed the prominent features affecting the model. The dyes including nile red, disperse orange #3, o-tolidine, m-cresol purple, indigo carmine, basic yellow1, and leuco malachite green are the most influential and sensitive dyes in classifying mango ripeness, and they contribute the most to the final decision. The above results demonstrate the interpretability of the DenseNet model.
All of these results show that DenseNet has superior performance when classifying the olfactory visualized ripeness of fruit. This exceptional ability is mainly attributed to the efficient use of the dense connections in the DenseNet model. These connections not only improve feature extraction and reuse but also significantly reduce the risk of overfitting, which is effective in addressing complex classification tasks.
Overall, our system combines colorimetric sensing arrays with the DenseNet model to successfully identify fruit ripeness with a validation set accuracy of 97.39%. This approach offers numerous advantages when compared to other well-known methods. As seen in Table 2, our method not only classified the ripeness of various kinds of fruits but also had higher accuracy than other well-known methods.
4. Conclusions
In summary, this study presents a new method for the non-destructive monitoring of fruit ripeness by integrating cross-reactive colorimetric sensing combinatorics with the DenseNet model. The colorimetric sensing combinatorics consisted of 25 dyes that were sensitive to volatile gases emitted by fruits, showing cross-reactivity to various types and concentrations of gases. The formation of a unique scent fingerprint can be identified using DCNN. By training on 12,484 images under varying lighting conditions, the DenseNet could learn autonomously and mitigate the impact of illumination on the experimental results, achieving an impressive accuracy rate of 97.39% on the validation test and 82.20% on the test set.
With its attractive features of portability, low cost, and high accuracy, this system has great potential for integration into existing smart packaging or monitoring systems. It provides a simple and non-destructive pfigureattern for ordinary consumers to evaluate the fruit ripeness and benefits the supply chain, especially for costly fruits from farmers and producers. However, there are still challenges for its commercial application. Current testing devices are not portable enough for on-site testing. Furthermore, various fruit varieties emit different complex VOCs. For a new fruit variety, a large number of data samples are required for training using deep learning models to achieve a commercial application.
Looking forward, we anticipate significant advancements in sensor technologies and analytical methods to mitigate these challenges. Enhanced sensor sensitivity and anti-interference capability, coupled with more advanced data analysis techniques, could significantly improve the system’s adaptability, and reduce the need for large datasets. In perspective, such technological and collaborative advancements not only promise to refine the system’s current capabilities but also to broaden its utility across various fields, including food safety and quality assurance.
M.Z.: Conceptualization, Methodology, Investigation, Data curation, Formal analysis, Writing–original draft. Z.Y.: Data curation. H.C.: Data curation. X.W.: Data curation. Y.Y.: Data curation. Y.W.: Conceptualization, Project administration, Funding acquisition, Supervision, Writing–review and editing. All authors have read and agreed to the published version of the manuscript.
Not applicable.
Not applicable.
The data presented in this study are available on request from the corresponding author. The data are not publicly available due to institutional policy regarding data protection.
The authors acknowledge the financial support from the National Natural Science Foundation of China (No. U20A2019).
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Figure 2. Colorimetric sensor array response to trans-2-hexenal. (a) Color differential maps were obtained by extracting R, G, and B values from the images at concentrations of 20 ppm, 100 ppm, 250 ppm, 500 ppm, 1000 ppm, and 1500 ppm of trans-2-hexenal, the RGB range is expanded from 3–10 to 0–255. (b) ED heatmaps of different concentrations of trans-2-hexenal. (c) The ED values of sensitive dyes in response to trans-2-hexenal, similar to the total ED of the colorimetric sensor array, show an increase with increasing concentrations of trans-2-hexenal.
Figure 3. Feasibility of using the colorimetric sensing combinatorics to detect multiple VOCs emitted by fruit. (a) ED value of the colorimetric sensing combinatorics increases with the rising concentration of the characteristic gases. (b) Color differential maps were obtained by extracting R, G, and B values from the images at concentrations of 500 ppm of nine characteristic gases, the RGB range is expanded from 3–10 to 0–255. (c) HCA produced a cluster tree that showed the nine gases could be distinguished at different concentrations, with a 100% success rate.
Figure 4. Colorimetric sensor detection combined with DCNN to detect fruit ripeness. (a) Schematic diagram of fruit ripeness label classification and deep learning for fruit ripeness recognition. (b) The training loss and the training accuracy of the DenseNet model. The training loss measures the dissimilarity between the predicted and actual results. The training accuracy increases with the number of epochs. (c) Confusion matrix of DenseNet for the validation set. (d) Comparison of the validation accuracy and F1_score on the validation set between four DCNN models (DenseNet, GoogleNet, Inception_v3, and ResNet18).
Comparison results among the four DCNN models.
Models | Validation Accuracy (%) | F1_Score | Test Accuracy (%) |
---|---|---|---|
DenseNet | 97.39 | 0.9712 | 82.20 |
GoogleNet | 97.17 | 0.9683 | 78.85 |
Inception_v3 | 96.03 | 0.9560 | 78.63 |
ResNet18 | 95.29 | 0.9498 | 76.73 |
Comparison with other well-known methods of fruit ripeness classification.
Method | Fruit | Classification Method | Accuracy/Correlation Coefficient | Ref. |
---|---|---|---|---|
RGB | Mango | Fuzzy logic | 87% | [ |
VIS/NIR | Watermelon | ANN | 80% | [ |
Acoustic method | Mango | - | 0.957 | [ |
Electronic Nose | Peaches and pears | - | 92% | [ |
RGB | Blueberry | KNN, etc. | 85–98% | [ |
HSV | Mango | Neural Network | 95% | [ |
Electrical method | Banana | - | 0.94 | [ |
RGB | Mango | Gaussian Mixture model and Fuzzy logic | Less than 90% in all varieties | [ |
Bioimpedance data | Strawberry | MLP network | 0.72 | [ |
Colorimetric sensor arrays | Banana, Peach, Mango | DenseNet | 97.39% | This study |
Supplementary Materials
The following supporting information can be downloaded at:
References
1. Hu, B.; Sun, D.-W.; Pu, H.; Wei, Q. Recent advances in detecting and regulating ethylene concentrations for shelf-life extension and maturity control of fruit: A review. Trends Food Sci. Technol.; 2019; 91, pp. 66-82. [DOI: https://dx.doi.org/10.1016/j.tifs.2019.06.010]
2. Harker, F.R.; Hunter, D.; White, A.; Richards, K.; Hall, M.; Fullerton, C. Measuring changes in consumer satisfaction associated with kiwifruit ripening: A new approach to understand human-product interactions with fruit. Postharvest Biol. Technol.; 2019; 153, pp. 118-124. [DOI: https://dx.doi.org/10.1016/j.postharvbio.2019.03.010]
3. Yang, Y.; Zheng, F.; Yu, A.; Sun, B. Changes of the free and bound volatile compounds in Rubus corchorifolius L. f. fruit during ripening. Food Chem.; 2019; 287, pp. 232-240. [DOI: https://dx.doi.org/10.1016/j.foodchem.2019.02.080]
4. Chen, H.; Zhang, M.; Bhandari, B.; Guo, Z. Applicability of a colorimetric indicator label for monitoring freshness of fresh-cut green bell pepper. Postharvest Biol. Technol.; 2018; 140, pp. 85-92. [DOI: https://dx.doi.org/10.1016/j.postharvbio.2018.02.011]
5. Wang, T.; Guo, Y.; Wan, P.; Sun, X.; Zhang, H.; Yu, Z.; Chen, X. A flexible transparent colorimetric wrist strap sensor. Nanoscale; 2017; 9, pp. 869-874. [DOI: https://dx.doi.org/10.1039/C6NR08265C]
6. Geng, Y.; Ali, M.A.; Clulow, A.J.; Fan, S.; Burn, P.L.; Gentle, I.R.; Meredith, P.; Shaw, P.E. Unambiguous detection of nitrated explosive vapours by fluorescence quenching of dendrimer films. Nat. Commun.; 2015; 6, 8240. [DOI: https://dx.doi.org/10.1038/ncomms9240] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/26370931]
7. Jia, R.; Tian, W.; Bai, H.; Zhang, J.; Wang, S.; Zhang, J. Amine-responsive cellulose-based ratiometric fluorescent materials for real-time and visual detection of shrimp and crab freshness. Nat. Commun.; 2019; 10, 795. [DOI: https://dx.doi.org/10.1038/s41467-019-08675-3]
8. Bruderer, T.; Gaisl, T.; Gaugg, M.T.; Nowak, N.; Streckenbach, B.; Müller, S.; Moeller, A.; Kohler, M.; Zenobi, R. On-line analysis of exhaled breath: Focus review. Chem. Rev.; 2019; 119, pp. 10803-10828. [DOI: https://dx.doi.org/10.1021/acs.chemrev.9b00005]
9. Suslick, K.S. An optoelectronic nose: “seeing” smells by means of colorimetric sensor arrays. MRS Bull.; 2004; 29, pp. 720-725. [DOI: https://dx.doi.org/10.1557/mrs2004.209]
10. Schoolaert, E.; Hoogenboom, R.; De Clerck, K. Colorimetric Nanofibers as Optical Sensors. Adv. Funct. Mater.; 2017; 27, 1702646. [DOI: https://dx.doi.org/10.1002/adfm.201702646]
11. Lewińska, I.; Speichert, M.; Granica, M.; Tymecki, Ł. Colorimetric point-of-care paper-based sensors for urinary creatinine with smartphone readout. Sens. Actuators B; 2021; 340, 129915. [DOI: https://dx.doi.org/10.1016/j.snb.2021.129915]
12. Li, X.; Li, S.; Liu, Q.; Cui, Z.; Chen, Z. A Triple-Channel Colorimetric Sensor Array for Identification of Biothiols Based on Color RGB (Red/Green/Blue) as Signal Readout. ACS Sustain. Chem. Eng.; 2019; 7, pp. 17482-17490. [DOI: https://dx.doi.org/10.1021/acssuschemeng.9b04740]
13. Wang, J. Generalized 2-D Principal Component Analysis by Lp-Norm for Image Analysis. IEEE Trans. Cybern.; 2016; 46, pp. 792-803. [DOI: https://dx.doi.org/10.1109/TCYB.2015.2416274]
14. Bouwmans, T.; Javed, S.; Zhang, H.; Lin, Z.; Otazo, R. On the Applications of Robust PCA in Image and Video Processing. Proc. IEEE; 2018; 106, pp. 1427-1457. [DOI: https://dx.doi.org/10.1109/JPROC.2018.2853589]
15. Song, Q.; Xiong, R.; Fan, X.; Liu, D.; Wu, F.; Huang, T.; Gao, W. Compressed Image Restoration via Artifacts-Free PCA Basis Learning and Adaptive Sparse Modeling. IEEE Trans. Image Process.; 2020; 29, pp. 7399-7413. [DOI: https://dx.doi.org/10.1109/TIP.2020.3002452]
16. Wei, L.; Huafu, C.; Qin, Y.; Xu, L. Analysis of fMRI Data Using Improved Self-Organizing Mapping and Spatio-Temporal Metric Hierarchical Clustering. IEEE Trans. Med. Imaging; 2008; 27, pp. 1472-1483.
17. Sanghoon, L.; Crawford, M.M. Unsupervised multistage image classification using hierarchical clustering with a bayesian similarity measure. IEEE Trans. Image Process.; 2005; 14, pp. 312-320. [DOI: https://dx.doi.org/10.1109/TIP.2004.841195] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/15762328]
18. Zhu, J.; Wang, H.; Guo, F.; Salmén, L.; Yu, Y. Cell wall polymer distribution in bamboo visualized with in situ imaging FTIR. Carbohydr. Polym.; 2021; 274, 118653. [DOI: https://dx.doi.org/10.1016/j.carbpol.2021.118653]
19. Bordbar, M.M.; Tashkhourian, J.; Hemmateenejad, B. Qualitative and quantitative analysis of toxic materials in adulterated fruit pickle samples by a colorimetric sensor array. Sens. Actuators B; 2018; 257, pp. 783-791. [DOI: https://dx.doi.org/10.1016/j.snb.2017.11.010]
20. Chen, C.; Zare, A.; Trinh, H.N.; Omotara, G.O.; Cobb, J.T.; Lagaunne, T.A. Partial Membership Latent Dirichlet Allocation for Soft Image Segmentation. IEEE Trans. Image Process.; 2017; 26, pp. 5590-5602. [DOI: https://dx.doi.org/10.1109/TIP.2017.2736419] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/28792897]
21. Nie, F.; Dong, X.; Hu, Z.; Wang, R.; Li, X. Discriminative Projected Clustering via Unsupervised LDA. IEEE Trans. Neural Networks Learn. Syst.; 2023; 34, pp. 9466-9480. [DOI: https://dx.doi.org/10.1109/TNNLS.2022.3202719] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36121958]
22. Ulfarsson, M.O.; Palsson, F.; Sigurdsson, J.; Sveinsson, J.R. Classification of Big Data with Application to Imaging Genetics. Proc. IEEE; 2016; 104, pp. 2137-2154. [DOI: https://dx.doi.org/10.1109/JPROC.2015.2501814]
23. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature; 2015; 521, pp. 436-444. [DOI: https://dx.doi.org/10.1038/nature14539] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/26017442]
24. Mater, A.C.; Coote, M.L. Deep learning in chemistry. J. Chem. Inf. Model.; 2019; 59, pp. 2545-2559. [DOI: https://dx.doi.org/10.1021/acs.jcim.9b00266] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/31194543]
25. Guo, L.; Wang, T.; Wu, Z.; Wang, J.; Wang, M.; Cui, Z.; Ji, S.; Cai, J.; Xu, C.; Chen, X. Portable Food-Freshness Prediction Platform Based on Colorimetric Barcode Combinatorics and Deep Convolutional Neural Networks. Adv. Mater.; 2020; 32, 2004805. [DOI: https://dx.doi.org/10.1002/adma.202004805] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/33006183]
26. Huang, Z.-J.; Luo, J.-Y.; Zheng, F.-Y.; Li, S.-X.; Liu, F.-J.; Lin, L.-X.; Huang, Y.-J.; Man, S.; Cao, G.-X.; Huang, X.-G. Long-term stable, high accuracy, and visual detection platform for In-field analysis of nitrite in food based on colorimetric test paper and deep convolutional neural networks. Food Chem.; 2021; 373, 131593. [DOI: https://dx.doi.org/10.1016/j.foodchem.2021.131593] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34838401]
27. Ma, P.; Xu, W.; Teng, Z.; Luo, Y.; Gong, C.; Wang, Q. An integrated food freshness sensor array system augmented by a metal–organic framework mixed-matrix membrane and deep learning. ACS Sens.; 2022; 7, pp. 1847-1854. [DOI: https://dx.doi.org/10.1021/acssensors.2c00255]
28. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Honolulu, HI, USA, 21–26 July 2017; pp. 4700-4708.
29.
30.
31. Niimi, J.; Deveau, A.; Splivallo, R. Geographical-based variations in white truffle Tuber magnatum aroma is explained by quantitative differences in key volatile compounds. New Phytol.; 2021; 230, pp. 1623-1638. [DOI: https://dx.doi.org/10.1111/nph.17259]
32. Maamoun, A.A.; El-akkad, R.H.; Farag, M.A. Mapping metabolome changes in Luffa aegyptiaca Mill fruits at different maturation stages via MS-based metabolomics and chemometrics. J. Adv. Res.; 2019; 29, pp. 179-189. [DOI: https://dx.doi.org/10.1016/j.jare.2019.10.009]
33. Fayek, N.M.; Farag, M.A.; Saber, F.R. Metabolome classification via GC/MS and UHPLC/MS of olive fruit varieties grown in Egypt reveal pickling process impact on their composition. Food Chem.; 2021; 339, 127861. [DOI: https://dx.doi.org/10.1016/j.foodchem.2020.127861] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32836025]
34. Riuaumatell, M.; Castellari, M.; Lopeztamames, E.; Galassi, S.; Buxaderas, S. Characterisation of volatile compounds of fruit juices and nectars by HS/SPME and GC/MS. Food Chem.; 2004; 87, pp. 627-637. [DOI: https://dx.doi.org/10.1016/j.foodchem.2003.12.033]
35. Wang, X.; Wang, Y.; Qi, H.; Chen, Y.; Guo, W.; Yu, H.; Chen, H.; Ying, Y. Humidity-Independent Artificial Olfactory Array Enabled by Hydrophobic Core–Shell Dye/MOFs@COFs Composites for Plant Disease Diagnosis. ACS Nano; 2022; 16, pp. 14297-14307. [DOI: https://dx.doi.org/10.1021/acsnano.2c04457] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/36043472]
36. Lim, S.H.; Kemling, J.W.; Feng, L.; Suslick, K.S. A colorimetric sensor array of porous pigments. Analyst; 2009; 134, pp. 2453-2457. [DOI: https://dx.doi.org/10.1039/b916571a] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/19918616]
37. Sakare, P.; Giri, S.K.; Mohapatra, D.; Modhera, B.; Babu, V.B. Lac dye-based intelligent colorimetric indicator for real-time freshness monitoring of packaged white button mushrooms (Agaricus bisporus). Postharvest Biol. Technol.; 2023; 206, 112552. [DOI: https://dx.doi.org/10.1016/j.postharvbio.2023.112552]
38. Li, Z.; Suslick, K.S. Colorimetric sensor array for monitoring CO and ethylene. Anal. Chem.; 2018; 91, pp. 797-802. [DOI: https://dx.doi.org/10.1021/acs.analchem.8b04321] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30547588]
39. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Boston, MA, USA, 7–12 June 2015; pp. 1-9.
40. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Las Vegas, NV, USA, 27–30 June 2016; pp. 2818-2826.
41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Las Vegas, NV, USA, 27–30 June 2016; pp. 770-778.
42. Andrade, E.H.A.; Maia, J.G.S.; Maria das Graças, B.Z. Aroma volatile constituents of Brazilian varieties of mango fruit. J. Food Compos. Anal.; 2000; 13, pp. 27-33. [DOI: https://dx.doi.org/10.1006/jfca.1999.0841]
43. Ghatak, B.; Ali, S.B.; Naskar, H.; Tudu, B.; Pramanik, P.; Mukherji, S.; Bandyopadhyay, R. Selective and sensitive detection of limonene in mango using molecularly imprinted polymer based quartz crystal microbalance sensor. Proceedings of the 2019 IEEE International Symposium on Olfaction and Electronic Nose (ISOEN); Fukuoka, Japan, 26–29 May 2019; pp. 1-3.
44. Lan, T.; Bao, S.; Wang, J.; Ge, Q.; Zhang, H.; Yang, W.; Sun, X.; Ma, T. Shelf life of non-industrial fresh mango juice: Microbial safety, nutritional and sensory characteristics. Food Biosci.; 2021; 42, 101060. [DOI: https://dx.doi.org/10.1016/j.fbio.2021.101060]
45. Do, J.; Salunkhe, D.; Olson, L. Isolation, identification and comparison of the volatiles of peach fruit as related to harvest maturity and artificial ripening. J. Food Sci.; 1969; 34, pp. 618-621. [DOI: https://dx.doi.org/10.1111/j.1365-2621.1969.tb12104.x]
46. Sumitani, H.; Suekane, S.; Nakatani, A.; Tatsuka, K. Changes in composition of volatile compounds in high pressure treated peach. J. Agric. Food. Chem.; 1994; 42, pp. 785-790. [DOI: https://dx.doi.org/10.1021/jf00039a037]
47. Visai, C.; Vanoli, M. Volatile compound production during growth and ripening of peaches and nectarines. Sci. Hortic.; 1997; 70, pp. 15-24. [DOI: https://dx.doi.org/10.1016/S0304-4238(97)00032-0]
48. Zhang, J.; Jha, S.K.; Liu, C.; Hayashi, K. Tracing of Chemical Components of Odor in Peels and Flesh from Ripe Banana on a Daily Basis Using GC-MS Characterization and Statistical Analysis for Quality Monitoring During Storage. Food Anal. Methods; 2019; 12, pp. 947-955. [DOI: https://dx.doi.org/10.1007/s12161-019-01435-5]
49. Zhu, X.; Li, Q.; Li, J.; Luo, J.; Chen, W.; Li, X. Comparative Study of Volatile Compounds in the Fruit of Two Banana Cultivars at Different Ripening Stages. Molecules; 2018; 23, 2456. [DOI: https://dx.doi.org/10.3390/molecules23102456] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/30257494]
50. Koh, A.; Kang, D.; Xue, Y.; Lee, S.; Pielak, R.M.; Kim, J.; Hwang, T.; Min, S.; Banks, A.; Bastien, P. et al. A soft, wearable microfluidic device for the capture, storage, and colorimetric sensing of sweat. Sci. Transl. Med.; 2016; 8, 366ra165. [DOI: https://dx.doi.org/10.1126/scitranslmed.aaf2593] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/27881826]
51. Shi, S.; Wang, Z.; Shi, J.; Wang, X.; Li, H. From points to parts: 3d object detection from point cloud with part-aware and part-aggregation network. IEEE Trans. Pattern Anal. Mach. Intell.; 2020; 43, pp. 2647-2664. [DOI: https://dx.doi.org/10.1109/TPAMI.2020.2977026] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/32142423]
52. Liu, Z.; Tang, H.; Amini, A.; Yang, X.; Mao, H.; Rus, D.L.; Han, S. Bevfusion: Multi-task multi-sensor fusion with unified bird’s-eye view representation. Proceedings of the 2023 IEEE International Conference on Robotics and Automation (ICRA); London, UK, 29 May–2 June 2023; pp. 2774-2781.
53. Shi, S.; Wang, X.; Li, H. Pointrcnn: 3d object proposal generation and detection from point cloud. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Long Beach, CA, USA, 15–20 June 2019; pp. 770-779.
54. He, C.; Zeng, H.; Huang, J.; Hua, X.-S.; Zhang, L. Structure aware single-stage 3d object detection from point cloud. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Seattle, WA, USA, 13–19 June 2020; pp. 11873-11882.
55. Shi, S.; Guo, C.; Jiang, L.; Wang, Z.; Shi, J.; Wang, X.; Li, H. Pv-rcnn: Point-voxel feature set abstraction for 3d object detection. Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition; Seattle, WA, USA, 13–19 June 2020; pp. 10529-10538.
56. Shin, K.; Kwon, Y.P.; Tomizuka, M. Roarnet: A robust 3d object detection based on region approximation refinement. Proceedings of the 2019 IEEE intelligent vehicles symposium (IV); Paris, France, 9–12 June 2019; pp. 2510-2515.
57. Mnih, V.; Heess, N.; Graves, A. Recurrent models of visual attention. Adv. Neural Inf. Process. Syst.; 2014; 27, pp. 2204-2212.
58. Othman, M.; Bakar, M.N.A.; Ahmad, K.A.; Razak, T.R. Fuzzy ripening mango index using RGB colour sensor model. Res. World; 2014; 5, 1.
59. Abdullah, N.E.; Madzhi, N.K.; Yahya, A.M.A.A.; Rahim, A.A.A.; Rosli, A.D. ANN Diagnostic System for Various Grades of Yellow Flesh Watermelon based on the Visible light and NIR properties. Proceedings of the 2018 4th International Conference on Electrical, Electronics and System Engineering (ICEESE); Kuala Lumpur, Malaysia, 8–9 November 2018; pp. 70-75.
60. Choe, U.; Kang, H.; Ham, J.; Ri, K.; Choe, U. Maturity assessment of watermelon by acoustic method. Sci. Hortic.; 2022; 293, 110735. [DOI: https://dx.doi.org/10.1016/j.scienta.2021.110735]
61. Brezmes, J.; Llobet, E.; Vilanova, X.; Saiz, G.; Correig, X. Fruit ripeness monitoring using an electronic nose. Sens. Actuators B; 2000; 69, pp. 223-229. [DOI: https://dx.doi.org/10.1016/S0925-4005(00)00494-9]
62. Li, H.; Lee, W.S.; Wang, K. Identifying blueberry fruit of different growth stages using natural outdoor color images. Comput. Electron. Agric.; 2014; 106, pp. 91-101. [DOI: https://dx.doi.org/10.1016/j.compag.2014.05.015]
63. Chhabra, M.; Gupta, A.; Mehrotra, P.; Reel, S. Automated detection of fully and partially riped mango by machine vision. Proceedings of the International Conference on Soft Computing for Problem Solving (SocProS 2011); Roorkee, India, 20–22 December 2011; Volume 2, pp. 153-164.
64. Soltani, M.; Alimardani, R.; Omid, M. Evaluating banana ripening status from measuring dielectric properties. J. Food Eng.; 2011; 105, pp. 625-631. [DOI: https://dx.doi.org/10.1016/j.jfoodeng.2011.03.032]
65. Nandi, C.; Tudu, B.; Koley, C. Machine vision based techniques for automatic mango fruit sorting and grading based on maturity level and size. Sensing Technology: Current Status and Future Trends II; Springer International Publishing: Cham, Switzerland, 2014; pp. 27-46.
66. Ibba, P.; Tronstad, C.; Moscetti, R.; Mimmo, T.; Cantarella, G.; Petti, L.; Martinsen, Ø.G.; Cesco, S.; Lugli, P. Supervised binary classification methods for strawberry ripeness discrimination from bioimpedance data. Sci. Rep.; 2021; 11, 11202. [DOI: https://dx.doi.org/10.1038/s41598-021-90471-5] [PubMed: https://www.ncbi.nlm.nih.gov/pubmed/34045542]
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
Artificial scent screening systems, inspired by the mammalian olfactory system, hold promise for fruit ripeness detection, but their commercialization is limited by low sensitivity or pattern recognition inaccuracy. This study presents a portable fruit ripeness prediction system based on colorimetric sensing combinatorics and deep convolutional neural networks (DCNN) to accurately identify fruit ripeness. Using the gas chromatography–mass spectrometry (GC-MS) method, the study discerned the distinctive gases emitted by mango, peach, and banana across various ripening stages. The colorimetric sensing combinatorics utilized 25 dyes sensitive to fruit volatile gases, generating a distinct scent fingerprint through cross-reactivity to diverse concentrations and varieties of gases. The unique scent fingerprints can be identified using DCNN. After capturing colorimetric sensor image data, the densely connected convolutional network (DenseNet) was employed, achieving an impressive accuracy rate of 97.39% on the validation set and 82.20% on the test set in assessing fruit ripeness. This fruit ripeness prediction system, coupled with a DCNN, successfully addresses the issues of complex pattern recognition and low identification accuracy. Overall, this innovative tool exhibits high accuracy, non-destructiveness, practical applicability, convenience, and low cost, making it worth considering and developing for fruit ripeness detection.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details

1 School of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China; Key Laboratory of Intelligent Equipment and Robotics for Agriculture of Zhejiang Province, Hangzhou 310058, China
2 School of Biosystems Engineering and Food Science, Zhejiang University, Hangzhou 310058, China; Key Laboratory of Intelligent Equipment and Robotics for Agriculture of Zhejiang Province, Hangzhou 310058, China; ZJU-Hangzhou Global Scientific and Technological Innovation Center, Hangzhou 311200, China