1. Introduction
The rapid growth of global population compounded by climate change is putting enormous pressure on the agricultural sector to increase the quality and quantity of food production. It is predicted that the global population will reach nine billion by 2050, and therefore, agricultural production must double to meet the increasing demands [1]. However, agriculture is facing immense challenges from the growing threats of plant diseases, pests and weed infestation [2,3,4,5] . The weed infestations, pests and diseases reduce the yield and quality of food, fibre and biofuel value of crops. Losses are sometimes cataclysmic or chronic, but on average account for about of the production of a few important food crops [1]. Weeds are undesired plants which compete against productive crops for space, light, water and soil nutrients, and propagate themselves either through seeding or rhizomes. They are generally poisonous, produce thorns and burrs and hamper crop management by contaminating crop harvests. That is why farmers spend billions of dollars on weed management, often without adequate technical support, resulting in poor weed control and reduced crop yield. Hence, weed control is an important aspect of horticultural crop management, as failure to adequately control weeds leads to reduced yields and product quality [6]. The use of chemical and cultural control strategies can lead to adverse environmental impacts when not managed carefully. A low-cost tool for identification and mapping of weeds at early growth stages will contribute to more effective, sustainable weed management approaches. Along with preventing the loss of crop yield by up to , early weed control is also useful in reducing the occurrence of diseases and pests in crops [2,7]. Many approaches have been developed for managing weeds, and they normally consider current environmental factors. Among these approaches, image processing is promising. In the image processing approach, unmanned aerial vehicles (UAVs) are used for monitoring crops and capturing the possible weeds in the fields. UAVs are found to be beneficial for agriculture usage due to their ability to cover large areas in a very short amount of time, and at the same time, do not cause any soil compaction or damage in the fields [8]. However, the interpretation of data collected from UAVs into meaningful information is still a painstaking task. This is because a standard data collection and classification require significant amounts of manual effort for segment size tuning, feature selection and rule-based classifier design.
The underline motivation of this paper was a real scenario. This scenario involved a chilli producing farm (AustChilli Group Australia) which is located in Bundaberg, Australia. As found in many studies [3,9,10,11] in the literature, the quality and quantity of the crop yields were being hampered due to the water and nutrients consumed by weed. Therefore, an effective solution for weed control in early stages of the crops is much needed. Hence, the objective of this study was to find an efficient and robust machine learning-based approach for detecting unwanted weed and parasites within crops. For this purpose, we analysed three machine learning algorithms, namely, random forest (RF), k-nearest neighbours (KNN) and support vector machine (SVM), and compared their performances at detecting weeds. In our previous work [12], we presented a weed detection mechanism using the RF algorithm. This study was an extended version, where we detected weeds using SVM and KNN and compared the results with the performance of RF. It is noteworthy to mention the condition of the farmland, which is one of the very crucial parameters for selecting the appropriate weed detection algorithm or detection mechanism. Generally, the field background for weed detection could be classified as one of two groups. The first one is a field early in the season, when weeds can be easily differentiated from the background (soil, plastic mulch, etc.), and crops (in seedling stage) can be identified as a simple spatial pattern (e.g., rows, or grid pattern) [13]. The second category is known as the closed crop canopy in agriculture, which has a fairly complicated background. In this case, weeds are detected within mature crops; this requires finding out unique features that differentiate weeds and crops (spectral signature, shape, etc.). This can be quite challenging and sometimes requires laborious annotation/ground truthing [14]. For our paper, we have done the weed detection at early stage of crop growth. The reason for this was that the seedlings are more hampered early on by weeds, since weeds consume the nutrients (water and fertiliser) which are much needed for the seedlings to grow [15]. Therefore, it is very important to detect and remove weeds in the early stages.
This paper is organised into five sections. Section 2 provides a literature review on machine learning algorithms used in weed detection. Section 3 presents the materials and experimental methods used for weed detection in an Australian chilli field. This approach includes several stages: UAV image collection, image pre-processing, feature extraction and selection, labelling images and applying machine learning classifiers. The simulation method and parameters are also presented in Section 3. Section 4 presents the simulation results of the classifiers along with performance analysis of RF, KNN and SVM. Finally, Section 5 concludes the paper with some potential directions for our future work.
2. Literature Review on Machine Learning Algorithms for Weed Detection
In this section, we explore relevant works conducted in recent years that use machine learning and image analysis for weed detection. Recent studies in the literature present a variety of classification approaches used to generate weed maps from UAV images [16,17,18,19]. However, as evidence in the recent state-of-art works shows [3,5,13,20], machine learning algorithms make more accurate and efficient alternatives than conventional parametric algorithms, when dealing with complex data. Among these machine learning algorithms, the RF classifier is becoming a very popular option for remote sensing applications due to its generalised performance and operational speed [3,4,5,13,21]. RF has been found to be desirable for high resolution UAV image classification and agricultural mapping [3,4,5,13,21]. Another popular machine learning classifier is SVM, which has been popularly used for weed and crop classification [22,23,24,25]. On the other hand, Kazmi et al. [26] used the KNN algorithm for creeping thistle detection in sugar beet fields. Table 1 presents an overview of recent works on machine learning based approach for weed detection.
Brinkhoff et al. [22] generated a land cover map to locate and classify perennial crops over a 6200 km area of the Riverina region located in New South Wales, Australia. To optimise accuracy, they used object-based image analysis techniques along with supervised SVM classification. They found the overall accuracy to be for an object count with twelve classes, whereas the accuracy was when weighted by object area. These results demonstrated the potential of using a time series of medium resolution remote sensing images to produce detailed land cover maps over intensive perennial cropping areas. Alam et al. [3] used an RF classifier to develop a real-time computer vision-based system to distinguish weeds from crops. The authors created their own dataset to train the classification model, and then they tested it with field data. Moreover, they developed a pulse width modulation-based fluid flow control system which controls an equipment to spray a desired amount of agrochemical based on the feedback from the vision-based system. Hence, the authors demonstrated the effectiveness of their real time vision-based agrochemical spraying system.
Faisal et al. [28] proposed a SVM algorithm to identify the weeds from chilli field images. The goal of their research was to explore the performance of the SVM classifier in an integrated weed management scheme. They collected images from chilli fields of Bangladesh that included five different weeds. Those images were then segmented by using a global thresholding-based binarisation technique to distinguish the plants from the ground, and extract features. From each image, fourteen features were classified into three groups—colour feature, shape features and moment invariants. Finally, SVM classifier was used to detect weeds. Their experimental results showed that the SVM achieved overall 97% accuracy on 224 test images. In [24], the authors proposed a weed detection approach for sugar beet cultivation where they combined several shape features to set up patterns which then were used to separate the highly similar sugar beets from the weeds. The images used in the research were collected from Shiraz University sugar beet fields. MATLAB toolbox was used to process these images. To differentiate between weeds and sugar beets, the authors explored different shape features—namely: shape factors, moment invariants and Fourier descriptors. Then, KNN and SVM classifiers were used; the overall accuracies were found to be and , respectively.
In [25], a weed detection approach was introduced where a histogram based on colour indices is used to identify soil, soybean and weed (broadleaf) classes by using colour indices as a feature, and a monochrome image is generated. To obtain grey-scale images, the image was scaled to an interval of 0–255, and image histograms were produced and normalised; BPNN and SVM classifiers were used to train with these histograms. The aim of this research was to determine an alternative feature vector that includes simple computational steps, and at the same time, ensure a high weed detection rate. The overall accuracies achieved by this technique for BPNN and SVM were and , respectively. In [14], the authors proposed an automated weed detection framework that was able to detect weeds at the various phases of plant growth. Colour, multispectral and thermal images were captured by using UAV-based sensors. Next, they converted the collected images into NDVI images and applied image processing techniques to those images, and by using colour images as the ground truth, they manually drew bounding boxes and hand-labelled vegetation bulbs. Finally, they applied the machine learning approaches to separate the weeds.
Gao et al. [5] investigated the effectiveness of a hyperspectral snapshot mosaic camera for the identification of weed and maize, where the images were collected from a laboratory of plant in Belgium. These raw images were processed, and for the band features, the calibration formula reflectance was acquired. These reflectance, NDVI and RVI features were generated in VB and NIR regions and one hundred eighty-five features were found. Then, a PCA-based feature reduction method was used. Next, feature selection algorithms were applied to extract distinctive features. Finally, a RF classifier was used to detect weed and crop. Overall, 81% accuracy was achieved for different types of weed recognition.
Castro et al. [13] proposed an automatic, RF-based image analysis approach to identify weeds from the early growing phases of the herbaceous crops. In this approach, the UAV images were used in the combination of digital surface models (DSMs) and orthomosaic techniques. After that, RF classifier was used to separate the weeds from the crops and soil, and the approach achieved 87.9% and 84% accuracies for sunflower and cotton fields. In [21], a simple flow diagram was proposed that can be used to track invasive water troops, both emerging and submerged; the important components of this workflow are the gathering of radiometrically calibrated multispectral imagery, image segmentation and using a machine learning model. The images were collected by using the eBee mapping drone. The orthomosaic of the multispectral imagery was generated by using Pix4Dmapper Pro 3.0. The images were then mosaiced and made into absolute maps of reflectivity, and they used ENVI 5.4 for image segmentation and extracted 14 spatial features. Moreover, they also generated an NDVI band and integrated it into the segmentation. Finally, they used an RF classifier for detection of infestations among the plants. The overall accuracies achieved for the above-water and submerged features were 92.19% and 83.76%, respectively.
In [4], the authors introduced an approach that can accurately estimate the different growth stages of avocado trees by using UAVs. They used a Parrot Sequoia® multi-spectral camera to capture multi-spectral images. To calculate the tree height, they used a canopy height model by separating the digital terrain model from the digital surface model. Then they utilised red edge and NIR bands brightness and different vegetation indices based on orthomosaic at-surface reflectance imagery. Finally, an RF algorithm was used which achieved 96% accuracy.
In [27], a weed mapping approach was proposed for precision agriculture, where UAV captured images were used from sunflower and maize crop fields. The research addressed the early growth stage problem of spectral similarity for crops and weed pixels with object-based image analysis (OBIA), for which they used an SVM algorithm integrated with feature selection techniques. The UAV was used to collect images of sunflower and maize fields located at the private farms, namely, La Monclova and El Mazorcal, in Spain. Next, these images were mosaicked using the Agisoft Photoscan software, and the subsample was labelled based on unsupervised feature selection techniques from these objects, whereas the automatic labelling was performed under the supervision of an expert. These objects were represented as colour histograms and data features based on remote-sensing measurements (first-order statistics, textures, etc.) for the classification of the objects. The overall accuracy of this SVM based approach was found to be around .
3. Materials and Experimental Methods
This section presents the materials and methodologies used in the proposed weed detection method. Figure 1 shows the workflow diagram showing different stages between the initial stage of inputting RGB images and the final stage of detecting weed. The materials and experimental methods of different stages are presented in the following subsections.
3.1. UAV Image Collection
As mentioned in the previous section, recent research publications in the field of agriculture have demonstrated successful applications of several imaging techniques for identifying diseases and weeds, and monitoring host plants. Sensors and camera-mounted UAVs can be used to capture images of field crops for the purpose. Depending on the type of camera, different types of images can be captured, such as RGB, thermal, multi-spectral, hyper-spectral, 3D and chlorophyll florescence. For our work, we have used RGB images captured by RGB cameras mounted in a drone (i.e., UAV). These images were captured over the chilli farm using a Phantom 3 Advanced drone mounted camera with the 1/2.3 CMOS sensor (the specifications of the drone and camera can be found here:
3.2. Image Preprocessing
The RGB UAV images were pre-processed and converted to orthomosaic images. In the image pre-processing step, the images were calibrated and mosaicked using Pix4D Mapping software following the default settings of “3D Maps”. Then they were geo-registered into a single orthophoto. The aerial images were geo-registered using ground control points [30] which were set with coordinates measured by a differential GPS with 0.02 m error (Leica Viva GS14 GNSS antenna and CS20 controller, Leica Geosystems, St. Gallen, Switzerland) and then converted to orthophotos. An orthophoto, also known as an orthophotograph or orthoimage, is an aerial photograph which is geometrically corrected to make the scale uniform so that the photo has the same lack of distortion as a map. An orthophoto consists of adjusted camera tilts, lens distortion and topographic relief. Therefore, it represents the earth’s surface accurately, and consequently, it can be used to measure true distance.
3.3. The Extraction and Selection of Features from Images
From a RGB image of a small part of the chilli farm, we extracted the reflectance of red, green and blue bands, from which we extracted the vegetation indices such as normalised red band, normalised green band and normalised blue band. The purpose of this normalisation was to reduce the effects of different lighting conditions on the colour channels. After that, excess green (), excess red (), excess green and red () and the greenness index were calculated, as shown in Table 2, where represents the reflectance of a specific band. The index gives the difference between the detected light values of green channel and the red and blue channels. For our machine learning approach, we used r, g, b, GI and indices, since they provide the optimum results of weed detection.
3.4. Labelling the Images
We need a labelled image to train our RF, KNN and SVM algorithms. Here, we used ENVI classic software to label a small section of the image as weeds, crops and bare field so that we could use it as the training data for the three machine learning algorithms. In ENVI, we manually labelled some of the areas as weeds, crops and bare field using ROIs (Region of interest) and then used the neural network classifier to label rest of area in the image. Figure 2 shows a labelled image where the crops are labelled in green, weeds are labelled in red and bare land is labelled in white. This labelled image was used to train our RF, KNN and SVM algorithms and detect weeds.
3.5. Machine Learning-Based Classification
Once image features were extracted and selected, these were used to build classification models. We have used the following machine learning classifiers as they have been efficiently used in the literature:
Random forest (RF) classifier: Breiman et al. [32] defined the RF as, “An ensemble of classification trees, where each decision tree employs a subset of training samples and variables selected by a bagging approach, while the remaining samples are used for internal cross-validation of the RF performance. The classifier chooses the membership classes having the most votes, for a given case.” RF has been proven to be highly suitable for high resolution UAV image classification and for agricultural mapping [3,4,5,13,21].
Support vector machine (SVM) classifier: SVM classifies data points based on hyperplanes which optimally differentiate the classes based on training data [24,25]. These hyperplanes are the surfaces defined by combinations of input features. SVM has been popularly used in literature to perform weed and crop classification [22,23,24,25].
K-nearest neighbours (KNN) classifier: KNN is a non-parametric algorithm, popularly used for regression and classification. The input consists of the k closest training examples in the feature space [33,34]. Kazmi et al. [26] used the KNN algorithm for creeping thistle detection in sugar beet fields.
As mentioned earlier, RF, SVM and KNN have been widely used for weed and crop classification; hence we selected and used these algorithms in our study, and compared the results of these classifiers. We have classified three different classes, namely, weeds, crops and unwanted area (bare land and plastic mulch) in the chilli farm in Australia.
3.6. Simulation Method and Parameters
MATLAB has been used for feature extraction from the pre-processed images and to simulate the machine learning based algorithms. For RF, we used 500 decision trees to generate the random forest algorithm, and we used the value of k = 4 for KNN. For any machine learning technique, we need to use a training data to train the algorithm. In this study, we used the labelled image shown in Figure 2 from ENVI to train the machine learning algorithm. We used this training data in our RF, KNN and SVM classifiers to detect weed in the chilli field. Additionally, we compare some performance metrics of RF, KNN and SVM classifiers, such as accuracy, recall (also known as sensitivity), specificity, precision, false positive rate and kappa coefficient [27,28]. We define the following parameters to be used in the performance metrics:
= True positive = number of records when weed is detected correctly.
= True negative = number of records when crop and bare land is detected correctly.
= False positive = number of records when weed is detected incorrectly.
= False Negative = number of records when crop and bare land is detected incorrectly.
P= Total positive = .
N= Total negative = .
Based on these formulas, we calculated the following performance metrics for RF classifier and compared them with KNN and SVM classifiers. The classification accuracy, recall, specificity, precision, false positive rate and kappa coefficient (kc) [35] were calculated from the following formulae [24,36,37]:
(1)
(2)
(3)
(4)
(5)
(6)
where A is accuracy andThe following section presents the simulation results of machine learning-based image processing for weed detection and compares the above-mentioned performance metrics of the machine learning techniques.
4. Simulation Results
Figure 3 shows the original RGB image of a small part of the farm; and the simulation results of identifying weeds in and within the rows of chilli using RF, KNN and SVM are shown in Figure 4. We have classified three different classes: weeds, crops and unwanted area (bare land and plastic mulch). We only show the weeds class in Figure 4.
We compared and contrasted the weed map with the original image of a small part of the chilli field, as shown in Figure 3. From visual analysis, we can see that the weed map was created properly for RF and SVM. However, we can see some false detection in the KNN map. Once the weed map was created, we used the QGIS software to find the coordinates of that weeded areas. Hence, the farm owner could identify the weeded area and apply manpower to remove the weeds. We compare the performance of RF, KNN and SVM in the following subsection.
Performance Analysis of RF, KNN and SVM
The performance metrics of RF, KNN and SVM were calculated using Equations (1)–(6) and are presented in Table 3.
These matrices are also presented in a bar chart in Figure 5, where we show the performance measures in terms of accuracy, recall, specificity, precision, false positive rate (FPR) and kappa coefficient. The accuracies for RF, KNN and SVM on the test dataset were 0.963, 0.628 and 0.94; RF shows better results than KNN and SVM. The recall and specificity for RF, KNN and SVM were 0.951 and 0.887; 0.621 and 0.819; and 0.911 and 0.890, respectively. The precision, FPR and kappa coefficient for RF, KNN and SVM were 0.949, 0.057 and 0.878; 0.624, 0.180 and 0.364; and 0.908, 0.08 and 0.825, respectively. Similarly to the accuracy metric, RF showed better results than the others in all other metrics. SVM also performed well. However, RF was more accurate than SVM. On the other hand, KNN did not perform well on our dataset. Therefore, it is evident that both RF and SVM are efficient classifiers for weed detection from UAV images.
5. Conclusions
Early weed detection is crucial in agricultural productivity, as weeds act as a pest to crops. This work aimed to detect weeds in a chilli field using image processing and machine learning techniques. The UAV images were collected from an Australian chilli farm, and these images were pre-processed using image processing techniques. Then features were extracted from the images to distinguish properties of weeds and the crop. Three different classifiers were tested using those properties: RF, SVM, and KNN. The experimental results demonstrate that RF performed better than the other classifiers in terms of accuracy and other performance metrics. RF and SVM offered and accuracy in weed detection from RGB images, respectively, whereas KNN offered only accuracy. In the future, we will explore multispectral and hyperspectral UAV images, and will apply deep learning algorithms to increase the accuracy of weed detection.
Author Contributions
Conceptualisation, N.I., M.M.R., S.W., C.-Y.X. and A.M.; methodology, N.I., M.M.R. and S.W.; software, N.I., M.M.R. and S.M.R.; validation, N.I., M.M.R. and S.W.; resources, C.-Y.X.; writing—original draft preparation, N.I., M.M.R., S.W. and S.A.W.; writing—review and editing N.I., M.M.R., S.W., A.M. and C.-Y.X., supervision S.M.; funding acquisition N.I., M.M.R. and S.A.W. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by New Staff Research Grant-2019, CQUniversity, grant number RSH/5339.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Acknowledgments
We would like to thank AustChilli Group, Australia, for allowing us to collect the images from their farm. This research was partially funded by research project grant-RSH/5339, funded by Central Queensland University, Australia.
Conflicts of Interest
The authors declare no conflict of interest.
Abbreviations
The following abbreviations are used in this manuscript:
FPR | False positive rate |
KNN | K-nearest neighbour |
ML | Machine learning |
NDVI | Normalised difference vegetation index |
OBIA | Object based image analysis |
RF | Random forest |
SVM | Support vector machine |
UAV | Unmanned aerial vehicle |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Figures and Tables
Figure 2. Labelled image to train the ML algorithms. (Crops are labelled in green, weeds are labelled in red and unwanted area is labelled as white).
Figure 4. Weed detection using three different machine learning algorithm (RF, KNN, SVM).
Machine learning techniques for weed detection.
Author | Year | Problem Definition | Targeted Crop | Dataset | Model/Tools | Accuracy |
---|---|---|---|---|---|---|
Alam et al. [3] | 2020 | Crop/weed detection and classification | Unspecified | Images were collected from private farm | RF | 95% |
Brinkhoff et al. [22] | 2020 | Land cover mapping | 9 perennial crops | Images were taken from the Riverina region in NSW, Australia | SVM | 84.8% |
Aaron et al. [14] | 2019 | Weed detection by UAV | Maize | Images were collected from private farm | NVDI digit, YOLOv3 Detector | 98% |
Zhang et al. [23] | 2019 | Weeds species recognition | 8 weed plants | 1600 weed images were taken from South China crop field | SVM | 92.35% |
Y-H Tu et al. [4] | 2019 | Measuring Canopy structure | Avocado tree | Avocado field Bundaberg, Australia | RF | 96% |
Adel et al. [24] | 2018 | Weed detection using shape feature | Sugar beet | Images were taken Shiraz University, Iran | SVM | 95% |
Abouzahir et al. [25] | 2018 | Weeds species detection | Soybean | Images were collected from Sâa José farm, Brazil | SVM | 95.07% |
J Gao et al. [5] | 2018 | Weeds recognition | Maize | Images were taken from crop field of Belgium | RF, KNN | 81% 76.95% |
Castro et al [13] | 2018 | Early Weed mapping | Sunflower, cotton | Images were taken from crop field of Spain | RF | 87.9% |
D Chabot et al. [21] | 2018 | Monitoring water aquatic vegetation | Stratiotes aloides | Trent-Severn Waterway in Ontario, Canada | RF | 92.19% |
Maria et al. [27] | 2016 | weed mapping using UAV-imagery | Sunflower, maize | Images were collected from private farm | SVM | 95.5% |
Faisal et al. [28] | 2012 | Classification of crops and weed | Chilli | Images were Collected chilli field | SVM | 97% |
Formulae of vegetation indices which were used in our work (adapted from [4,29,31]).
Vegetation Index | Formula |
---|---|
Normalised red band | |
Normalised green band | |
Normalised blue band | |
Greenness Index | |
Excess green | |
Excess red | |
Excess green and red |
Comparison of performance metrics for the RF classifier, KNN classifier and SVM classifier.
Performance Metric | RF | KNN | SVM |
---|---|---|---|
Accuracy | 0.96 | 0.63 | 0.94 |
Recall/Sensitivity | 0.95 | 0.62 | 0.91 |
Specificity | 0.89 | 0.81 | 0.89 |
Precision | 0.95 | 0.62 | 0.91 |
False Positive Rate | 0.06 | 0.18 | 0.08 |
Kappa coefficient | 0.88 | 0.36 | 0.83 |
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
This paper explores the potential of machine learning algorithms for weed and crop classification from UAV images. The identification of weeds in crops is a challenging task that has been addressed through orthomosaicing of images, feature extraction and labelling of images to train machine learning algorithms. In this paper, the performances of several machine learning algorithms, random forest (RF), support vector machine (SVM) and k-nearest neighbours (KNN), are analysed to detect weeds using UAV images collected from a chilli crop field located in Australia. The evaluation metrics used in the comparison of performance were accuracy, precision, recall, false positive rate and kappa coefficient. MATLAB is used for simulating the machine learning algorithms; and the achieved weed detection accuracies are 96% using RF,
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
Details




1 School of Engineering and Technology, Central Queensland University, Rockhampton, QLD 4700, Australia;
2 School of Engineering and Technology, Central Queensland University, Rockhampton, QLD 4700, Australia;
3 Institute for Future Farming Systems, Central Queensland University, Bundaberg, QLD 4670, Australia;
4 School of Engineering and Technology, Central Queensland University, Rockhampton, QLD 4700, Australia;
5 School of Engineering and Technology, Central Queensland University, Rockhampton, QLD 4700, Australia;