1. Introduction
The human impact on the marine ecosystem has increased in recent decades [1]. Activities that have a major impact include oil drilling, fishing, and wind turbine deployment. An important factor in monitoring marine biodiversity and maintaining sustainable fish stocks is marine imaging [2,3]. Possible applications are, for example, creating a time series of different species by detecting and classifying species in the images. However, computational support is needed to make the best use of the vast amounts of data generated, e.g., by stationary underwater observatories [4] or seafloor observation systems [5]. A lot of work exists in the context of (semi-)automated detection and classification of species in marine images [3,6,7,8,9,10,11,12]. All these works employ some kind of machine learning algorithm to render a data-driven model of the task to be performed (like object detection or classification). Such a machine learning approach towards (semi-)automatic image interpretation requires a training set of images (or image patches) and expert annotations, usually represented as (taxonomic) labels associated with the images collected with some image annotation software such as BIIGLE 2.0 [13]. In almost all works published, these images are in fact image patches, marked by domain experts in large images showing an underwater scenery containing multiple objects. The detection and extraction of these patches showing single objects can be done by experts, sometimes supported by computational methods that often employ unsupervised learning [14,15,16] or even citizen scientists [17]. However, one task that cannot be supported straightforwardly with computational methods or non-experts is the final classification of objects to taxonomic categories or morphotypes, and this task is addressed in this work.
One main problem in providing computational support for taxonomic classification by employing supervised learning classifiers is the amount of manual expert work required to collect a training set of labeled image patches of sufficient size for all object classes. To collect such a set, several challenges must be faced, some of them very special for underwater computer vision applications:
- Limited background knowledge: Often, it is not known a priori which species can occur in the data set.
- Expensive expert annotation: Quality controlled annotations are expensive in the context of marine imaging because expert knowledge from the domain of marine biology is needed for the annotation process.
- Low abundant classes: It is time-consuming to manually find a sufficient number of examples of rare species for the training set.
One approach that is particularly well suited to tackling these three challenges is active learning. The core idea of active learning is to select training samples automatically from the set of samples to be classified instead of leaving the selection of training samples to human experts. This automated selection of training samples is usually done in an iterative fashion. First, a training sample is selected automatically from the set of all samples to be classified. Next, the selected sample is labeled by an expert, and the process continues by going back to the first step. This is repeated until enough training samples have been selected and labeled. Performing the first step—selecting a training sample—requires an explicit description of a so-called sampling strategy. In light of the three challenges listed above, we define the following criteria for an efficient sampling strategy in this context. In order to make efficient use of the data and domain expert’s time collecting a training set, the sampling strategy should…
a. … avoid samples that do not show any instance of a class (i.e., species),
b. … prioritize samples that show an instance of a class that is not yet in the training set,
c. … prioritize samples that show an instance of a class that is underrepresented in the training set, and
d. … prioritize samples that can help to discriminate a class better from the other classes.
A number of works have been published in the field of active learning in recent years [7,18,19], and one popular method is active learning using uncertainty sampling [18]. The basis of uncertainty sampling is the estimation of a classifier’s uncertainty regarding the classification of each sample. This allows the automatic selection of the sample with the highest estimated uncertainty to be labeled by the experts in a training step to increase the potential of a classifier’s ability to discriminate the according class from the other classes. In [19], the authors propose a two-class active learning method that generates a clustering prior to the actual classification. Then, the algorithm assigns a higher priority to examples the closer they are to the classification boundary and the closer they are to a cluster centroid. In [7], an initial clustering is also performed, and relevance scores are assigned to the clusters. The relevance score is supposed to represent the extent to which the cluster can obtain samples that are likely to have greater potential to enhance the classifier’s performance. The relevance scores are then used to determine a cluster from which a sample is randomly drawn. The method has been shown to perform very well on a marine image data set. However, the method employs so called hand crafted feature representations to classify the images, i.e., so-called dominant color features, an established color feature representation method that is often applied in image retrieval contexts. The dominant color feature for an image patch is extracted by grouping the pixels (i.e., their rgb-colors) into a number of five clusters using the modified median cut algorithm [20]. The mean of all color vectors in the cluster containing the highest number of color vectors is the image patch’s dominant color feature. Dominant color features are not expected to work well on other datasets containing species that are visually distinguished by shape rather than color. This is likely to be the case for datasets from a different location, or perhaps even a dataset taken at the same location but with a new hardware setup. In recent years, convolutional neural networks have been successfully proposed as a very powerful approach to computer vision problems, making the selection and tuning of classic hand crafted features like dominant colors obsolete.
In this paper, we propose ALMI, a new active learning method for the object classification in marine images using a generic deep learning feature representation. ALMI takes sub-images, referred to as image patches in the following, and returns for each patch a class label, describing its content. ALMI is built on two conceptual ideas: First, it combines uncertainty sampling with relevance sampling to automatically select the next sample to be classified by domain experts and added to the training set. Second, it achieves a new level of flexibility by employing deep learning features instead of hand-crafted features like dominant colors (see above) proposed in prior works. We use two different data sets from marine imaging to demonstrate our method and evaluate its performance in comparison with other methods. The two data sets differ regarding the location and water depth where they were recorded, and consequently regarding the taxonomic composition of species they contain. Moreover, the images from one data set were taken from a cabled fixed underwater observatory (FUO) while the images from the other data set were taken from a moving towed ocean floor observation system (OFOS). Our experiments show that our method is able to perform well on image sets that differ in various aspects. In both data sets, our generic approach outperforms the state-of-the-art methods without any extensive tuning towards the individual data set.
The data sets used as input for ALMI in our experiments are explained in more detail in the next section. Section 3 describes the proposed method ALMI, and the results of the evaluation are given in Section 4. Section 5 will discuss the evaluation and wrap our findings up.
2. Materials
The first data set (see Figure 1) was created from an open-access, still image data set taken at the Hausgarten observatory with an Ocean Floor Observation System (OFOS) [21,22]. The original images are publicly available as described in the Data Availability Statement at the end of this paper. The Hausgarten observatory currently includes 21 stations located between (N 78.5°, E 05°) and (N 80°, E 11°) between Greenland and Svalbard. The OFOS was towed to a research vessel and took images of size3504×2336 at a depth of 2500 m. From the OFOS images, sub-images showing one object (like a sea star or a crustacean for instance) were extracted by the authors of [6]. The resulting data set will be referred to as Hausgarten dataset (HG). The dataset HG was used in [6] to evaluate the COATL learning architecture and will be used in our experiments in Section 4.1. The HG data set used in this work contains 1815 image patches grouped into 9 classes.
The second data set (see one example image in Figure 2) was created using images from the Lofoten-Vesterålen (LoVe) Ocean Observatory. The original images are publicly available as described in the Data Availability Statement at the end of this paper. LoVe is a cabled fixed underwater observatory located at (N 68° 54.474′, E 15° 23.145′) in the Norwegian Sea about 22 km offshore. The observatory monitors a coral reef at a depth of about 260 m. Among other sensors, the observatory is equipped with a high-resolution digital camera taking images of the coral reef. One image of size5184×3456 pixels is taken once per hour. The change detection method proposed in [16] was used to extract sub-images containing at most one object per sub-image from 24 LoVe images in an unsupervised fashion. The resulting dataset will be referred to as LoVe data set (LV). The LoVe dataset was used in [7] to evaluate an active learning method and will be used in our experiments in Section 4.2. The LV dataset used in this work contains 3031 image patches grouped into 6 classes. It mainly consists of one image patch class “no object” showing no objects of interest (see Figure 2 on the right).
3. Methods
The proposed active learning workflow ALMI (see Figure 3) takes a setI=Ii|1≤i≤Nof imagesIi(from here on, we will use the term image instead of image patch) as input and assigns the images to classes that have a semantic meaning. First, a fully automatic initialization step is performed to prepare the data for the semi-automatic labeling process where semantic classes are found and training samples of all classes are labeled and added to the training set. The initialization step starts by extracting from each imageIi, a feature vectorfithat represents the image in a lower-dimensional (here 300 dimensional) vector space. Next, the feature vectors are grouped into M clustersCj,(1≤j≤M)where features that are similar to each other belong to the same cluster. For1≤i≤N, we denote bycithe unique cluster index j withfi∈Cj . Moreover, a relevance score is computed for each cluster. This score estimates the potential of the cluster’s items to improve the classifier’s learning performance in learning new classes not represented in the training set (see Section 3.2 below).
Next, the training set is composed, and in each iteration, the following three steps are performed:
-
A sample imageIiis chosen automatically according to the sampling efficiency criterion (defined below in the Sampling efficiency algorithm section)
- An expert classifies the sample into a class found in a previous iteration or into a new class.
- The classifier is retrained to update the uncertainties used in the sampling criterion (step 1).
The trained classifier can then be used to classify the remaining samples. 3.1. Feature Extraction
For further processing, for each imageIi, a feature vectorfi is computed that describes the image in a lower dimensional vector space. Due to the limited background knowledge problem formulated in the introduction, the image feature representation cannot be built with hand-crafted features using heuristics without a strong loss in generalization (see Section 1). Instead, we propose to use the InceptionV3 Net [23], a fully convolutional deep learning network, to extract featuresfifor any network inputIi. These features however are abstract and are learned automatically during a pre-training step (see below).
A deep learning network takes the imageIi as input and passes it through a number of so-called layers that transform the input and pass it to the next layer until the the image is classified in the last layer. In the case of fully convolutional networks like InceptionV3, the layers mainly consist of a number of filters. In the InceptionV3 Net (see Figure 4), the layers are grouped into so-called inception modules inspired by the Inception Net described in [24]. The inception modules take the output of the previous inception module as input (or the original image in case of the first inception module) and perform multiple convolutions with different kernel sizes. The convolution results are then stacked on top of each other and passed to the next inception module. The output of the last inception module can be seen as a feature vector that is passed to the last layer for classification The filter-weights are not predefined by human experts but are learned during the training, where images are classified and the filter-weights are adjusted in an iterative process to optimize the classification performance. In case of the InceptionV3 Net about25×106parameters (mainly filter-weights) are learned during the training.
To make sure that the filter-weights are set properly, we use an InceptionV3 Net that was pretrained on a large set of images, the ImageNet [25]. ImageNet is a list of web images that provides access to more than14×106 annotated and quality-controlled web images. The list includes images showing examples of a variety of concepts such as sports, foot, animal, fish, etc. The subset of marine animals contains 1348 images. The InceptionV3 Net pre-trained on the ImageNet data set used in this work was downloaded using tensorflow [26].
To generate a feature vectorfifor one imagesIi, the images are fed into the pretrained InceptionV3 Net and propagated through the layers. The output of the last layer of the last inception module will be denoted byfi˜. As these featuresfi˜ are 2048-dimensional, we reduce the dimension in order to enhance the computation time and performance of the classifier. To do so, we use Principal Component Analysis (PCA) [27]. PCA is a method for dimension reduction that can be thought of as understanding the feature vectors as datapoint in the 2048-dimensional euclidean space and transforming them into a new coordinate system that is determined in the following way. The first axis is determined to minimize the sum of the squared distances between itself and each data point. The other axes are determined one by one in the way that each axis minimizes the sum of squared distances between itself and the data points under the condition of being orthogonal to all previously determined axes. After transforming the feature vectors in this way, all but the first 300 coordinates of each feature vector can be omitted without loosing too much information (compare the explained variance in Section 4.1 and Section 4.2). The feature vectors obtained this way will be denoted byfi.
3.2. Cluster Relevance
The feature vectors are grouped into M clusters (i.e., groups of feature vectors that are similar to one another) using a cluster method that takes the feature vectors as input and returns for each feature vectorfia cluster indexci(1≤ci≤M) as output. The choice of the clustering method is not crucial in this context and is in general not dependent on the data set or the kind of imaging setting. The method we use for the dataset HG is agglomerative clustering. Agglomerative clustering starts with N clusters and assigns each feature vector to its own cluster. Next, the number of clusters is reduced by iteratively choosing two clusters according to a given criterion (here: wards criterion) and merge them until only m clusters remain. Here, we use wards criterion that chooses the two clusters to be merged in the way that the increase of the in-cluster variance is minimal.
In the case of the dataset LV, the imagesIi were extracted automatically from larger images using the change detection method BFCD [16] which only is applicable to images from a fixed camera. In its core, BDFC extracts the imagesIicontaining maximum one object per image as sub-images from a time-series of large scale images. This is done by clustering the pixel-wise differences of the large scale images to the pixels of the mean image of the large scale images. In that process not only the imagesIiare returned but also a cluster indexci(1≤i≤M) for each image is returned, so no additional clustering is required for the dataset LV.
For both data sets, the relevance scores of the clusters are computed as follows. For a cluster j (1≤j≤M), let
Sj=fi|1≤i≤N∧ci=j
denote the set of feature vectors that belong to cluster j and let
mj=1Sj·∑fi∈Sjfi
denote the centroid of cluster j. With the mean of all feature vectors
C=1N·∑i=1Nfi
and the Euclidean distanced2, the relevance scorerjof cluster j is defined as the distance
rj=d2C,mj
between the centroid of cluster j and the mean of all feature vectors.
3.3. Sampling Efficiency Algorithm Motivated by the criteria (a–d) defined in the introduction, we implement a sampling algorithm to select the next training sample in two steps:
Step 1
—Cluster selection: For selecting, a cluster, let the activity scoreajdenote the number of times a sample has been selected from cluster j in the previous iterations. By defining
xj=∞ifaj=0rj ajelse
and selecting the cluster
j˜=arg maxj(xj)
it is ensured that the frequency that a sample is drawn from cluster j is approximately proportional torj.
Step 2
—Training sample selection: Let{ω1,…,ωK(T) }denote theK(T)classes that are present in the training set during the sample selection in iteration t. IfK(T)≥2 , uncertainty sampling [18] is used to draw a sample from the clusterj˜selected in step 1: Given a samplefi to be classified, a chosen classifier (e.g., the support vector machine (SVM) [28] that is used in the experiments in this paper) computes for each classωkthe probabilitypi,k(1≤k≤K(T))thatfibelongs toωk. With
δi(t)=1if sample i has been labeled before iteration t0else
the characteristic function that indicates if a sample has been labeled in an iterationt′<t, the uncertainty of the classifier regarding the classification of a featurefican then be expressed as
ui=0δi(t)=11−max1≤k≤K(T)(pi,k)else
To select the sample where the classifier is most uncertain, the samplefi˜with
i˜=arg max{1≤i≤N|ci=j˜}ui
is selected. In caseK(T)<2, a classifier can not be trained and a sample is drawn randomly with uniform distribution from the clusterj˜selected in step 1.
3.4. Classification Uncertainty
As a last step in each iteration, the classifier has to be trained to obtain the uncertainties that are used in the next iteration. In the t-th iteration, a number of t samples have been labeled by the expert. The labels of the labeled samples are propagated to the remaining samples using the clusters found in Section 3.2. To each cluster j, the labell^jis assigned that occurs most often in cluster j, according to
l^j=arg max1≤k≤K(T)({1≤i≤N|δi(t)=1∧ci=j∧li=ωk})
The labels assigned to the clusters are then used to assign a label to each sample, according to
l˜i=liifδi(t)=1l^ci else
The featuresfiand their labelsl˜iare then used to train the classifier. The trained classifier is then used to predict for each samplefiand each classωkthe probabilitypi,k(1≤j≤K(T))thatfibelongs toωk. These probabilities are then used in the next iteration to compute the uncertainties during the selection of the next sample.
4. Evaluation
ALMI is evaluated on the LoVe dataset and the Hausgarten dataset. The real-life application with the human expert iteratively labeling the data as described above is simulated with the data sets LV and HG that have been entirely labeled with gold standard classificationsgiby domain experts in advance. During each iteration, when the label for an imageIiis queried, the a priori determined labelgiis assigned toIi . As a classifier, the Support Vector Machine [28] (SVM) is used. The main idea of the SVM for two classes is to find a hypersurface that separates the classes in the training set. To do so, the SVM transform the samples into a higher-dimensional vector space until a separating hyperplane can be found. The samples and the hyperplane are then transformed back to the original vector space ending up with a hypersurface that separates the training data. A new sample can then be classified by determining on which side of the hypersurface the sample is located.
For each of the data sets, published results of a state-of-the-art method are available for comparison. Each of these methods adds the training samples one by one in an iterative fashion to the training data set similar to the iterations of the computer-assisted labeling described in Section 3.3 and Section 3.4. At the end of each iteration t, the classification performance is evaluated on a test setT=Ii1 ,…,IiN^ ⊂Iwith the numberN^of test samples as described in the following two subsections.
To compute a classifier’s performance onT, letgiτ denote the gold standard label assigned to imageIiτ ∈Tby human experts. Furthermore, lethiτ(t)denote the label assigned to featurefiτ by the classifier when trained with t training samples. An often-used method to evaluate a classifier’s performance is to compute the accuracy defined by
a(t)=1T1≤τ≤N^|hiτ(t)=giτ
which describes the proportion of correctly classified samples in all classified samples. However, for a fair evaluation of the methods on the LoVe dataset and the Hausgarten dataset this performance measure will be changed slightly to match the evaluation in [6] and [7] as described in the following two subsections.
4.1. Evaluation on the Dataset HG
In this experiment, the proposed method is evaluated on the dataset HG. First, the result of the principal component analysis is inspected. As described in Section 3.1, the PCA is used to reduce the InceptionV3 Net features from a length of 2048 to a length of 300. For this dataset, the explained variance of the first 300 principal components was determined to be94.2% . Next, the proposed method is compared to the COATL-approach proposed in [6]. The core idea of COATL is to use different classifiers depending on the number of available expert labels. No classifications are made until five labels are available. From five to 20 available labels, a K-Nearest-Neighbors approach is used. From 20 to 400 labels, an SVM is used. From 400 to 1500 labels, an H2SOL [6] is used. Moreover, when more than 1500 labels are available, a convolutional neural network is used.
As proposed in [6], after each iteration t, the performance is evaluated on all samples except for the t labeled samples. By doing so, the test data set consists ofN−timages after iteration t. To avoid testing on a too small test dataset, only 1500 iterations are performed which leaves 315 test samples after the last iteration. In both methods COATL and ALMI, a number ofnkfirst classifications are neglected. In case of COATL, this number is set tonk=5. In case of ALMI, all image classifications are neglected before more than one class has been learned, i.e.,K(T)>1withK(T) as the number of classes learned after T iterations. That is why a slight modification of the accuracy given in Equation (12) is used in this experiment.
aHGt=1N−t×1≤i≤N|hi(t)=gi∧δ(t)=0ift≥5∧K(T)≥20else
The accuracyaHGt is computed for the proposed method and for COATL after each iteration. That is done 10 times for each method, and the average accuracies for each method and each number of samples t are shown in Figure 5.
The results show that our new proposed generic method shows a steeper learning rate than COATL, even without any particular tuning for this data. To achieve an accuracy of75%, the proposed method just needs 68 labels (4.5%of the training data), while COATL needs 1023 labels (68.2%of the training data). An accuracy of80%is achieved by the proposed method with 150 labels (10%of the training data), while COATL does not achieve an accuracy of80%.
After 150 iterations when ALMI has an accuracy of80.0%, the accuracy of COATL is69.9% . To show which species are effected, the confusion matrix of COATL and ALMI after 150 iterations is shown in Figure 6.
The rows represent the true labels while the columns represent the predicted labels, i.e., the number in rowιcolumnκrepresents how often an instance of classιhas been predicted as classκin average over the 10 runs. The numbers in brackets on the main diagonal show the class-wise accuracies and have been computed as follows. For a classιlet
-
TPιdenote the number of instances of classιthat have been correctly classified as classι,
-
TNιdenote the number of instances of any classκ≠ιthat have not been classified as classι,
-
FPιdenote the number of instances of any classκ≠ιthat have incorrectly been classified as classι, and
-
FNιdenote the number of instances of classιthat have been incorrectly classified as classκ≠ι.
The accuracy of classιis then defined as
a(ι)=TPι+TNιTPι+TNι+FPι+FNι
On the first glance, the class-wise accuracies obtained by the proposed method ALMI seem to be better or equal to the class-wise accuracies obtained by COATL except for the class “burrowing purple anemone”. In fact, also for the class “shrimp” COATL performs a little better, as ALMI has more false negatives than COATL. That is not reflected by the accuracy as the number of true negatives is quite high compared to the number of false negatives for this underrepresented class leading to a accuracy of >0.95 for both methods. As the overall accuracy shows, ALMI still outperforms COATL. As can be seen in the upper right triangles of the confusion matrices, this is mainly due to fact that ALMI fixes the problem that COATL tends to assign species incorrectly to more abundant classes. 4.2. Evaluation on the Data Set LV
In this experiment, ALMI is evaluated on the dataset LV. First, the feature vectors are inspected. As described in Section 3.1, the InceptionV3 Net features are reduced from a length of 2048 to a length of 300 using principal component analysis. For this dataset, the explained variance of the first 300 principal components was determined to be95.6%.
Next, ALMI is compared to the active learning approach based on dominant color features, described in [7]. As in [7], 200 runs of the experiment have been conducted. Both ALMI and the method described in [7] do not classify any sample before the number K of classes in the training set exceeds 1. Several data-specific aspects had to be considered in this evaluation. First, the test dataset is not strictly separated from the training dataset and the classification performance is evaluated after each iteration on the whole dataset including the labeled training data. Second, the expert labels available after t iterations are included in the evaluation. To do so, let for1≤i≤N
h^i(t)=liifδi=1hi(t)else
denote the labels that are used to train the classifier.
Third, to take the strong data imbalance and the dominating abundance of images with no objects (see Figure 2 top right) into account, the performance measurement had to be adapted. Otherwise, the performance measurement would easily measure very high accuracies even for a naive classifier, classifying all images to the no object class. Thus, to neglect the correctly classified no object-samples in the HG experiments the following changes were applied to the number of images considered in the evaluation
N^=1≤i≤N|¬h^i(t)=0∧gi=0
and to the performance measure
aLVt,r=1N^×1≤i≤N|h^it=gi∧h^it≠0ifK(T)≥20else
The accuracyaLVt,ris computed for ALMI and COATL after each iteration t. The superscripts t and r denote here that the accuracyaLVt,r has been measured in the r-th run of the experiment in iteration t (i.e., with t labeled samples). Along with the mean, Figure 7 shows the standard deviations computed and visualized as follows. With
μt=1200∑r=0200aLVt,r
denoting the mean of the accuracy at iteration t averaged over all runs of the experiment, the standard deviation of the accuracies after iteration t is defined as
σt=∑r=0200(μt−aLVt,r)2
The area between the curves ofμt−σtandμt+σtis then filled with a semi-transparent color.
Figure 7 shows that also on this dataset the results of ALMI outperform previously published results on the same dataset. To achieve an accuracy of90%, ALMI needs 258 labels (8.5% of the dataset), while the method in [7] needs 279 labels (9.2% of the dataset). When the method in [7] achieves90%accuracy, ALMI has already reached94%accuracy. In other words, ALMI has about40% fewer misclassifications when trained with 279 labels. Regarding the standard deviation, the plot shows two things. First, after about 110 iterations, the accuracies achieved by ALMI vary less than the accuracies achieve by the method proposed in [7]. Second, when the method proposed in [7] reaches an accuracy of0.9%, the mean accuracy of ALMI is about one time the standard deviation larger than0.9.
To test these findings about improving accuracy for statistical significance, we apply two tests in order to check whether the variance and/or the mean of the accuracy obtained after 279 iterations differs significantly depending on whether ALMI or the reference method from [7] is used. For this, letA=(aLV279,1,…,aLV279,200)denote the accuracy values obtained by ALMI after 279 iterations and letB=(bLV279,1,…,bLV279,200) denote the according accuracy values obtained by the method proposed in [7]. As a first test, we use the Levene test [29] to test if the variances of A and B differ significantly. In the Levene test, the null hypothesis states that the variances of A and B are equal. Next, a p-value is computed that describes the probability that the variances of A and B differ more or equal than the actually observed variances under the assumption of the null hypotheses. In our experiment, the Levene test results in a p-value of about3×10−9. This is by far smaller than the typically chosen threshold of0.05 which shows that the null hypothesis should be rejected and the difference of variances is highly significant. As a second test, we apply a one-sided Welch’s test [30] to test if the mean of A is significantly larger than the mean of B. In the one-sided Welch’s test, the null hypothesis states that the mean of A is lower or equal than the mean of B. Next, a p-value is computed that describes the probability thatmean(A)−mean(B)is larger or equal than the actually observed difference between the means of A and B under the assumption of the null hypothesis. In our experiment, the Levene test results in a p-value of about3×10−12. This is by far smaller than the typically chosen threshold of0.05which shows that the null hypothesis should be rejected and the difference of the accuracies is highly significant.
5. Discussion and Conclusions
The aim of this work was to present a generic method for marine image classification that shows an improved learning performance due to the use of generic features and a reasoned choice of training samples in order to increase the efficiency of the manual annotation task performed by human experts. The proposed method ALMI is a single label-image classification method, i.e., the images of the processed dataset are required to contain maximum one object per image. However, if that is not the case, single-object images can be extracted from large scale images prior to using ALMI fully automatic. Some methods are proposed in [14,15,16] where the method proposed in [16] expects images from fixed cameras. The other two methods can be applied to any kind of dataset, e.g., image from OFOSs, FUOs, or semi-mobile platforms such as pan/tilt units.
To evaluate the extent to which the method meets this objective, its performance was compared to other related works. The evaluation focused on several aspects. First, the method was evaluated on very disparate data sets in order to assess the effectivity of the generic feature approach. Second, results from previously published evaluations of existing methods on the same data sets had to be available so results can be reproduced. Third, the evaluation was done in the same way as the previously published evaluations in order to visualize the progress. Regarding the evaluation in Section 4.2, one may observe that the data was not split into test set and training set, which is of course common practice. However, in active learning, it makes sense to leave the training data in the test set to avoid a decrease of the test set’s quality. The decrease of the test set’s quality is more prominent in a setup where object classes are underrepresented, and the true positive non-object samples are not considered in the accuracy: As discussed above, an important feature of a good sampling strategy is to draw samples from (potentially underrepresented) classes that contain actual objects of interest. If these samples are removed from the test set, the test set’s quality decreases faster with a “good” sampling strategy than with a strategy that draws many “no object” samples. This is especially illustrated by the following two points, which become only apparent when the training data are removed from the test set and become apparent more quickly with a good sampling strategy than with a sampling strategy that selects many “no object” samples.
- When all the samples of an underrepresented class are in the training set, the underrepresented class is not part of the test set anymore.
- When all the samples that are not in the “no object” class are in the training set, the accuracy is 0 because the test set only contains “no object” images that are not counted as true positives.
- it can achieve higher accuracies than previously published methods and
- it has a steeper learning curve than, i.e., ALMI achieves a certain level of accuracy with less training samples.
These effects are more prominent on the data set HG and are especially remarkable on the data set LV, because on this data set the previously published method outperformed the results of other known methods to such an extent that further improvement seemed difficult to achieve. Considering the large differences between the two marine image data sets, this all suggests that ALMI has the potential to apply to a wide range of marine image data sets and makes us confident that it can be useful in the biodiversity estimation in different types of marine habitats.
Author Contributions
T.M. and T.W.N.: Conceptualization, methodology, validation, formal analysis, investigation, data curation, writing-review and editing, visualization, and resources; T.M.: software and writing-original draft preparation; T.W.N.: supervision, project administration, and funding acquisition; All authors have read and agreed to the published version of the manuscript.
Funding
This work has been funded by the Bundesministerium für Wirtschaft und Energie (BMWI, FKZ 0324254D) and financially supported through Equinor Contract No. 4590165310. The authors would like to thank Equinor for the financial support and for providing field images.
Data Availability Statement
Links to the orignal Hausgarten images described in Section 2 paragraph 1 can be downloaded at https://doi.pangaea.de/10.1594/PANGAEA.615724?format=textfile (Last accessed on 30 December 2020). The original LoVe images described in Section 2 paragraph 2 are available under https://love.equinor.com/ (Last accessed on 30 December 2020).
Conflicts of Interest
The authors declare no conflict of interest.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2021. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
In recent years, an increasing number of cabled Fixed Underwater Observatories (FUOs) have been deployed, many of them equipped with digital cameras recording high-resolution digital image time series for a given period. The manual extraction of quantitative information from these data regarding resident species is necessary to link the image time series information to data from other sensors but requires computational support to overcome the bottleneck problem in manual analysis. As a priori knowledge about the objects of interest in the images is almost never available, computational methods are required that are not dependent on the posterior availability of a large training data set of annotated images. In this paper, we propose a new strategy for collecting and using training data for machine learning-based observatory image interpretation much more efficiently. The method combines the training efficiency of a special active learning procedure with the advantages of deep learning feature representations. The method is tested on two highly disparate data sets. In our experiments, we can show that the proposed method ALMI achieves on one data set a classification accuracy A > 90% with less than N = 258 data samples and A > 80% after N = 150 iterations, i.e., training samples, on the other data set outperforming the reference method regarding accuracy and training data required.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer