1 Introduction
The past few decades have seen various nature-inspired algorithms being highlighted to resolve numerical optimization issues. These algorithms are key players in unravelling a multitude of engineering optimization problems due to worldwide investigation and their exploitability. They are characterised by their mimicry of living organisms’ behaviour in nature, like the general fauna living on land or in water, respectively. As such, metaheuristic searching optimization has recently gained a huge interest in being utilised in a wide range of well known optimization problems and engineering applications. It has applications in various fields, such as power optimization [1–5]; text clustering [6]; smart traffic management [7]; robotic [8]; networking [9–11]; data security [12, 13], engineering [14–16]; and machine learning [17–25].
Regardless of the different concepts and natural inspirations behind the approaches for meta-heuristic searching optimization, they have one common fundamental structure. That is the utilisation of a heuristic-based selective search method in the solution space to find the best solution that optimizes a given objective function. In the case of a multi-optimization problem, a set of objective functions are applied as long as the set of constraints is preserved. The factors that have attracted researchers in recent times to these algorithms are the rapid advancement of hardware performance and their ability to solve numerous problems in the engineering field; they are also attractive due to the simplicity of their objective function and constraints. Several nature-inspired searching optimization frameworks have been proposed based on the multitude of natural occurrences seen worldwide; some of these natural occurrences include the particle swarms, krill herds hunting behaviour of bats, black holes, bee food searching behaviour, ant colonies, improvisation process of jazz musicians, and evolutionary algorithms such as genetic algorithm, and differential evolution [26–35].
There were several attempts for modifying and enhancing the performance of the nature-inspired algorithms by updating the architecture of the algorithms for handling different case studies. In [36] the authors introduced an improved version of the Harmony Search (HS) algorithm, called the Improved Harmony Search (IHS) algorithm. The IHS algorithm combines the power of HS with fine-tuning capabilities of mathematical techniques to achieve high-quality solutions with fewer fitness function evaluations. The authors demonstrated the effectiveness of the proposed approach in several test problems, where it outperforms other evolutionary and mathematical programming techniques reported in the literature. The work by [37] proposed a discrete variation of the Grey Wolf Optimizer (GWO) called the Discrete GWO (DGWO) for scheduling dependent tasks in cloud computing environments. The scheduling process in DGWO is formulated as a minimization problem for computation and data transmission costs. The algorithm utilizes the largest order value (LOV) method to convert the continuous candidate solutions produced by GWO to discrete candidate solutions. On the other hand, the island model is another way to enhance the nature-inspired algorithms is a commonly used technique in nature-inspired algorithms, such as genetic algorithms and Cuckoo Search Algorithm evolutionary algorithms. It involves partitioning the population into multiple sub-populations or islands and applying the optimization algorithm independently to each island [38, 39].
Data clustering (DC) refers to classifying similar objects into a group whose content differed significantly from the objects contained in another group. DC is an unsupervised learning process as the objects are placed in unspecified and predetermined clusters. In contrast, classification is a method of learning with supervision, whereby objects are classified into predetermined groups (clusters). However, clustering is associated with the issue of the absence of antecedent knowledge of the dataset provided, as well as the challenging selection of various input parameters like the number of clusters, the number of nearest neighbours, and more. Wrongly selected parameters would inevitably result in bad outcomes. Moreover, below-par precision has also bogged down the algorithms in the case of datasets that host clusters of dissimilar complex shapes, densities, sizes, noise, and outliers [40].
Various real-world applications have implemented DC methods widely. The main aim of this approach is to partition data objects such that the accumulated distances between data objects and their respective centroids are minimized. By clustering, objects within a cluster should have as much similarity as possible while being significantly different from objects in other clusters. In other words, DC can be viewed as an optimization problem where the objective is to partition a given set of data points into a fixed number of clusters such that the within-cluster similarity is maximized and the between-cluster similarity is minimized.
Among the common approaches to solving DC, problems are to formulate it as a meta-heuristic optimization problem [41–46]. The study by [27] recently devised one of the meta-heuristic optimization methods known as a "black hole," which replicates the natural action of a black hole (BH) of drawing in neighbouring stars. The concept of BHA and its interaction with the neighbouring stars formed the basis of the BHA algorithm. In this regard, the work presented by [27] has flaws in terms of exploration as the process of obtaining an optimal resolution necessitates too many reiterations. The BHA and its enhanced versions have been utilised to tackle several well known optimization problems recently [47–61].
Recently, several metaheuristics have been enhanced by incorporating a multi-swarm or multi-population approach, including Genetic Algorithm (GA) [62], Artificial Bee Colony (ABC) [63], and Particle Swarm Optimizer (PSO) [15, 64–66], and Nomadic People Optimizer (NPO) [33] due to their capability to use different populations with their parameters set and they can simultaneously implement search space. As a result, they have significantly enhanced the performance of the original metaheuristic [67, 68]. This paper proposed multiple BHA optimization as a generalization to BHA optimization, in which the algorithm no longer depended on one best resolution. Instead, a set of best solutions were generated and called MBHA, which was maintained for some time in the search process. Furthermore, the algorithm’s objective function was replaced with an objective function of higher effectiveness to resolve the clustering issue. Additionally, it was also compared with the original BHA algorithm according to several datasets.
The rest of this article is organised as follows: The Section 2 will focus on some earlier reported data clustering methods, while Sections 3 and 4 focus on BHA and the proposed MBHA, respectively. Finally, Section 5 summarizes the experimental results, while Section 6 summarises the work.
2 Background
This section aims to offer an overview of the data clustering optimization problem and the black hole optimization algorithm. First, the section explains data clustering as an optimization problem, providing the necessary mathematical formulation. Additionally, the section presents a review of the most significant related works. The second subsection explains the original version of Black Hole Algorithm (BHA), and discusses its advantages and drawbacks.
2.1 The problem of data clustering
Clustering is a crucial approach to unsupervised data classification that involves grouping a set of vectors or patterns (such as data items, observations, or feature vectors) in a multi-dimensional space [69–71]. The process of DC is determined by the dataset categorization concept using a specific number of clusters while reducing the intra-object distance within each cluster. The rearrangement of a given set of data patterns is referred to as cluster analysis; it is usually represented by one of two things: 1) a vector of measurements; or 2) a point in a multi-dimensional space. The procedure is conducted to create clusters that are differentiated by similarity attributes [72]. Some of the common application areas of DC are image processing, analysis of medical images, as well as statistical data analysis. They are also useful in various science and engineering fields and are sometimes used interchangeably with statistical data analysis. The differences across clusters can be attributed to their sizes, shapes, and densities, as seen in Fig 1.
[Figure omitted. See PDF.]
(I) and (II) the difference between the data before and after performing the clustering. I. Before, II. After.
Nevertheless, noise present in the data presented may pose a challenge for cluster detection, whereby the ideal cluster is fundamentally designated as “a set of points that is compact and solitary. Although humans are commonly ascribed to their cluster seeking proficiency in probably three dimensions, automatic algorithms remain as the go-to for high-dimensional data.” This fact, alongside the undesignated number of clusters that are yet to be described for a provided dataset, has consistently generated thousands of clustering algorithms underlined in publications [73]. Meanwhile, the learning task can be described in pattern recognition, in which the data analysis section is commonly linked with predictive modelling. In this case, the training data is allocated to predict the unknown test data behaviour. Assessment of the data similarity may require the use of “distance measures; the problem may be designed thus: given N records data, each record is assigned to only one of the K clusters. After that, clustering is done using several criteria that serve as the process objective function (OF). The minimizing of the sum of squared ED between each record and the center of the related cluster” is one of the commonly observed features. This is shown below.(1)where ‖Qi−Zj‖ is the Euclidian Distance (ED) between a data record Oi and the cluster center Zj. N and K are the numbers of data records and the number of clusters, respectively.
Combining a nature-inspired optimization algorithm with a clustering algorithm has led to the generation of optimal solutions. The study by [74] has displayed the method of adaptive time-dependent transporter ant for clustering (ATTA-C), which underlines alterations to the standard “Ant Colony Optimizer (ACO)” Ant-based clustering algorithm. It aims to subject high dissimilarities to a penalty, enhance the spatial separation between clusters, and facilitate clustering procedures. Achieving this requires the calculation of the fitness value for each clustering solution, which is carried out using a neighbourhood function (NF). Meanwhile, the study [75] has underlined a novel Particle Swarm Optimizer (PSO) approach for clustering issues, which is implementable in the case of a known or unknown number of clusters. The algorithm is termed CPSO and proceeds according to the gbest neighborhood topology, encoding cluster centroids in particles and possibly generating new partitions during optimization. This occurs either by the removal or splitting of these clusters, until the allocated number of clusters is yielded.
Furthermore, an improved version of the Firefly Algorithm (FA) was proposed by [76] for a given dataset, in which the FA is employed and implemented for the training set to obtain the cluster centre via random selection of 75% of the dataset provided. Meanwhile, the remaining 25% dataset is termed a test dataset and utilized to investigate the FA algorithm performance [77]. The Krill Herd Algorithm (KHA) is mostly used to display a simulated herding pattern of each krill individual. The density-based approach utilized allows the discovery of clusters, subsequently undermining the region of adequately high density into clusters of krill individuals arbitrarily shaped in the climate. The objective goal of the krill movement is the minimum distance between individual krill from the food source and highly-dense herds. That is considered via foraging movement and random diffusion. In the case of a density-based cluster, it can be described as a set of density-linked objects of maximum concerning density-reach capability and noisy objects. The study [44] has previously suggested an artificial bee colony (ABC) clustering approach be subjected to categorical data. A one-step k-modes procedure is first developed for this particular approach before it is incorporated with the ABC to yield a categorical data cluster. Meanwhile, the study by [78] introduced C-ESA as a hybridization of the K-means clustering algorithm and Elephant Search Algorithm (ESA) for data clustering and obtaining the best centroid location and clustering precision enhancement.
In [79], a map/reduce programming for the ABC algorithm has been designed, capable of configuring and incorporating data in a multi-node environment. The ABC allowed the speediest completion time during execution, displaying its high efficiency for all types of data due to the parallelism attribute it offers. It also provides the amalgamation of local and global search techniques to achieve a trade-off between exploration and exploitation capabilities in obtaining optimized clusters. Similarly, the designed map/reduce programming utilizing ABC mechanisms is incorporated in a single node and multi-mode Hadoop platform, whereby the mapper phase generates the best fitness value by mimicking the behaviour of the employed bee. Meanwhile, the reducer phase achieves the probability value for cluster optimization by mimicking the onlooker and employed bees. The resulting experimental outcomes have predicted consecutively run times of varying dataset sizes in single-node and multi-node climates. Upon evaluating the performance displayed by the ABC scheme alongside the conventional Differential Evolution (DE) and PSO schemes, the ABC method was found to show superior results for optimal cluster selection compared to the remaining options. Furthermore, it also minimized the time for execution and errors in classification in the optimal cluster selection for multi-node Hadoop cluster architecture.
Meanwhile, fresh heuristic gravitational-based for data clustering has been described by [80], which answers to the excess centroid movement. Owing to the excess of centroid velocity history in the gravitational clustering algorithm, this serves as a way of improving the balance between exploration and exploitation capabilities. The technique includes an initialization phase that uses the variance and median approach so to avoid random initialisation effects. Following that, the centroid’s accumulated velocity history is removed, leaving only the force of the data points in the cluster associated with the centroid to influence its position throughout any iteration.
Besides, an alternative clustering method that is effective and superior shown by [81] has opted for the application of a nature-inspired krill herd algorithm. The problem is translated into an optimization search problem via objective function minimisation to distinguish the optimum centre of each cluster. Then, multiple real and synthetic databases are reviewed, with comparison studies undertaken to elucidate the purpose of the ESKH-C technique. The technique is specifically implemented to attain quality clustering using dissimilar dimensional real data and synthetic databases alike. The predicted outcome of confidence results from the simulation studies also indicated that the technique could group optimal cluster groups having different data shapes, sizes, dimensions, and densities. In [82], modified Bee Colony Optimization (MBCO) has been implemented, with its hybridization with k-means serving as a way of its application to data clustering. The technique is synonymous with bees’ traits of forgiveness and a fair chance, which is seen for trustworthy bees or their opposite alike. It is also associated with the probability-based selection (Pbselection) approach that allocates unassigned data points in every iteration. The paper by [83] has revealed a semi-supervised K clustering framework, whereby a K-means clustering framework is initially used for the gene data. Following this, an enhanced semi-supervised K means clustering is implemented for greedy iteration to identify the K mean clustering and obtain improved outcomes. Simulations have subsequently proven that a global semi-supervised K clustering algorithm offers superior capacity for optimization and cluster effect in comparison with MDO algorithm.
Overcoming the issue of local optimum in K-Means also resulted in [84], in which a new clustering framework is designed via hybridized Crow Search Optimization (CSA). A novel population-based metaheuristic optimization algorithm is rooted in the crows’ intelligent behaviour. Similarly, a K-Means clustering algorithm called CSAK means has also been suggested, whereas [43] has recently designed an Elephant Herding Optimization suited for clustering tasks. In this method, intra-cluster distance and cost function are reduced.
2.2 Black Hole Algorithm
Based on the black hole phenomena, the BHA is based on the core premise of an expanse of space housing a large volume of mass. The mass is concentrated within, making it impossible for any adjacent object to escape its gravitational pull. If one were to fall victim to the event, one would be obliterated from the cosmos, including light. The method is made up of two parts: 1) star movement and 2) star re-initialization upon entering the D-dimensional hypersphere around the BH (i.e., the event horizon). It functions as follows: the first step is the initialization of the N+1 stars, xi∈RD, i = 1,…,N+1 in the search space, where N = population size. The best value after subjection to a fitness evaluation is then recognized as the black hole xBH. Because it is known to be static, no movement is visible until other stars reach a higher resolution. As a result, the number of individuals searching for the best value is equal to N, and in each generation, a star is shifting towards the BH as seen in the following equation [27]:(2)where rand is a random number in the range [0,1].
Furthermore, the BHA suggests that a star that comes too close to the BHA and passes through the event horizon would be removed. The following equation describes the radius of the event horizon (R) [27]:(3)where fi and fBH are the BH’s and ith star’s fitness values, respectively. N represents the number of stars considered as the candidate solutions.
When R is greater than the distance between a potential solution and the BH (the best solution), the related candidate is automatically collapsed, causing the formation of a new possible solution that is distributed arbitrarily over the search space. The BHA is characterized by a simple structure requiring no parameters and can be easily implemented. Compared to the other heuristics, the BHA converges to the global optimum in all iterations, unlike the other heuristics that can be trapped in locally optimal solutions [27, 85]. Although using BHA as a clustering method is associated with outstanding results, it has drawbacks due to a lack of balance between exploration and exploitation capabilities. Finding a better solution than the existing BHA will alter the direction of a star, thereby changing the star’s orientation into a new BH. Furthermore, the event horizon must be conceptualized due to the stars’ possible rapid convergence for the solution space to be absorbed by the BH. This problem is caused by the lack of exploration capabilities by the BHA. It does not, however, provide intensified processes for exploration or information collection regarding previously found solutions; instead, it is just a restart approach that is applied to each star [86].
3 Multi-population Black Hole Algorithm
The weakness in the exploration capability of the Black Hole Algorithm (BHA) stems from its low diversity population. The algorithm tends to converge too quickly to local optima, which limits its ability to explore the search space and find global optima [87]. Therefore, in case of the exploitation capabilities are being performed more than the exploration capabilities, the chances of being trapped in a local optimum are increased. In this paper, an enhanced version of the BHA algorithm was proposed and called the “Multi-Population Black Hole (MBH) Algorithm” for the problem of data clustering. MBHA is based on the original BHA algorithm but uses multiple populations instead of a single one. Each population comprises several candidate solutions (stars) that undergo random generation in the search space. Then, the populations are initialised and each of their fitness values is assessed, whereby the best candidate having the best fitness value is chosen as the black hole. At the same time, the rest reverts to becoming normal stars. As the black hole can absorb stars around it, such a process of star absorption occurs after the black hole and stars are initialised, at which the stars move. The absorption process has been formulated as seen below:(4)where xi(t+1) and xi(t) are the location of the ith star at iteration t and t+1, xBH is the location of the black hole in the search space, c is a constant, rand is a random number in the interval [0, 1], and N is the number of stars (candidate solutions) in the population. The constant c is utilized to restrict solutions scattering in the space, as well as to yield a higher convergence speed for the algorithm.
While running the algorithm, a star (or the BH) in a population may arrive at a location offering lesser cost compared to the current black hole or not reach it. This results in the concept of Search Counter (SC), which defines the number of times a population evolves without finding an improved fitness value. Therefore, if a star reaches a better location, there will be a probability of generating a new star for that population (probgenerating_star), and this probability is formulated as follows:(5)where SCmax is the maximum value of SC.
After checking the probability of generating a new star, the SC will be reset to zero. This probability helps the population that loses many stars due to the cessation of evolution for some time to acquire new stars and give them a longer life span. A population loses some of its stars due to crossing the event horizon in case of the limitations of a black hole in space shaped as a sphere. The black hole will suck in every star that ventures into its event horizon, whereby every star death is characterized by a new replacement star of probability (probreplace) that is arbitrarily distributed in the search space. The probreplace is formulated as probgenerating_star, which will help the progressing population to keep its number of stars as large as possible. The calculation for the radius of the event horizon in the BHA algorithm is done using Eq (3).
A population must be omitted if the number of its stars becomes less than the minimum allowed a number of stars in a population. At each iteration, there will be a probability of generating a new population (probgenerating_population), which will help to explore the entire search space and avoid the local minima at a minimum number of iterations (speed up the convergence to global optima in early iterations).(6)where rand is a random number in the interval [0.1]. The solutions of the new population are generated in two ways: 1) arbitrarily in the search space, and 2) arbitrarily chosen from other populations. The ratio rg is used to mix between the two ways and is formulated as follows:(7)where itr is the iteration of generating the new population and max iterations refers to the total number of iterations. Therefore, the search process during the early iterations is considered to be a global search (rg) is small, and the solutions are arbitrarily generated in the search space. As the iterations continue, it becomes a local search (rg) is become larger and the solutions are taken from other populations. Note that the value of rg can be also selected as a constant. Thus, to generate a new population, there are two cases: if rg is less than 50% of the total number of iterations then generate a new random population, otherwise, generate the population based on the position of the global best black hole (BHG) as shown in the following equations:(8)(9)where Xi represents a new star in the population P, while r1 and r2 represent a randomly selected population, and a randomly selected star from that population, and rand is a random number in the range [0,1]. This work can overcome BH’s weaknesses and make a good balance between global search and local search. The key processes for the enhanced BHA algorithm are subsequently summarised using the following pseudocode in Fig 2, while the flowchart is given in Fig 3.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
4 Results and discussion
MBHA performance was assessed by carrying out two sets of experiments. Firstly, several mathematical objective functions with multiple local minima were used to further evaluate the developed algorithm and to compare it with the original BHA and other related works. Secondly, MBHA algorithm has been validated and tested based on six benchmark datasets, and compared to other powerful state-of-art algorithms.
4.1 Evaluation on benchmark test functions
To highlight MBHA for superior exploration compared to the standard BH, further verification has been carried out via a set of multi-model types of objective functions in a multi-dimensional space. Table 1 has succinctly outlined elements like the functions and their key features such as the “Name, Dimensions (D), Upper and Lower Boundaries (UB, LB),” the optimal solution (Opt) values are also stated in Table 1 while the parameter setting is in Table 2. These parameter settings were utilized in their default values as specified in the original versions. Moreover, each function is also associated with the generation of the convergence curve of the search, which is then differentiated from the actual BHA algorithm. The simulation was done using Matlab 2018a on a PC with the following specifications: Core i7, 16 GB RAM, 3.6 GHz, 64-bit Windows 10 OS.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
The performance of the new MBHA was benchmarked against 9 popular metaheuristics which are Genetic Algorithm (GA) [88], Arterial Bee Colony (ABC) algorithm [89], Particle Swarm Optimization (PSO) [90], Levy Firefly Algorithm (LFFA) [91], Grey Wolf Optimizer (GWO) [92], Ant Colony Optimization Algorithm (ACO) [30], Bat algorithm (BA) [93], Flower Pollination Algorithm (FPA) [94], and Blackhole (BH) [27]. The assessments and experiments were carried out accordingly, with MBHA and BHA being subjected to 30 different runs each. As a result, the best mean, error rate, and standard deviation were calculated, as seen in Table 3 for each algorithm.
[Figure omitted. See PDF.]
Although the statistical results presented in Table 3 provides a first insight into the performance of MBHA, a Wilcoxon Signed-Rank Test pair-wise statistical test with a statistical significance value ∝ = 0.05 is utilized for a better comparison. A Wilcoxon signed rank test is needed to compare the performance of MBHA against standard BHA and PSO individually.
The null hypothesis (H0) for Wilcoxon signed Rank test is that there is no significant median difference between the mean pair of samples. The results are compared to other methods at a 95% level of confident. Here, if the Wilcoxon statistic is less or equal to the alpha (α = 0.05), then H0 will be rejected. To perform the statistical calculations, the SPSS statistics Software Version. In Table 4, the statistical results of BHA, PSO, GA, ABC and ACO algorithms compared to MBHA are given.
[Figure omitted. See PDF.]
The convergence curve was also generated for the searching pattern of the 6 functions and compared with that of the original BHA. As can be seen from Figs 4 to 9 the MBHA has shown faster convergence curves. For the test problems, MBHA showed a better fitness value than BHA during all the optimization processes. This means the MBHA is more efficient than BHA and more suitable for the optimization problem.
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
[Figure omitted. See PDF.]
4.2 Evaluation on benchmark dataset
To ensure a fair comparison with existing methods, the same datasets used in the original version of the black hole algorithm and related works were utilized. Using different datasets would make it difficult to compare performance. Although testing on multiple datasets is important, consistency in dataset selection was prioritized. Six datasets were utilised to evaluate the performance of the suggested algorithm for data clustering: Iris, Wine, Glass, Cancer, Contraceptive Method Choice (CMC), and Vowel. Table 5 has outlined each of their specific attributes, whereby the datasets were all obtained from the UCI ML laboratory.
1. Iris dataset. Consisting of 150 arbitrary samples of flowers, the dataset’s samples possessed four features of the iris flower and were grouped into 3 groups that were made up of 50 instances.
2. Wine dataset. For this dataset, the quality of the wine was depicted according to its physicochemical attributes, as they were originally harvested from an identical Italian region but of 3 different cultivars. The three wine types were associated with 178 instances each, with 13 numeric features representing the number of 13 components in each wine type.
3. CMC dataset. This dataset is a subset of the 1987 National Contraceptive Prevalence Survey carried out in Indonesia. The sample size was made up of married women who, during the interview period, were either not pregnant or unknown of their pregnancy. It underlined the problem of anticipating recent choices of contraceptive techniques (i.e. no use, short-term use, or long-term use) per a woman’s socioeconomic and demographic attributes.
4. Cancer dataset. This dataset is representative of the Wisconsin breast cancer database; it is made up of 9 components with 683 instances; the 9 components are “Clump Thickness, Cell Size Uniformity, Cell Shape Uniformity, Marginal Adhesion, Single Epithelial Cell Size, Bare Nuclei, Bland Chromatin, Normal Nuclei, and Mitoses.” Every instance was attributed to being possibly benign or malignant, respectively.
5. Glass dataset. This dataset is made up of 214 objects with 9 features that included “refractive index, silicon, potassium, sodium, calcium, magnesium, aluminium, barium, and iron.” Meanwhile, six types of glass were used in the data sampling process; these are “non-float processed building windows, float processed building windows, containers, tableware, float-processed vehicle windows, and headlamps.”
6. Vowel dataset. This dataset is made up of 871 Indian Telugu vowel sounds; the dataset also has 3 attributes that correspond to the 1st, 2nd, and 3rd vowel frequencies, as well as 6 overlapping classes.
[Figure omitted. See PDF.]
The comparison stage was conducted by calculating four statistical values after executing the algorithms for 30 run times; the output was the sum of intra-cluster distances. These four values are (Best, Average, Worst, and Standard deviation). Additionally, all algorithms have been compared based on the value of the error rate. These two measurements can be defined as follows:
1. The sum of the distances between the clusters as a measure of internal quality: The calculation and summing up of the intra-cluster distances between the data centre and each data object is shown in Eq 1. A higher cluster quality is typically correlated to a smaller sum of intra-cluster distances, in which the sum of the distances between the clusters was one of the fitness components evaluated in this study.
2. Error Rate (ER) as an external quality measure: The equation below displays the percentage of misplaced data objects:
(10)
Several metaheuristic methods were compared with the performance of the proposed algorithm, such as K-means [72], PSO [35], ACO [95], KH [77], GSA [41], BB-BC [41], CS [96], TS [97], and BHA [27]. In addition, MBHA was also subject to a comparison with 9 of the recently modified hybrid meta-heuristics reported in the literature; these metaheuristics included: K-means++ [98], IKH [98], BSF-ABC [99], ACPSO [66], H-KHA [81], K-MCI, MCI [100], and NM- PSO, K-NM-PSO [99]. The results of the comparison based on the standard meta-heuristics clustering frameworks and the modified hybrid meta-heuristic for a better benchmarking of the MBHA are shown in Tables 6 and 8.
[Figure omitted. See PDF.]
A summary of error rate and intra-cluster distances is shown in Table 6. Each of the algorithms was implemented for 30 runs, and after the simulation runs, the values for the best, average, worst, standard deviation and error rate were for each algorithm. In the Table, the values in bold were the best-derived values using algorithms for each dataset. The results of the experiments showed that MBHA outperformed BHA and K-means. Further comparisons showed that the suggested technique achieved the least standard deviation compared to the other algorithms, implying that the MBHA is always at its minimum value.
Furthermore, the Iris dataset depicted MBHA algorithms having a convergence of 96.522 for each run. In contrast, the wine dataset indicated that the MBHA revealed the superior solution for worst 16,294.230, average 16,293.400, and standard deviations 0.7623. Moreover, the CMC dataset also showed the best, worst, and mean solutions obtained by MBHA of 5528.800, 5531.220, and 5530.000, with a standard deviation of 0.3466. On the other hand, the K-means, PSO, ACO, KH, GSA, BB-BC, CS, TS, and BHA failed to result in the best solutions.
Besides, the Cancer dataset via MBHA algorithm resulted in the best solution of 2961.950, whereas the glass dataset obtained 208.760 as its best optimum value. Meanwhile, MBHA showed the worst value of 211.569, in comparison with the algorithms K-means, PSO, ACO, KH, GSA, BB-BC, CS, TS, and BH, attaining the worst values over 30 runs were 227.350, 283.52, 280.08, 247.085, 248.367.21, 243.208, 227.022, 280.080, 213.956, respectively, for the glass dataset.
Similarly, the MBHA also obtained the best optimum value for the vowel dataset, which was 148,941.00. Therefore, it could be conclusively stated that the MBHA algorithm had achieved the near-best value in all runs and reassured its capacity to yield superior optimal solutions, notwithstanding a small standard deviation in a minimum number of iterations.
The algorithms were further compared statistically to check for significant differences in their performances; the statistical comparison was made using the Friedman and Iman–Davenport tests. Table 7 presents the performance of the algorithms based on the employed statistical tests.
[Figure omitted. See PDF.]
Table 8 compared the average instar-cluster distances and error rate of various clustering algorithms; MBHA yielded the best performance and conclusively revealed superior performance for all six datasets. The Iris dataset resulted in a 0.00010 standard deviation for the proposed algorithm, which was a value that was remarkably less in comparison with the remaining clustering algorithms. However, its best solution of 96.51300 and worst solution of 96.53200 was both superior compared to the remaining.
[Figure omitted. See PDF.]
For the Wine dataset, the proposed MBHA algorithm obtained an average value of 16,293.400 outperforming the rest of the algorithms, excluding ACPSO. Meanwhile, the CMC dataset obtained exceedingly superior performance for the proposed algorithm; the worst solution of 5531.220 was relatively better compared to the remaining algorithms by a wide margin.
The Cancer dataset revealed that the proposed MBHA depicted the best solution of 2961.950 and an average solution of 2963.900. Its standard deviation was 0.0072 and reassuringly superior compared to K-means++, IKH, BSF-ABC, ACPSO, H-KHA, K-MCI, NM-PSO, K-NM-PSO, and MCI. In contrast, the Glass dataset obtained the best solution of 199.860 using the K-MCI algorithm, while the final dataset of Vowel yielded 148,943.00 of the best average solution by MBH. Hence, this conclusively highlighted the effectiveness of MBHA to resolve complex optimization problems, simply due to the best results generated by almost all of the datasets and upon comparison with the remaining comparative algorithms. The outcomes were specifically achieved by adding the element of new operators.
5 Conclusion
Black Hole Algorithm (BHA) is a newly developed optimization method that offers a promising solution for addressing complex global optimization problems. However, one of the limitations of the BHA algorithm is that the lack of balancing between the exploration and exploitation, which increases the chances of trapping in local minima, thereby preventing it from finding the optimal solution. To overcome this issue, an enhanced version of the BHA based on a new multi-population architecture, has been employed in this work by applying effective enhancements including a global exploration operator that facilitates the rapid convergence of the algorithm towards optimal solutions. The proposed algorithm is called “Multi-Population Black Hole Algorithm (MBHA)”.
Simulation results demonstrate that the proposed algorithm is able to significantly reduce computation time and achieve its set objectives, thereby prompting further evaluation on data clustering problems. Furthermore, the outcomes confirm the suitability of the proposed algorithm for resolving clustering problems as compared with previous reports. Despite the numerous advantages of the MBHA algorithm, several aspects require further elucidation and investigation in future research. Firstly, the algorithm was only benchmarked on nine test functions, thus necessitating the use of more benchmark problems to provide a comprehensive assessment of its capabilities. Secondly, the issue of number of populations and their sizes presents a fascinating research area that deserves in-depth exploration. Lastly, improving the convergence of the MBHA algorithm represents a crucial research topic that warrants further investigation.
In conclusion, the proposed MBHA represents an effective optimization method that offers a viable alternative for solving complex global optimization problems. Nevertheless, further research is necessary to investigate the ability of the algorithm to handle different hard optimization problems, such as, feature selection, hyperparameters tuning for Support Vector Machine (SVM), and training artificial neural networks (ANN).
Acknowledgments
Authors would like to thank Data Ana
Citation: Salih SQ, Alsewari AA, Wahab HA, Mohammed MKA, Rashid TA, Das D, et al. (2023) Multi-population Black Hole Algorithm for the problem of data clustering. PLoS ONE 18(7): e0288044. https://doi.org/10.1371/journal.pone.0288044
About the Authors:
Sinan Q. Salih
Roles: Conceptualization, Methodology, Software, Validation, Writing – original draft
E-mail: [email protected] (SQS); [email protected] (AAA)
Affiliation: Technical College of Engineering, Al-Bayan University, Baghdad, Iraq
ORICD: https://orcid.org/0000-0003-0717-7506
AbdulRahman A. Alsewari
Roles: Supervision, Writing – review & editing
E-mail: [email protected] (SQS); [email protected] (AAA)
Affiliation: Data Analytics & AI research Group, College of Computing and Digital Technology, Faculty of Computing Engineering and the Built Environment, Birmingham City University, Birmingham, United Kingdom
ORICD: https://orcid.org/0000-0002-7802-6628
H. A. Wahab
Roles: Conceptualization, Formal analysis, Methodology, Writing – original draft
Affiliation: Faculty of Computing, Kuantan, Malaysia
Mustafa K. A. Mohammed
Roles: Data curation, Methodology, Visualization
Affiliation: University of Warith Al-Anbiyaa, Karbala, Iraq
ORICD: https://orcid.org/0000-0002-1850-6355
Tarik A. Rashid
Roles: Formal analysis, Visualization
Affiliation: Computer Science and Engineering Department, University of Kurdistan Hewler, Erbil, Iraq
ORICD: https://orcid.org/0000-0002-8661-258X
Debashish Das
Roles: Formal analysis
Affiliation: Data Analytics & AI research Group, College of Computing and Digital Technology, Faculty of Computing Engineering and the Built Environment, Birmingham City University, Birmingham, United Kingdom
Shadi S. Basurra
Roles: Investigation, Supervision, Writing – review & editing
Affiliation: Data Analytics & AI research Group, College of Computing and Digital Technology, Faculty of Computing Engineering and the Built Environment, Birmingham City University, Birmingham, United Kingdom
1. Zhang S, Zhang H, He Q, Bian K, Song L. Joint Trajectory and Power Optimization for UAV Relay Networks. IEEE Commun Lett. 2018;22: 161–164.
2. Bryson AE. Applied optimal control: optimization, estimation and control. Routledge; 2018.
3. Hassan MH, Kamel S, Salih SQ, Khurshaid T, Ebeed M. Developing chaotic artificial ecosystem-based optimization algorithm for combined economic emission dispatch. IEEE Access. 2021. https://doi.org/10.1109/ACCESS.2021.3066914
4. Guo H, Tao H, Salih SQ, Yaseen ZM. Optimized parameter estimation of a PEMFC model based on improved Grass Fibrous Root Optimization Algorithm. Energy Reports. 2020;6: 1510–1519.
5. Tao H, Salih SQ, Saggi MK, Dodangeh E, Voyant C, Al-Ansari N, et al. A Newly Developed Integrative Bio-Inspired Artificial Intelligence Model for Wind Speed Prediction. IEEE Access. 2020;8: 83347–83358. https://doi.org/10.1109/ACCESS.2020.2990439
6. Sangaiah AK et al. Arabic text clustering using improved clustering algorithms with dimensionality reduction. Cluster Comput. 2018.
7. Sekar E.V. et al. A framework for smart traffic management using hybrid clustering techniques. Cluster Comput. 2017; 1–16.
8. Li Z. et al. Discrete cuckoo search algorithms for two-sided robotic assembly line balancing problem. Neural Comput Appl. 2018;30: 2685–2696.
9. Chiang H.-S. et al. A novel artificial bee colony optimization algorithm with SVM for bio-inspired software-defined networking. Int J Parallel Program. 2018; 1–19.
10. Bhagat Tiyasha, Welde Tesfaye, Tung Al-Ansari, et al. Evaluating Physical and Fiscal Water Leakage in Water Distribution System. Water. 2019;11: 2091.
11. Bacanin N, Arnaut U, Zivkovic M, Bezdan T, Rashid TA. Energy Efficient Clustering in Wireless Sensor Networks by Opposition-Based Initialization Bat Algorithm. Computer Networks and Inventive Communication Technologies. 2022. pp. 1–16.
12. Zamli KZ, Alhadawi HS, Din F. Utilizing the roulette wheel based social network search algorithm for substitution box construction and optimization. Neural Comput Appl. 2022.
13. Alhadawi HS, Salih SQ, Salman YD. Chaotic Particle Swarm Optimization Based on Meeting Room Approach for Designing Bijective S-Boxes. Proceedings of International Conference on Emerging Technologies and Intelligent Systems. 2022. pp. 331–341.
14. Ghosh A., Mal P., Majumdar A. Advanced Optimization and Decision-Making Techniques in Textile Manufacturing. CRC Press; 2019.
15. Salih SQ, Alsewari AA, Yaseen ZM. Pressure Vessel Design Simulation: Implementing of Multi-Swarm Particle Swarm Optimization. Proc 2019 8th Int Conf Softw Comput Appl. 2019; 120–124.
16. Malik A, Rai P, Heddam S, Kisi O, Sharafati A, Salih SQ, et al. Pan Evaporation Estimation in Uttarakhand and Uttar Pradesh States, India: Validity of an Integrative Data Intelligence Model. Atmosphere (Basel). 2020;11: 553.
17. Bottou L., Curtis F.E., Nocedal J. Optimization methods for large-scale machine learning. Siam Rev. 2018;60: 223–311.
18. Allawi MF, Salih SQ, Kassim M, Ramal MM, Mohammed AS, Yaseen ZM. Application of Computational Model Based Probabilistic Neural Network for Surface Water Quality Prediction. Mathematics. 2022;10: 3960.
19. Tao H, Al-Sulttani AO, Salih Ameen AM, Ali ZH, Al-Ansari N, Salih SQ, et al. Training and Testing Data Division Influence on Hybrid Machine Learning Model Process: Application of River Flow Forecasting. Shahid S, editor. Complexity. 2020;2020: 1–22.
20. Yaseen ZM, Naghshara S, Salih SQ, Kim S, Malik A, Ghorbani MA. Lake water level modeling using newly developed hybrid data intelligence model. Theor Appl Climatol. 2020;141: 1285–1300.
21. Shehu A., Falope T, Ojim G, Abdullahi Y, Abba S. A Novel Machine Learning based Computing Algorithmin Modeling of Soiled Photovoltaic Module. Knowledge-based Eng Sci. 2022;3: 28–36.
22. Bacanin N, Zivkovic M, Stoean C, Antonijevic M, Janicijevic S, Sarac M, et al. Application of Natural Language Processing and Machine Learning Boosted with Swarm Intelligence for Spam Email Filtering. Mathematics. 2022;10: 4173.
23. Thakkar HK, Shukla H, Sahoo PK. Metaheuristics in classification, clustering, and frequent pattern mining. Cognitive Big Data Intelligence with a Metaheuristic Approach. Elsevier; 2022. pp. 21–70. https://doi.org/10.1016/B978-0-323-85117-6.00005–4
24. Bezdan T, Stoean C, Naamany A Al, Bacanin N, Rashid TA, Zivkovic M, et al. Hybrid Fruit-Fly Optimization Algorithm with K-Means for Text Document Clustering. Mathematics. 2021;9: 1929.
25. Salih SQ. A New Training Method Based on Black Hole Algorithm for Convolutional Neural Network. J Sourthwest Jiaotong Univ. 2019;54: 1–10.
26. Gandomi AH, Alavi AH. Krill herd: A new bio-inspired optimization algorithm. Commun Nonlinear Sci Numer Simul. 2012.
27. Hatamlou A. Black hole: A new heuristic optimization approach for data clustering. Inf Sci (Ny). 2013.
28. Yang XS, Deb S. Cuckoo search via lévy flights. 2009 World Congress on Nature & Biologically Inspired Computing (NaBIC). 2009. pp. 210–214.
29. Feng X, Yang T, Yu H. A new multi-colony fairness algorithm for feature selection. Soft Comput. 2017;21: 7141–7157.
30. Dorigo M., Birattari M. Ant colony optimization. Springer; 2010.
31. Yang XS, Deb S. Engineering optimisation by cuckoo search. Int J Math Model Numer Optim. 2010.
32. Zong WG, Joong HK, Loganathan GV. A new heuristic optimization algorithm: harmony search. Simulation. 2001;76: 60–68.
33. Salih SQ, Alsewari AA. A new algorithm for normal and large-scale optimization problems: Nomadic People Optimizer. Neural Comput Appl. 2020;32: 10359–10386.
34. Storn R, Price K. Differential Evolution—A Simple and Efficient Heuristic for Global Optimization over Continuous Spaces. J Glob Optim. 1997;11: 341–359. https://doi.org/10.1023/A:1008202821328
35. Kennedy J, Eberhart R. Particle swarm optimization. Proceedings of ICNN’95—International Conference on Neural Networks. IEEE; 1995. pp. 1942–1948 vol.4. https://doi.org/10.1109/ICNN.1995.488968
36. Mahdavi M, Fesanghary M, Damangir E. An improved harmony search algorithm for solving optimization problems. Appl Math Comput. 2007;188: 1567–1579.
37. Abed-alguni BH, Alawad NA. Distributed Grey Wolf Optimizer for scheduling of workflow applications in cloud environments. Appl Soft Comput. 2021;102: 107113.
38. Gozali AA, Kurniawan B, Weng W, Fujimura S. Solving university course timetabling problem using localized island model genetic algorithm with dual dynamic migration policy. IEEJ Trans Electr Electron Eng. 2020;15: 389–400.
39. Abed-alguni BH, Paul D. Island-based Cuckoo Search with elite opposition-based learning and multiple mutation methods for solving optimization problems. Soft Comput. 2022;26: 3293–3312.
40. Sarstedt M., Mooi E. Cluster analysis, in A concise guide to market research. Springer; 2019.
41. Hatamlou A., Abdullah S., Hatamlou M. Data clustering using big bang–big crunch algorithm,. Innov Comput Technol. 2011; 383–388.
42. Hatamlou A., Abdullah S., Othman Z. Gravitational search algorithm with heuristic search for clustering problems. 3rd Conference on Data mining and optimization (DMO). 2011.
43. Jaiprakash K.P., Nanda SJ. Elephant Herding Algorithm for Clustering, in Recent Developments i. Mach Learn Data Anal. 2019;5: 112–115.
44. Ji J. et al. A novel artificial bee colony based clustering algorithm for categorical data. PLoS One. 2015;10: e0127125. pmid:25993469
45. Kowalski P.A. et al. Nature Inspired Clustering–Use Cases of Krill Herd Algorithm and Flower Pollination Algorithm. Interactions Between Computational Intelligence and Mathematics Part 2. 2019. pp. 83–98.
46. Bagirov A M., Karmitsa N, Taheri S. Metaheuristic Clustering Algorithms. Partitional Clustering via Nonsmooth Optimization. 2020. pp. 165–183.
47. Pashaei E., Aydin N. Binary black hole algorithm for feature selection and classification on biological data. Appl Soft Comput. 2017;56: 94–106.
48. Bouchekara H. Optimal power flow using black-hole-based optimization approach. Appl Soft Comput. 2014;24: 879–888.
49. Lenin K., Reddy B.R. and MSK . Dwindling of active power loss by enhanced black hole algorithm. Int J Res Electron Comm Tech. 2014;1: 11–15.
50. Rodrigues D. et al. Black hole algorithm for non-technical losses characterization. 6th Lat Am Symp Circuits Syst (LASCAS). 2015.
51. Kacha L, Zitouni A, Djoudi M. KAB: A new k-anonymity approach based on black hole algorithm. J King Saud Univ—Comput Inf Sci. 2022.
52. Qasim OS, Al-Thanoon NA, Algamal ZY. Feature selection based on chaotic binary black hole algorithm for data classification. Chemom Intell Lab Syst. 2020;204: 104104.
53. Pashaei E, Pashaei E. Gene selection using hybrid dragonfly black hole algorithm: A case study on RNA-seq COVID-19 data. Anal Biochem. 2021;627: 114242. pmid:33974890
54. Azizipanah-Abarghooee R. et al. Short-term scheduling of thermal power systems using hybrid gradient based modified teaching–learning optimizer with black hole algorithm. Electr Power Syst Res. 2014;108: 16–34.
55. Nemati M., Momeni H. Black holes algorithm with fuzzy Hawking radiation. Int J Sci Technol Res. 2014;3: 85–88.
56. Bouchekara HR. Optimal design of electromagnetic devices using a black-hole-based optimization technique. IEEE Trans Magn. 2013;49: 5709–5714.
57. Doraghinejad M. Nezamabadi-pour H. Black hole: a new operator for gravitational search algorithm. Int J Comput Intell Syst. 2014;7: 809–826.
58. Eskandarzadehalamdary M., Masoumi B., Sojodishijani O. A new hybrid algorithm based on black hole optimization and bisecting k-means for cluster analysis. Electrical Engineering (ICEE), 2014 22nd Iranian Conference. 2014.
59. Yaghoobi S., Hemayat S. and HM . Image gray-level enhancement using Black Hole algorithm. i. 2nd International Conference on Pattern Recognition and Image Analysis (IPRIA). 2015.
60. Pashaei E., Ozen M. and NA . An application of black hole algorithm and decision tree for medical problem. 15th International Conference on Bioinformatics and Bioengineering (BIBE). 2015.
61. Premalatha K and RB. A nature inspired swarm based stellar-mass black hole for engineering optimization. international Conference on Electrical, Computer and Communication Technologies (ICECCT). 2015.
62. Pourvaziri H. BN . A hybrid multi-population genetic algorithm for the dynamic facility layout problem. Appl Soft Comput. 2014;24: 457–469.
63. Biswas S. et al. Co-evolving bee colonies by forager migration: A multi-swarm based Artificial Bee Colony algorithm for global search space. Appl Math Comput. 2014;232: 216–234.
64. Yazdani D. et al. A novel multi-swarm algorithm for optimization in dynamic environments based on particle swarm optimization. Appl Soft Comput. 2013;13: 2144–2158.
65. Salih SQ, Alsewari AA, Al-Khateeb B, Zolkipli MF. Novel Multi-Swarm Approach for Balancing Exploration and Exploitation in Particle Swarm Optimization. In Proceesdings of 3rd International Conference of Reliable Information and Communication Technology 2018 (IRICT 2018). Springer; 2018. pp. 196–206.
66. Liang JJ, Suganthan PN. Dynamic multi-swarm particle swarm optimizer. Proceedings—2005 IEEE Swarm Intelligence Symposium, SIS 2005. 2005. https://doi.org/10.1109/SIS.2005.1501611
67. Li C., Yang S. and MY . An adaptive multi-swarm optimizer for dynamic optimization problems. Evol Comput. 2014;22: 559–594. pmid:24437666
68. Li C. et al. Multi-population methods in unconstrained continuous dynamic environments: The challenges. Inf Sci (Ny). 2015;296: 95–118.
69. Golalipour K, Akbari E, Hamidi SS, Lee M, Enayatifar R. From clustering to clustering ensemble selection: A review. Eng Appl Artif Intell. 2021;104: 104388.
70. Zhang X, Lin Q, Mao W, Liu S, Dou Z, Liu G. Hybrid Particle Swarm and Grey Wolf Optimizer and its application to clustering optimization. Appl Soft Comput. 2021;101: 107061.
71. Anderberg MR. Cluster analysis for applications. 1973.
72. Jain AK. Data clustering: 50 years beyond K-means. Pattern Recognit Lett. 2010;31: 651–666.
73. Kumar Y and PKS. A chaotic teaching learning based optimization algorithm for clustering problems. Appl Intell.: 1–27.
74. Handl J., Knowles J. and MD . Ant-based clustering and topographic mapping. Artif Life. 2006;12: 35–62. pmid:16393450
75. Cura T. A particle swarm optimization approach to clustering. Expert Syst Appl. 2012;39: 1582–1588.
76. Senthilnath J., Omkar S. and VM . Clustering using firefly algorithm: performance study. Swarm Evol Comput. 2011;1: 164–171.
77. Singh V. and Sood K M.M. Krill Herd clustering algorithm using dbscan technique. Int J Comput Sci Eng Technol. 2013;4: 197–200.
78. Tian Z. et al. Elephant search algorithm on data clustering. 12th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery (ICNC-FSKD). 2016.
79. Ilango S.S. et al. Optimization using artificial bee colony based clustering approach for big data. Cluster Comput. 2018; 1–9.
80. Alswaitti M, Ishak MK, Isa NAM. Optimized gravitational-based data clustering algorithm. Eng Appl Artif Intell. 2018;73: 126–148.
81. Abualigah LM, Khader AT, Hanandeh ES, Gandomi AH. A novel hybridization strategy for krill herd algorithm applied to clustering techniques. Appl Soft Comput. 2017;60: 423–435.
82. Das P., Das SD D.K. A Modified Bee Colony Optimization (MBCO) and it’s hybridization with k-means for an application to data clustering. Appl Soft Comput. 2018.
83. Mai X., Cheng SW J. Research on semi supervised K-means clustering algorithm in data mining. Cluster Comput. 2013; 1–8.
84. Lakshmi K, Visalakshi NK, Shanthi S. Data clustering using K-Means based on Crow Search Algorithm. Sādhanā. 2018;43: 190.
85. Kumar S, Datta D, Singh SK. Black Hole Algorithm and Its Applications. Studies in Computational Intelligence. 2015. pp. 147–170.
86. Piotrowski A.P., Napiorkowski PMR J.J. How novel is the “novel” black hole optimization approach? Inf Sci (Ny). 2014;267: 191–200.
87. Abdulwahab HA, Noraziah A, Alsewari AA, Salih SQ. An Enhanced Version of Black Hole Algorithm via Levy Flight for Optimization and Data Clustering Problems. IEEE Access. 2019;7. https://doi.org/10.1109/access.2019.2937021
88. Davis L. Handbook of genetic algorithms. 1991.
89. Karaboga D, Basturk B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J Glob Optim. 2007;39: 459–471.
90. Zambrano-Bigiarini M., Clerc RR M. Standard particle swarm optimisation 2011 at cec-2013: A baseline for future pso improvements. IEEE Congress on Evolutionary Computation. 2013.
91. Yang X-S. Firefly algorithm, Levy flights and global optimization, in Research and development. Intelligent systems XXVI. 2010. pp. 209–218.
92. Mirjalili S, Mirjalili SM, Lewis A. Grey wolf optimizer. Adv Eng Softw. 2014;69: 46–61.
93. Yang X-S. Engineering optimizations via nature-inspired virtual bee algorithms. International Work-Conference on the Interplay Between Natural and Artificial Computation. Springer.; 2005.
94. Yang XS. Flower pollination algorithm for global optimization. International conference on unconventional computing and natural computation. Berlin, Heidelberg: Springer; 2012. pp. 240–249. https://doi.org/10.1007/978-3-642-32894-7_27
95. Zhang C., Ouyang D. and JN . An artificial bee colony approach for clustering. Expert Syst Appl. 2010;37: 4761–4767.
96. Boushaki S.I., Kamel OB N. A new quantum chaotic cuckoo search algorithm for data clustering. Expert Syst Appl. 2018;96: 358–372.
97. Liu Y, Yi Z, Wu H, Ye M, Chen K. A tabu search approach for the minimum sum-of-squares clustering problem. Inf Sci (Ny). 2008.
98. Jensi R. GWJ . An improved krill herd algorithm with global exploration capability for solving numerical function optimization problems and its application to data clustering. Appl Soft Comput. 2016;46: 230–245.
99. Ghafarzadeh H. AB . An Efficient Hybrid Clustering Method Using an Artificial Bee Colony Algorithm and Mantegna Lévy Distribution. Int J Artif Intell Tools. 2016;25: 1550034.
100. Krishnasamy G., Kulkarni RP A.J. A hybrid approach for data clustering based on modified cohort intelligence and K-means. Expert Syst Appl. 2014;41: 6009–6016.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer
© 2023 Salih et al. This is an open access article distributed under the terms of the Creative Commons Attribution License: http://creativecommons.org/licenses/by/4.0/ (the “License”), which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Abstract
The retrieval of important information from a dataset requires applying a special data mining technique known as data clustering (DC). DC classifies similar objects into a groups of similar characteristics. Clustering involves grouping the data around k-cluster centres that typically are selected randomly. Recently, the issues behind DC have called for a search for an alternative solution. Recently, a nature-based optimization algorithm named Black Hole Algorithm (BHA) was developed to address the several well-known optimization problems. The BHA is a metaheuristic (population-based) that mimics the event around the natural phenomena of black holes, whereby an individual star represents the potential solutions revolving around the solution space. The original BHA algorithm showed better performance compared to other algorithms when applied to a benchmark dataset, despite its poor exploration capability. Hence, this paper presents a multi-population version of BHA as a generalization of the BHA called MBHA wherein the performance of the algorithm is not dependent on the best-found solution but a set of generated best solutions. The method formulated was subjected to testing using a set of nine widespread and popular benchmark test functions. The ensuing experimental outcomes indicated the highly precise results generated by the method compared to BHA and comparable algorithms in the study, as well as excellent robustness. Furthermore, the proposed MBHA achieved a high rate of convergence on six real datasets (collected from the UCL machine learning lab), making it suitable for DC problems. Lastly, the evaluations conclusively indicated the appropriateness of the proposed algorithm to resolve DC issues.
You have requested "on-the-fly" machine translation of selected content from our databases. This functionality is provided solely for your convenience and is in no way intended to replace human translation. Show full disclaimer
Neither ProQuest nor its licensors make any representations or warranties with respect to the translations. The translations are automatically generated "AS IS" and "AS AVAILABLE" and are not retained in our systems. PROQUEST AND ITS LICENSORS SPECIFICALLY DISCLAIM ANY AND ALL EXPRESS OR IMPLIED WARRANTIES, INCLUDING WITHOUT LIMITATION, ANY WARRANTIES FOR AVAILABILITY, ACCURACY, TIMELINESS, COMPLETENESS, NON-INFRINGMENT, MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Your use of the translations is subject to all use restrictions contained in your Electronic Products License Agreement and by using the translation functionality you agree to forgo any and all claims against ProQuest or its licensors for your use of the translation functionality and any output derived there from. Hide full disclaimer