Fuzzy C Mean Find the Global Optimal Solution
Comput Intell Neurosci. 2020; 2020: 1386839.
Hybrid Fuzzy Clustering Method Based on FCM and Enhanced Logarithmical PSO (ELPSO)
Jian Zhang
School of Mechanical Engineering, Tongji University, Shanghai 200092, China
Zongheng Ma
School of Mechanical Engineering, Tongji University, Shanghai 200092, China
Received 2019 Nov 10; Revised 2020 Feb 5; Accepted 2020 Feb 14.
This article has been cited by other articles in PMC.
- Data Availability Statement
-
The data used to support the findings of this study are available from the corresponding author upon request.
Abstract
Fuzzy c-means (FCM) is one of the best-known clustering methods to organize the wide variety of datasets automatically and acquire accurate classification, but it has a tendency to fall into local minima. For overcoming these weaknesses, some methods that hybridize PSO and FCM for clustering have been proposed in the literature, and it is demonstrated that these hybrid methods have an improved accuracy over traditional partition clustering approaches, whereas PSO-based clustering methods have poor execution time in comparison to partitional clustering techniques, and the current PSO algorithms require tuning a range of parameters before they are able to find good solutions. Therefore, this paper introduces a hybrid method for fuzzy clustering, named FCM-ELPSO, which aim to deal with these shortcomings. It combines FCM with an improved version of PSO, called ELPSO, which adopts a new enhanced logarithmic inertia weight strategy to provide better balance between exploration and exploitation. This new hybrid method uses PBM(F) index and the objective function value as cluster validity indexes to evaluate the clustering effect. To verify the effectiveness of the algorithm, two types of experiments are performed, including PSO clustering and hybrid clustering. Experiments show that the proposed approach significantly improves convergence speed and the clustering effect.
1. Introduction
In order to obtain effective information on huge quantities of data quickly and accurately, many methods have been proposed. As an unsupervised learning method, clustering analysis is one of the vital means in dealing with these data whose objective is to partition an unlabeled dataset into a number of clusters, such that elements in same cluster show a high level of similarity, while elements from different clusters show a high level of dissimilarity. The clustering technique has been studied extensively in a variety of application fields such as data mining, machine learning, pattern recognition, and image segmentation [1–3].
Clustering algorithms can be further divided into two basic categories: hard and fuzzy [4]. Hard clustering methods assign each object to a single group, while fuzzy clustering methods introduce membership degrees between objects and the different clusters of the dataset and assign each element of a dataset to multiple clusters simultaneously in accordance with the membership function matrix. Therefore, the latter can handle overlapping partitions.
The most popular fuzzy clustering algorithm is fuzzy c-means (FCM) which was proposed by Bezdek et al. [5] and has been widely used in multiple domains [6, 7]. The goal of FCM is to minimize the criterion function and obtain a more accurate membership matrix gradually. But the random selection of center points makes iterative process fall into the saddle points or local optimal solution easily. Furthermore, if the datasets contain severe noise points or if the datasets are high dimensional, such as bioinformatics [8], the alternating optimization often fails to find the global optimum.
However, these shortcomings have motivated the proposal of alternative approaches for fuzzy clustering, many of which are extensions of FCM. A kernel-based FCM (KFCM) was proposed by Zhang and Chen [9], which replaces the Euclidean distance metric with a kernel metric to achieve better mapping for nonlinear separable datasets. Lin [10] proposed a novel evolutionary kernel intuitionistic FCM clustering algorithm (EKIFCM) that combines intuitionistic fuzzy sets (IFSs) with KFCM and utilizes genetic algorithms (GA) to optimize the parameters of EKIFCM simultaneously. Although these FCM versions aim to achieve good performance in fuzzy clustering, they do not improve the random initialization process of FCM and still fall into local optimum easily [11].
The probability of finding the global optimum may be increased by stochastic methods such as evolutionary or metaheuristic optimization algorithms. As one of the most famous metaheuristic methods, PSO has become one of the most popular metaheuristics and an important tool for many applications due to its versatility and simplicity, and it is found that it can provide better initial centroids for the FCM algorithm to improve the FCM results, and thus this has motivated the proposal of many PSO-based methods for hard clustering [12] and some PSO-based methods for fuzzy clustering [11, 13, 14]. Cura [15] presented a new PSO approach to the clustering problem, employing the pure PSO technique to solve both clustering problems with both known and unknown numbers of clusters, which provides a new idea for clustering.
Izakian and Abraham [16] proposed a hybrid fuzzy clustering method based on FCM and fuzzy PSO (FPSO), and their experiments show better results than FPSO and FCM. The quantum-behaved particle swarm optimization (QPSO) with a fully connected topology is coupled with the FCM, forming a new version of hybrid method called QPSO-FCM [17]. However, these PSO-based methods are much slower compared to the traditional methods which may limit their practical applications.
Another problem with PSO-based clustering methods, according to Alam [12], is the need to tune a range of parameters before they are able to find a better solution. For overcoming these shortcomings, hybrid methods for fuzzy clustering based on fuzzy c-means and improved particle swarm optimization (FCM-IDPSO) were proposed by Silva Filho et al. [18], who introduced the IDPSO for adjusting the parameters dynamically during training and tackling the two main problems of PSO-based clustering methods. Many improved PSO-FCM clustering methods have been successfully applied to practical applications [19–22]. It is worth noting that the complex structure of PSO-based methods and the huge amount of computation make the algorithm have room for further improvement.
In recent years, many excellent hybrid methods have been proposed for optimal cluster analysis, which do not use PSO as optimization algorithm, such as CRO-FCM [23] which uses chemical-based metaheuristic obtaining optimal cluster centers for FCM; ETLBO-FCM [24] incorporates elicit teaching learning-based optimization and FCM to overcome the major limitations of FCM; Rahul et al. [25] introduced bat optimization to FCM and utilized maxi-min classifier to determine the count of clusters, and the results showed that the clustering accuracy is improved. These studies have greatly promoted the development of clustering algorithms.
One of the main contributions of this paper is to introduce a new version PSO with enhanced logarithmic decreasing strategy (ELPSO) for clustering. Based on this strategy, ELPSO takes different inertia weight values during various periods adaptively and thus provides better balance between exploration and exploitation and avoids falling into local minima quickly, thereby obtaining better solutions. The other contribution of this paper is to propose a new method for the fuzzy clustering problem using hybridization combining FCM with ELPSO, named FCM-ELPSO, which makes use of the merits of both algorithms. This hybrid method introduces ELPSO for training process and uses ELPSO's global exploration to find a suitable initial clustering prototype for FCM and the local exploration to avoid falling into local optimum and utilizes the fast convergence of FCM to improve the results and convergence time. Both clustering methods are tested based on UCI datasets independently, and the results are compared to other PSO-based clustering methods, respectively.
The structure of the paper is as follows. Section 2 outlines all necessary prerequisites. In Section 3, a new version of PSO for clustering, named ELPSO, and the hybrid method (ELPSO-FCM) are proposed. Section 4 includes the results of experiments based on UCI datasets. In Section 5, main conclusions are covered.
2. Theoretical Basis
In this section, we briefly describe some basic concepts of FCM, original PSO (or standard PSO, SPSO) and some improved versions of PSO with different inertia weight strategies, and a cluster index which is used in the hybrid method for evaluating the clustering effect.
2.1. FCM
We define S={s 1,…, s j ,…, s N } as a clustering dataset of N objects indexed by j; each object s j is represented by a vector of quantitative variable. We define B={β 1,…, β i ,…, β C } as the prototypes of C clusters listed by i and U=[u ij ] C×N as a fuzzy partition matrix, where u ij indicates the membership of the j th object with the i th prototype. s j , β i ∈ R Q , where Q is the data dimensionality. The constraints on u ij are as follows:
(1)
(2)
(3)
The goal of the FCM algorithm is to find the optimal prototype matrix and the corresponding membership degree matrix that minimize an objective function given by the following equation:
where m(m > 1.0) is the fuzzy weighting exponent and d ij is the Euclidean distance that indicates the dissimilarity from data vectors s j to cluster center β i .
The parameter d ij is obtained by the following equation:
To minimize the criterion J, the clustering prototypes β i and the membership degrees u ij are updated according to equations (6) and (7), respectively.
(6)
(7)
After computing the memberships of all the objects, the new prototypes of the clusters are calculated. The process stops when the prototypes stabilize. That is, the prototypes from the previous iteration are of close proximity to those generated in the current iteration, normally less than an error threshold.
2.2. Original Particle Swarm Optimization
PSO is originally introduced in terms of social and cognitive behavior of bird flocking and fish schooling. The potential solutions are called particles which fly through the problem space by following the current best particles. Each particle keeps track of its coordinates in the problem space which are associated with the best solution that has been achieved so far. The solution is evaluated by the fitness value, which is also stored. This value is called pbest. Another best value that is tracked by the PSO is the best value, obtained so far by any particle in the swarm. The best value is a global best and is called gbest. The search for the better positions follows the rule as equations (8) and (9):
(8)
where x l and v l are position and velocity vector of the particle l, respectively; ω is the inertia weight; c 1 and c 2 are positive constants, called acceleration coefficients which control the influence of pbest l and gbest in search process; and r 1 and r 2 are random values in the range [0, 1]. The fitness value of each particle's position is determined by a fitness function, and PSO is usually executed with repeated application of (8) and (9) until a specified number of iterations have been exceeded or the velocity updates are close to zero over a number of iterations.
2.3. Some Improved Versions of PSO with Different Inertia Weight Strategies
Using statistical theory to analyze the variance of the basic parameters of PSO, including inertia weight and accelerating constants, it can be considered that the inertia weight has tremendous impact on the overall performance of PSO [26]. Many studies have shown that larger inertia weight values have better global search capabilities, and smaller inertia weight values have advantages in local exploitation [27]. So, different adaptive strategies of inertia weight are proposed to achieve a better balance between exploration capabilities and development capabilities and get more stable and satisfactory results, such as linear, nonlinear, fuzzy rules, random, and other strategy-based inertia weights.
In this section, three kinds of inertia weight strategies will be emphatically reviewed which are widely used in a variety of application domains, and the process of corresponding algorithms can be found in [28–30]. The method proposed in this paper will be compared with the above algorithm in Section 4.
2.3.1. Linear Inertia Weight Strategy
The monotonic decreasing inertia weight adjustment strategy is introduced into PSO by Eberhart [28] and aimed to enhancing the fine-tuning ability of PSO. But linear inertia weight strategy cannot achieve the accurate balance between local search and global search due to the nonlinearity and complexity of the PSO search process. So, it does not always perform better than an appropriate fixed inertia weight when the inertia weight decreases gradually as the iteration proceeds.
2.3.2. Natural Exponential Inertia Weight Strategy
Inspired by linear decreasing inertia weight strategy, Chen et al. [29] proposed two inertia weight strategies of natural exponential functions. Based on their experimental settings, these natural exponential strategies have a faster convergence speed in the early stage of PSO search process compared with the linear adjustment strategy.
2.3.3. Random Inertia Weight Strategy
It is difficult to predict whether in a given time the exploration or exploitation would be better in the dynamic environment. So, randomness is introduced into the inertia weight strategy of PSO to address this problem in [30]. Using particle swarms to track and optimize dynamic systems, a new way of calculating the inertia weight value is proposed.
2.4. Cluster Index PBM(F)
Pakhira et al. [31] proposed a validity index called the PBM index. The index is developed for both crisp and fuzzy clustering; however, here we review only the fuzzy version of the index called the PBM(F) index. The index is defined as
where ; D C =max i,n=1 C ‖β i − β n ‖; C is the number of clusters; and is the center of dataset S.
J m is different from J and considered to be
(11)
where N is the total number of patterns in the dataset, U(S)=[u ij ] C×N is a partition matrix for the data, and β i is the centroid of the i th cluster; here, the fuzzy parameter m 1 is set to 1.5.
The factor, E 1/J m , includes the sum of weighted intracluster distances for the complete dataset taken as a single cluster and that for the c cluster system. This factor is a measure of the compactness of a C cluster system. The factor D C is the maximum intercluster separation in a C cluster system. This factor signifies between cluster separation. Higher values of the PBM(F) index indicate better clustering in the sense that the clusters are well separated and relatively compact.
3. Proposed Algorithms
In this section, we will introduce the new version PSO with enhanced logarithmic decreasing strategy in detail, named ELPSO, and give the algorithmic process for clustering application; next, based on ELPSO and FCM, a hybrid algorithm is formed for combining the merits of these two algorithms, called FCM-ELPSO.
3.1. Enhanced Logarithmical PSO (ELPSO)
In order to adjust the performance of particle swarm and balance the global search and local search capabilities of the swarm in flight process, a simple and effective inertia weight adjustment strategy was introduced into PSO and a new version of PSO, called enhanced logarithmic decreasing PSO (ELPSO), was developed. The new strategy function is formulated as follows:
where t is the current iteration and z is the regulatory factor for fine-tuning ability of PSO, whose value can be set to 1.05 by experience. Equations (13) and (14) show the new velocity formula and position formula of particle l at instant t using the new inertia weights:
(13)
The size of each element is consistent in equations (13) and (14) except the parameters R 1l and R 2l . In order to increase the randomness of particle swarm search, we set the random value R as a matrix. Random matrixes of each particle will be initialized during every iteration, and the range of each element in the matrix is [0, 1].
Here, we give the method for clustering which employs the pure ELPSO technique.
Let the position of particle l, represented by X l , be the prototype matrix, whose size is C × Q, where C is the right cluster number and Q is the dimension of the datasets l ∈ [1, P] in which P is the size of population. In this way, X l may be expressed as follows:
(15)
Therefore, a swarm represents a number of candidates' cluster center for the data vector. Each data vector belongs to a cluster according to its membership function and thus a fuzzy membership is assigned to each data vector. Each cluster has a cluster center per iteration and presents a solution which gives a vector of cluster centers. This method determines the position vector X l for every particle, updates it, and then changes the position of cluster center. And the fitness function for evaluating the generalized solutions is stated as
where J(X l ) is the objective function of FCM, as shown in equation (4), calculated for particle l. The smaller J(X l ) is, the better is the clustering effect and the higher is the fitness function f(X l ).
The fake code is shown as follows.
3.2. The Hybrid Methods for Fuzzy Clustering Based on Fuzzy c-Means and Improved Particle Swarm Optimization
Although FCM requires fewer function evaluations, it usually falls into local optima. In this section, the FCM algorithm is integrated with ELPSO algorithm to form a hybrid clustering algorithm called FCM-ELPSO which maintains the merits of both FCM and ELPSO algorithms. This hybrid method introduces ELPSO for training process and uses ELPSO's global exploration to find a suitable initial clustering prototype for FCM and the local exploration to avoid falling into local optimum and utilizes the fast convergence of FCM to improve the results and convergence time.
The fake code is shown as follows.
4. Experiments and Results
This section is divided into two parts: ELPSO clustering and hybrid clustering, can use Algorithm 1 and Algorithm 2 for obtaining corresponding results separately. All experiments are based on the platform Matlab 2016b and executed on an Intel core i7-8750H 2.20 GHz computer running Microsoft Windows 10.
For evaluating the performance of the proposed algorithms, nine well-known UCI Machine Learning Repository datasets have been selected: Abalone, Ecoli, Glass, Image segmentation, Page blocks classification, Spectf, Steel plates faults, Ultrasonic flowmeter diagnostics, and Yeast. These datasets include examples of low, medium, and high dimensional data with various partitions. A detailed description of the datasets is shown in Table 1.
Table 1
Descriptions of the real datasets.
Datasets | Objects | Variables | Groups |
---|---|---|---|
Abalone | 4177 | 3 | 8 |
Ecoli | 336 | 7 | 8 |
Glass | 214 | 9 | 6 |
Image segmentation | 2310 | 19 | 7 |
Page blocks classification | 5473 | 10 | 5 |
Spectf | 267 | 44 | 2 |
Steel plates faults | 1941 | 27 | 7 |
Ultrasonic flowmeter diagnostics | 361 | 43 | 4 |
Yeast | 2000 | 8 | 10 |
4.1. ELPSO Clustering
The ELPSO, original PSO, and three improved versions with different inertia weight strategies shown in Section 2.3 will be tested here for evaluating the performance of these heuristic algorithms. Based on Abalone, Ecoli, Glass, and Image segmentation datasets, each method runs 30 times independently and total 500 iterations within every time.
According to the methodology used by Izakian and Abraham [16], criterion J is introduced to evaluate the clustering effect. The lower values of J, the better clustering effect could be obtained. Therefore, the experimental data with the minimum final value of criterion J were considered as the optimal result. The average value recorded was to account for the stochastic nature of the algorithms. For a better view of the results, the best values and the average values of J are shown in Figures 1 –4, respectively.
The clustering results of Abalone dataset. (a) The best result in 30 times. (b) Average result in 30 times.
The clustering results of Ecoli dataset. (a) The best result in 30 times. (b) Average result in 30 times.
The clustering results of Glass dataset. (a) The best result in 30 times. (b) Average result in 30 times.
The clustering results of Image segmentation dataset. (a) The best result in 30 times. (b) Average result in 30 times.
Since the inertia weight plays an important role on the overall performance of the algorithm, in order to ensure that the variables are unique, all parameters are set consistently except the inertia weight. The parameter values for each algorithm are set as follows.
The population: all algorithms are set to 30 uniformly; ELPSO: c 1 =c 2 = 2, ω is dynamically adjusted according to the proposed strategy using equation (12), and z is set to 1.05; the parameters in other algorithms are set to be consistent with ELPSO and their inertial weight strategy reference literature [28–30].
The results are shown as follows.
For a better observation of experimental results, we extract the curves of the first 200 iterations separately and place them in the overall iteration graph. In this way, we can perceive the convergence trend of each algorithm explicitly. In addition, the criterion J of the 50th, 200th, and 500th iterations is listed in Tables 2 and 3, respectively; these results represent the optimal value and average value in experiments.
Table 2
Best results for criterion J (the best results are highlighted in bold).
Datasets | Iterations | PSO | LPSO | EPSO | RPSO | ELPSO |
---|---|---|---|---|---|---|
Abalone | 50 | 7610.0432 | 7928.2152 | 7321.0800 | 7456.6338 | 7209.1766 |
200 | 7229.6062 | 7399.8084 | 7197.7556 | 7210.1214 | 7197.7447 | |
500 | 7197.7448 | 7237.2808 | 7197.7447 | 7198.3222 | 7197.7447 | |
Ecoli | 50 | 6.7325 | 6.3524 | 6.4713 | 6.8044 | 5.8502 |
200 | 6.0905 | 5.9078 | 5.8227 | 6.0385 | 5.3947 | |
500 | 5.3443 | 5.6718 | 5.3314 | 5.4742 | 5.3310 | |
Glass | 50 | 240.1781 | 213.2250 | 191.1327 | 220.7926 | 178.3978 |
200 | 176.6745 | 184.5977 | 155.7761 | 168.4756 | 154.1854 | |
500 | 154.5077 | 174.6182 | 154.1481 | 159.4427 | 154.1460 | |
Image segmentation | 50 | 18079261 | 14459116 | 16710175 | 10112452 | 7263604 |
200 | 13038282 | 11080451 | 6606879 | 9258174 | 5940593 | |
500 | 6082020 | 9604572 | 5780101 | 8724927 | 5690913 |
Table 3
Average results for criterion J (best results are highlighted in bold).
Datasets | Iterations | PSO | LPSO | EPSO | RPSO | ELPSO |
---|---|---|---|---|---|---|
Abalone | 50 | 7681.8732 | 7756.7018 | 7375.2852 | 7444.9952 | 7210.9449 |
200 | 7224.8618 | 7380.1867 | 7197.7683 | 7211.3830 | 7197.7447 | |
500 | 7197.7452 | 7297.5891 | 7197.7447 | 7199.7810 | 7197.7447 | |
Ecoli | 50 | 6.9441 | 6.9788 | 6.5282 | 6.6143 | 5.9576 |
200 | 5.8466 | 6.1840 | 5.4144 | 5.7297 | 5.3656 | |
500 | 5.3791 | 5.9953 | 5.3618 | 5.5613 | 5.3537 | |
Glass | 50 | 245.4625 | 240.3651 | 215.4950 | 219.1814 | 184.6535 |
200 | 213.7898 | 220.0770 | 177.9124 | 193.2778 | 162.8663 | |
500 | 156.9724 | 187.9902 | 154.6609 | 168.6809 | 154.1477 | |
Image segmentation | 50 | 17945971 | 17274272 | 15048314 | 16121031 | 8161111 |
200 | 13228454 | 14092457 | 7037767 | 12326004 | 6154794 | |
500 | 6811563 | 12371116 | 6298898 | 10979083 | 6050846 |
Figures 1 –4 show the result for the five approaches represented by five colored curves. In each figure, the horizontal axis represents the number of iterations, and the vertical axis represents criterion J. A smaller value of J indicates better results.
The optimal result in 30 iterations represents the extreme ability of algorithm, but the average result over 30 iterations can better illustrate the performance of algorithm. It is clearly seen from Figures 1 to 4 that ELPSO converges more quickly and has obvious advantage of convergence speed in best graphs and average graphs than other algorithms, especially in the first fifty iterations.
Tables 2 and 3 show that ELPSO always achieves the smallest value for criterion J in 50th, 200th, and 500th iterations, better than other four algorithms regardless of the best values or average values. Although in the Abalone dataset, EPSO finally obtained the same optimal value as ELPSO, but its earlier convergence rate was slower than ELSPO. From the results obtained from the four sets of datasets, LPSO is more likely to fall into a local optimum in five algorithms, and ELPSO has never been trapped in a local optimum due to its appropriate inertia weight selection strategy.
The results of the tests lead to the conclusion that the proposed ELPSO is efficient and has rapid convergence, can counterpoise the global search and local search more effectively, and can reveal very encouraging results in terms of quality of solution found.
4.2. Hybrid Methods Clustering
In this section, the FCM-ELPSO proposed in this work is compared to other four PSO-based hybrid algorithms which are FCM-SPSO, FCM-LPSO, FCM-EPSO, and FCM-RPSO. In addition, GA-FCM is added to the test. For evaluating the performance of all of the above algorithms, eight UCI datasets are selected: Ecoli, Glass, Image segmentation, Page blocks classification, Spectf, Steel plates faults, Ultrasonic flowmeter diagnostics, and Yeast, as shown in Table 1.
To quantitatively evaluate the convergence effect, the fundamental criterion can be described as follows: the distance between different objects in the same cluster should be as close as possible; the distance between different objects in different cluster should be as far as possible. The criterion J is still introduced to evaluate the clustering effect, as the same in Section 4.1. Additionally, an effective cluster validity index is introduced into the evaluation system, namely, PBM(F), which has been described in detail. It is worth reminding again that for a given dataset and the determined number of clusters, higher values of the PBM(F) index indicate better clustering in the sense that the clusters are well separated and relatively compact.
Each algorithm is run 30 times with random initializations for every dataset, and the partition that corresponds to the best criterion value is selected. Once the partition is selected, its corresponding PBM(F) is calculated. Furthermore, the average and standard deviation of the 30 repetitions are also computed for criterion J and validity index PBM(F). The parameters of the PSO part in these five algorithms are the same values as in Section 4.1, and the fuzzy parameter m in the FCM part is set to 2. The results are shown as follows.
Table 4 shows the best objective function values expressed in equation (4) obtained from the five clustering algorithms. For a more careful observation, the average values are provided separately in Table 5. It should be noted that the hybrid methods always converge before reaching the aforementioned maximum number of iterations [16]. Hence, it can be considered that under the same stopping condition, the performance of the algorithms depend on their results.
Table 4
Best results for criterion J (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 5.3561 | 5.3460 | 5.3490 | 5.3540 | 5.3457 | 5.3326 |
Glass | 157.4681 | 155.3780 | 154.4778 | 155.1152 | 154.7951 | 154.1496 |
Image segmentation | 6.0142e + 06 | 5.9676e + 06 | 5.8362e + 06 | 5.8933e + 06 | 5.8689e + 06 | 5.7221e + 06 |
Page blocks classification | 8.5735e + 09 | 8.5614e + 09 | 8.5621e + 09 | 8.5643e + 09 | 8.5616e + 09 | 8.5612e + 09 |
Spectf | 5.8436e + 05 | 5.8049e + 05 | 5.7739e + 05 | 5.7739e + 05 | 5.7739e + 05 | 5.7739e + 05 |
Steel plates faults | 4.3874e + 14 | 4.2944e + 14 | 4.3438e + 14 | 4.3463e + 14 | 4.2936e + 14 | 4.2900e + 14 |
Ultrasonic flowmeter diagnostics | 3.6411e + 08 | 3.6287e + 08 | 3.6312e + 08 | 3.6364e + 08 | 3.6310e + 08 | 3.6276e + 08 |
Yeast | 12.2630 | 12.0382 | 11.8642 | 11.8538 | 11.8746 | 11.8413 |
Table 5
Average results for criterion J (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 5.4132 | 5.4074 | 5.3909 | 5.3943 | 5.3878 | 5.3649 |
Glass | 160.6247 | 158.4809 | 158.7457 | 159.0322 | 159.6960 | 154.9908 |
Image segmentation | 6.1894e + 06 | 6.0791e + 06 | 6.1117e + 06 | 6.1396e + 06 | 6.0897e + 06 | 6.0510e + 06 |
Page blocks classification | 9.9203e + 09 | 9.8668e + 09 | 9.1186e + 09 | 9.3430e + 09 | 9.3317e + 09 | 8.6631e + 09 |
Spectf | 5.8960e + 05 | 5.8260e + 05 | 5.7883e + 05 | 5.7838e + 05 | 5.7849e + 05 | 5.7754e + 05 |
Steel plates faults | 4.5225e + 14 | 4.5225e + 14 | 4.8656e + 14 | 4.7123e + 14 | 4.6599e + 14 | 4.4662e + 14 |
Ultrasonic flowmeter diagnostics | 3.6819e + 08 | 3.6763e + 08 | 3.6795e + 08 | 3.6904e + 08 | 3.6796e + 08 | 3.6387e + 08 |
Yeast | 13.1546 | 12.1272 | 11.9522 | 11.9556 | 11.9585 | 11.8478 |
Tables 4 and 5 show that FCM-ELPSO always achieves the smallest value for criterion J. For further illustrating the performance of these algorithms, we introduce the standard deviation to describe the deviation degree of the mean values. The smaller the standard deviation value is, the smaller the convergence range is and the more robust the algorithm is. Table 4 shows the standard deviation for criterion J.
In Table 6, FCM-ELPSO gets the smallest standard deviation on five datasets, Glass, Page blocks classification, Spectf, Ultrasonic flowmeter diagnostics, and Yeast. FCM-SPSO gets two, Image segmentation and Steel plates faults, and FCM-LPSO has one, Ecoli. It can be seen that ELPSO has a smaller convergence range and higher robustness.
Table 6
Standard deviation for criterion J (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 0.0301 | 0.0297 | 0.0223 | 0.0245 | 0.0273 | 0.0272 |
Glass | 2.8639 | 1.7319 | 2.4125 | 1.8354 | 2.0823 | 1.1018 |
Image segmentation | 1.7584e + 05 | 7.2663e + 04 | 1.7584e + 05 | 1.4599e + 05 | 1.5925e + 05 | 2.7719e + 05 |
Page blocks classification | 1.2927e + 09 | 1.3478e + 09 | 7.8227e + 08 | 1.4926e + 09 | 1.0365e + 09 | 4.6222e + 08 |
Spectf | 1.4432e + 03 | 912.4835 | 1.9985e + 03 | 1.5790e + 03 | 1.6325e + 03 | 261.1258 |
Steel plates faults | 3.6258e + 13 | 1.7151e + 13 | 3.8546e + 13 | 2.4850e + 13 | 3.4818e + 13 | 2.3890e + 13 |
Ultrasonic flowmeter diagnostics | 3.9146e + 06 | 5.1001e + 06 | 3.4344e + 06 | 4.2497e + 06 | 3.7826e + 06 | 1.0618e + 06 |
Yeast | 0.0472 | 0.0443 | 0.0627 | 0.0596 | 0.0705 | 0.0065 |
Tables 7 –9 show the corresponding values for validity index PBM(F).
Table 7
Best results for validity index PBM(F) (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 0.3265 | 0.3274 | 0.3303 | 0.3308 | 0.3327 | 0.3340 |
Glass | 3.0827 | 3.1856 | 3.7173 | 3.2463 | 4.0843 | 3.9848 |
Image segmentation | 564.4213 | 566.3507 | 577.5760 | 573.2499 | 576.2624 | 587.5084 |
Page blocks classification | 8.1091e + 04 | 8.5060e + 04 | 8.5086e + 04 | 7.7729e + 04 | 8.7911e + 04 | 8.3118e + 04 |
Spectf | 26.4198 | 18.8760 | 35.7040 | 35.4333 | 33.3814 | 34.2975 |
Steel plates faults | 1.0651e + 07 | 1.0634e + 07 | 1.0566e + 07 | 1.0686e + 07 | 1.0685e + 07 | 1.0692e + 07 |
Ultrasonic flowmeter diagnostics | 4.0962e + 03 | 4.1664e + 03 | 4.0975e + 03 | 4.0582e + 03 | 4.1587e + 03 | 4.2225e + 03 |
Yeast | 0.1082 | 0.1105 | 0.1395 | 0.1374 | 0.1378 | 0.1526 |
Table 8
Average results for validity index PBM(F) (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 0.3195 | 0.3162 | 0.3174 | 0.3198 | 0.3214 | 0.3238 |
Glass | 2.6173 | 2.7959 | 2.9268 | 2.8091 | 2.9334 | 3.3451 |
Image segmentation | 261.9715 | 255.1723 | 409.8047 | 401.2907 | 361.9401 | 487.0102 |
Page blocks classification | 6.7247e + 04 | 6.8971e + 04 | 6.5584e + 04 | 6.7081e + 04 | 6.5419e + 04 | 7.0711e + 04 |
Spectf | 19.5535 | 12.4250 | 26.0889 | 28.1726 | 27.2532 | 31.5303 |
Steel plates faults | 9.6322e + 02 | 1.0076e + 07 | 9.5767e + 06 | 9.7646e + 06 | 9.7419e + 06 | 1.0186e + 07 |
Ultrasonic flowmeter diagnostics | 3.6912e + 03 | 3.7835e + 03 | 3.7067e + 03 | 3.6883e + 03 | 3.7202e + 03 | 4.0045e + 03 |
Yeast | 0.0871 | 0.0884 | 0.1207 | 0.1177 | 0.1191 | 0.1434 |
Table 9
Standard deviation for validity index PBM(F) (best results are highlighted in bold).
Datasets | GA-FCM | FCM-SPSO | FCM-LPSO | FCM-EPSO | FCM-RPSO | FCM-ELPSO |
---|---|---|---|---|---|---|
Ecoli | 0.0085 | 0.0088 | 0.0060 | 0.0072 | 0.0065 | 0.0053 |
Glass | 0.2473 | 0.1993 | 0.2953 | 0.2056 | 0.3306 | 0.1647 |
Image segmentation | 200.4316 | 196.6843 | 207.5596 | 198.1861 | 216.7850 | 168.7607 |
Page blocks classification | 7.1675e + 03 | 5.8166e + 03 | 6.7870e + 03 | 6.3222e + 03 | 7.4694e + 03 | 2.8312e + 03 |
Spectf | 7.0519 | 2.9708 | 7.7021 | 6.6620 | 6.5179 | 2.1903 |
Steel plates faults | 5.8661e + 05 | 4.2121e + 05 | 6.4348e + 05 | 5.2614e + 05 | 6.2299e + 05 | 3.9786e + 05 |
Ultrasonic flowmeter diagnostics | 306.4910 | 331.9834 | 232.4805 | 245.6173 | 290.7265 | 196.3196 |
Yeast | 0.0133 | 0.0107 | 0.0116 | 0.0129 | 0.0134 | 0.0053 |
FCM-ELPSO gets the maximum on five datasets of best results for validity index PBM(F), as shown in Table 7, and FCM-RPSO performs better in Glass and Page blocks classification, while FCM-LPSO is good at Spectf. In terms of average results and standard deviation, FCM-ELPSO performs better than other algorithms. And it is noticed that the performance of GA-FCM is not as good as the hybrid clustering algorithms based on PSO.
Comparing the results of two cluster validity indexes, it is possible to notice that the best criterion J is not always associated with the best value for the PBM(F) because the cluster validity indexes are not applicable to all datasets. However, the experimental results can still prove that FCM-ELPSO performs better and has better robustness. The hybrid algorithm combines the merits of both algorithms to prevent premature convergence and trapping into local optimum effectively and improves the convergence speed slightly and obtains satisfactory results.
5. Conclusion
This paper proposes ELPSO to better balance between exploration and exploitation, which avoids falling into local optimum and has excellent convergence ability. In order to overcome the shortcomings of the PSO-based fuzzy clustering algorithms, ELPSO and FCM are combined to form a hybrid method called FCM-ELPSO, which utilizes the global search property of ELPSO to produce suitable initial clustering prototypes for FCM. FCM-ELPSO can correct the clustering direction during training constantly. So, as a randomized initialization approach, the hybrid method has the capability to alleviate the problems faced by FCM, which has some demerits of initialization and falling in local minima. The experiments test the ELPSO and the hybrid algorithm separately. Experimental results show that ELPSO and FCM-ELPSO perform well in the UCI datasets. In particular, FCM-ELPSO can produce higher quality clusters with a smaller standard deviation on the selected datasets compared with other clustering methods, especially in the high dimension and large data cases.
For future work, we will explore the practical application of the proposed methods in different fields, such as image segmentation, text mining, and medical problems. Furthermore, we will research novel initialization methods of PSO to improve the performance for complex datasets.
Acknowledgments
This work was supported by the National Key R&D Program of China (2018YFB1308400).
Data Availability
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
1. Pimentel B. A., de Souza R. M. C. R. A multivariate fuzzy c-means method. Applied Soft Computing. 2013;13(4):1592–1607. doi: 10.1016/j.asoc.2012.12.024. [CrossRef] [Google Scholar]
2. Melin P., Castillo O. A review on type-2 fuzzy logic applications in clustering, classification and pattern recognition. Applied Soft Computing. 2014;21:568–577. doi: 10.1016/j.asoc.2014.04.017. [CrossRef] [Google Scholar]
3. Li X., Song J., Zhang F., Ouyang X., Khan S. U. MapReduce-based fast fuzzy c-means algorithm for large-scale underwater image segmentation. Future Generation Computer Systems. 2016;65:90–101.S0167739X16300486 [Google Scholar]
4. Xu R. Survey of Clustering Algorithms. Piscataway, NJ, USA: IEEE Press; 2005. [Google Scholar]
5. Bezdek J. C., Ehrlich R., Full W. FCM: the fuzzy c-means clustering algorithm. Computers & Geosciences. 1984;10(2-3):191–203. doi: 10.1016/0098-3004(84)90020-7. [CrossRef] [Google Scholar]
6. Kesemen O., Tezel Ö., Özkul E. Fuzzy c-means clustering algorithm for directional data (FCM4DD) Expert Systems with Applications. 2016;58:76–82. doi: 10.1016/j.eswa.2016.03.034. [CrossRef] [Google Scholar]
7. Verma H., Agrawal R. K., Sharan A. An improved intuitionistic fuzzy c-means clustering algorithm incorporating local information for brain image segmentation. Applied Soft Computing. 2016;46:543–557. doi: 10.1016/j.asoc.2015.12.022. [CrossRef] [Google Scholar]
8. Olman V., Fenglou Mao F., Hongwei Wu H., Ying Xu Y. Parallel clustering algorithm for large data sets with applications in bioinformatics. IEEE/ACM Transactions on Computational Biology and Bioinformatics. 2009;6(2):344–352. doi: 10.1109/tcbb.2007.70272. [PubMed] [CrossRef] [Google Scholar]
9. Zhang D.-Q., Chen S.-C. Clustering incomplete data using kernel-based fuzzy C-means algorithm. Neural Processing Letters. 2003;18(3):155–162. doi: 10.1023/b:nepl.0000011135.19145.1b. [CrossRef] [Google Scholar]
10. Lin K.-P. A novel evolutionary kernel intuitionistic fuzzy C-means clustering algorithm. IEEE Transactions on Fuzzy Systems. 2014;22(5):1074–1087. doi: 10.1109/tfuzz.2013.2280141. [CrossRef] [Google Scholar]
11. Kuo R. J., Lin T. C., Zulvia F. E., Tsai C. Y. A hybrid metaheuristic and kernel intuitionistic fuzzy c-means algorithm for cluster analysis. Applied Soft Computing. 2018;67:299–308. doi: 10.1016/j.asoc.2018.02.039. [CrossRef] [Google Scholar]
12. Alam S. Research on particle swarm optimization based clustering: a systematic review of literature and techniques. Swarm and Evolutionary Computation. 2014;17:1–13. doi: 10.1016/j.swevo.2014.02.001. [CrossRef] [Google Scholar]
13. Chen S., Xu Z., Tang Y. A hybrid clustering algorithm based on fuzzy c-means and improved particle swarm optimization. Arabian Journal for Science & Engineering. 2014;39(12):8875–8887. doi: 10.1007/s13369-014-1424-9. [CrossRef] [Google Scholar]
14. Lu P. Fuzzy C-means based on cooperative QPSO with learning behavior. Proceedings of the International Conference on Intelligent Science & Big Data Engineering; June 2015; Suzhou, China. Springer International Publishing; [Google Scholar]
15. Cura T. A particle swarm optimization approach to clustering. Expert Systems with Applications. 2012;39(1):1582–1588. doi: 10.1016/j.eswa.2011.07.123. [CrossRef] [Google Scholar]
16. Izakian H., Abraham A. Fuzzy c-means and fuzzy swarm for fuzzy clustering problem. Expert Systems with Applications. 2011;38(3):1835–1838. doi: 10.1016/j.eswa.2010.07.112. [CrossRef] [Google Scholar]
17. Sengupta S., Basak S., Peters R. A. Data clustering using a hybrid of fuzzy C-means and quantum-behaved particle swarm optimization. Proceedings of the 2018 IEEE 8th Annual Computing and Communication Workshop and Conference (CCWC); December 2017; Las Vegas, NV, USA. [Google Scholar]
18. Silva Filho T. M., Pimentel B. A., Souza R. M. C. R., Oliveira A. L. I. Hybrid methods for fuzzy clustering based on fuzzy c-means and improved particle swarm optimization. Expert Systems with Applications. 2015;42(17-18):6315–6328. doi: 10.1016/j.eswa.2015.04.032. [CrossRef] [Google Scholar]
19. Tam N. T., Hai D. T., Son L. H., Vinh L. T. Improving lifetime and network connections of 3D wireless sensor networks based on fuzzy clustering and particle swarm optimization. Wireless Networks. 2018;24(5):1477–1490. doi: 10.1007/s11276-016-1412-y. [CrossRef] [Google Scholar]
20. Chrouta J., Zaafouri A., Jemli M. Modeling and identification of irrigation station using fuzzy c-mean clustering algorithms based on particle swarm optimization. Proceedings of the International Conference on Systems & Control; 2015; Minneapois, MN, USA. IEEE; [CrossRef] [Google Scholar]
21. Xuan T. P., Siarry P., Oulhadj H. Integrating fuzzy entropy clustering with an improved PSO for MRI brain image segmentation. Applied Soft Computing. 2018;65:230–242. doi: 10.1016/j.asoc.2018.01.003.S1568494618300097 [CrossRef] [Google Scholar]
22. Singh N., Choudhary N., Gautam R., Tiwari S. Smart Innovations in Communication and Computational Sciences. Vol. 2. Berlin, Germany: Springer; 2019. Model order reduction using fuzzy C-means clustering and particle swarm optimization. [CrossRef] [Google Scholar]
23. Nayak J., Naik B., Behera H. S., Abraham A. Hybrid chemical reaction based metaheuristic with fuzzy c-means algorithm for optimal cluster analysis. Expert Systems with Applications. 2017;79:282–295. doi: 10.1016/j.eswa.2017.02.037. [CrossRef] [Google Scholar]
24. Nayak J., Naik B., Kanungo D. P., Behera H. S. A hybrid elicit teaching learning based optimization with fuzzy c-means (ETLBO-FCM) algorithm for data clustering. Ain Shams Engineering Journal. 2016;9(3):379–393. doi: 10.1016/j.asej.2016.01.010.S2090447916000289 [CrossRef] [Google Scholar]
25. Kumar R., Dwivedi R., Jangam E. Communications in Computer and Information Science. Berlin, Germany: Springer; 2019. Hybrid fuzzy C-means using bat optimization and maxi-min distance classifier. [CrossRef] [Google Scholar]
26. Peng Y., Peng X. Y., Liu Z. Q. Statistic analysis on parameter efficiency of particle swarm optimization. Acta Electronica Sinica. 2004;32(2):209–213. [Google Scholar]
27. Tian D., Shi Z. MPSO: modified particle swarm optimization and its applications. Swarm and Evolutionary Computation. 2018;41:49–68. doi: 10.1016/j.swevo.2018.01.011.S2210650217307137 [CrossRef] [Google Scholar]
28. Eberhart R. C., Shi Y. Comparing inertia weights and constriction factors in particle swarm optimization. Proceedings of the 2000 IEEE Congress on Evolutionary Computation; July 2000; La Jolla, CA, USA. IEEE; [CrossRef] [Google Scholar]
29. Chen G., Huang X., Jia J., Min Z. Natural exponential inertia weight strategy in particle swarm optimization. Proceedings of the 2006 6th World Congress on Intelligent Control and Automation; June 2006; Dalian, China. [CrossRef] [Google Scholar]
30. Eberhart R. C., Shi Y. Tracking and optimizing dynamic systems with particle swarms. Proceedings of the 2001 Congress on Evolutionary Computation; May 2001; Seoul, Korea. IEEE; [CrossRef] [Google Scholar]
31. Pakhira M. K., Bandyopadhyay S., Maulik U. Validity index for crisp and fuzzy clusters. Pattern Recognition. 2004;37(3):487–501. doi: 10.1016/j.patcog.2003.06.005. [CrossRef] [Google Scholar]
Fuzzy C Mean Find the Global Optimal Solution
Source: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7104327/