International Journal of Fuzzy Systems, Vol. 16, No. 1, March 2014 39 A Hybrid Fuzzy Clustering Method with a Robust Validity Index Horng-Lin Shieh Abstract1 A robust validity index for fuzzy c-means (FCM) algorithm is proposed in this paper. The purpose of fuzzy clustering is to partition a given set of training data into several different clusters that can then be modeled by fuzzy theory. The FCM algorithm has become the most widely used method in fuzzy clustering. Although, there are some successful applications of FCM have been proposed, a disadvantage of FCM is that the number of clusters must be predetermined. After clustering, it is often necessary to evaluate the fitness of the results obtained by FCM. Such assessment techniques are called cluster validity. In this paper, a new cluster validity index is proposed to evaluate the fitness of clusters obtained by FCM and four examples show the results of proposed index have good performances than other cluster validities. Keywords: Clustering algorithm, fuzzy c-means (FCM) algorithm, robust, validity index. 1. Introduction tering partition a given set of sampling data into several different clusters by membership functions. Let X denote the universal set, then the membership function, A , by which fuzzy set A is usually defined as: (1) A : X [0,1] . In fuzzy clustering algorithm, the degree of a data point, x X={x1, x 2,…, xn} d , belonging to cluster A can be denoted by A (x ) . The value of a fuzzy membership function belongs to any number between 0 and 1, and is meant to be a mathematical characterization of a “set” which may not be precisely defined. There are many fuzzy clustering methods proposed in the literature [6-13]. In fuzzy clustering algorithms, the fuzzy c-means algorithm (FCM) [2, 8-10] proposed by Bezdek has become the most popular approach for both theoretical and practical applications in recent decades. Let U= {ik }cn M fcn be a partition matrix where ik is the membership value of xi belonging to class k, and V={ v1, v2,…,vc } is a set of cluster centers. The FCM minimize the following objective function with respect to ik and vk : n c Jm(U,V) = ( ik ) m || xi vk ||2 , In data processing technique, the clustering algorithms are widely used for grouping together similar data into a number of clusters. It attempts to partition unlabeled input vectors into clusters such that data points within a cluster are more similar to each other than those belonging to different clusters. There are two kinds of clustering algorithm: hard and soft clustering [1]. In hard clustering, such as K-nearest neighbors (KNN) [2][3] and k-means [4], each data point is assigned to exactly one cluster, while in soft clustering, such as the fuzzy clustering algorithm, a membership value is assigned representing the degree to which each data point belongs to a cluster. In real applications, sampling data often have uncertain attributes and so cannot be correctly partitioned into one cluster. The fuzzy set proposed by Zadeh [5] is a solution for dealing with this problem. The fuzzy clus- and M fcn U [ ik ] 0 ik 1,i, j; c n ik 1, 0 ik n . k 1 i 1 (3) To optimize (2), FCM algorithm alternates between optimizations of Jm over U with V fixed and Jm over V with U fixed, producing a sequence of { U ( s ) ,V ( s ) }. Specifically, the (s+1)th value of V={ v1 ,v2 ,…,vc } is computed using the (s)th value of U in the right-hand side of (4). n v k i1 n m ik x i . (4) m ik i1 The (s+1) st value of U is obtained by (5): ik Corresponding Author: Horng-Lin Shieh is with the Department of Electrical Engineering, St. John’s University, 499, Sec. 4, Tam King Road, Tamsui District, New Taipei City, Taiwan, 251. E-mail: [email protected] Manuscript received 05 Oct. 2011; revised 10 Oct. 2013; accepted 17 Feb. 2014. (2) k 1 i 1 || xi vk ||1/( m 1) , c (5) 1/( m 1) || xi v j || j 1 where 1<m< is the fuzzification parameter; 2 c n the number of centers, and ||.|| denotes the inner product norm induced on d . © 2014 TFSA International Journal of Fuzzy Systems, Vol. 16, No. 1, March 2014 40 In literature, there were many successful applications of the fuzzy set, such as data mining [14, 15], decision making [16-18], robot control [19], and function approximation [20]. However, the FCM algorithm has one drawback, that is, FCM needs to know the number of clusters, which is not always possible in some real applications. In 1994, Yager and Filev [21] proposed a mountain function to obtain the initial cluster centers which can lead to better clustering results by FCM. In 1994, Chiu [22] modified the mountain method to construct a potential function for calculating the clustering center of the sampling data. The proposed method is also called the subtractive clustering (SC) algorithm. In SC algorithm, the feature points are likened to potential sources. The potential of cluster center has a maximal value at the location of the feature point which decreases rapidly at any point away from the feature point [23]. The potential of each data point xi is defined as: n P ( xi ) exp( d ( xk , xi )) , (6) k 1 where 4 / ra , ra is a positive constant, and the d(xk,xi) is the distance between data points xi and xk. It is reasonable to assume that the peaks of the potential function correspond to cluster centers and that the valleys correspond to the decision boundaries between the clusters [23]. Suppose xk has the highest potential, then xk is selected as the first cluster center, denoted by x1* , and its corresponding potential value is p1* . After the first cluster center is selected, the potential of each data point can be revised by: (7) P( xi ) P( xi ) p1*e d ( xk , xi ) , where 4 / rb , and rb is a positive constant. To find the next cluster center, x2* , the new revised potential is maximized and the effects of this cluster center are again removed. The process is repeated until p*j / p1* , where is a given threshold and p *j is the potential value of x*j . The choice of is an important factor affecting the clustering results: if is too large, then too few data points will be accepted as cluster centers; if is too small, then too many cluster centers will be generated [22]. The problems of SC algorithm is that each cluster center obtained by SC algorithm is located a certain data point, but it is not a precise location of center. In this paper, the SC algorithm is used for identifying the initial cluster centers of FCM, and a novel robust validity index is proposed for FCM algorithm to indicate the fitness of partitions of a data set with noise. The proposed index combines the compactness in each clus- ter and the separation between clusters to obtain the correct number of clusters of the FCM algorithm. 2. Cluster Validity A validity index is a function which assigns to the output of the clustering algorithm a value which is intended to measure the quality of the clustering provided by the output. A validity index for finding an optimal c, denoted c*, which can completely describe the data structure, becomes the most studied topic in cluster validity. The quality a partition obtained by the FCM algorithm is evaluated by how closely the data points are associated to the cluster centers. The membership function indicates the close relationship between data and clusters. If the value of one of the memberships of a particular data point is larger than the others, then that point is identified as being a part of the subset of the data represented by the corresponding cluster centers [24]. If there are c clusters of a data set, then each data point has c memberships which represent the close degree between the data point with the cluster centers. So, it is desirable to summarize the information contained in the memberships by a single number which represents the fitness of the data point as classified by the clustering algorithms. The four validity indexes frequently used are [25]: (a) Partition coefficient (PC): The PC index proposed by Bezdek in 1981 is the first validity index for FCM. The PC index is define as (8): c n PC(c) = 1 ( ik )2 , n k 1 i 1 (8) c and ik 1, k 1 where ik is the membership value of xi belonging to cluster k, and 1/c PC ( c) 1 . The optimal cluster number c* of PC index can be find by solving ( arg max 2 c n 1 PC(c)). But the PC index only considers the fuzzy membership degree ik that indicates the average relative amount of membership sharing done between pairs of fuzzy subsets in U, by combining into a single number, the average contents of pairs of fuzzy algebraic products [26], without considering the data structure of the clusters. So, the optimal value of c* is obtained when ik =1, for a certain cluster k, and ij =0, when j k , i.e. the PC index has maximum value on every hard partition. Therefore, there is a special case where each data point formed one cluster, i.e. ii =1 and ij =0, when j i . A modification of the PC index proposed by Dav´e [27], called MPC, is defined as: MPC (c) = 1 c (1 PC (c)) . c 1 (9) Horng-Lin Shieh: A Hybrid Fuzzy Clustering Method with a Robust Validity Index The value of MPC is in [0, 1]. The MPC is a normalized version of the PC index. When PC(c) =1/c, MPC =0 and PC(c) =1, MPC =1. So, it has the same disadvantage as the PC index. (b) Partition entropy (PE): The PE proposed by Bezdek [28] is defined as: PE ( c ) 1 c n ik ln( ik ) , n k 1 i 1 (10) where 0 PE ( c ) ln(c ) . The optimal cluster number c* of PE index can be find by solving ( arg min2 c n 1 PE(c)). The PE index considers measuring the amount of fuzziness in a given U. The problem of PE index is analog to the PC index, i.e. the PE index will take the minimum value on every hard partition. These two indexes only evaluate the fuzziness of U, but do not consider the data structure of the clusters. (c) Xie and Beni (XB) index: The XB index proposed by Xie and Beni [29] in 1991 is defined as follows: c XB( c ) n 2 2 ik || xi vk || k 1 i 1 n mini j (|| vi v j ||) , (11) where vi and vj represent the centers of cluster i and j, respectively, and ||vi -vj || represents the Euclidean distance between vi and vj. XB tries to solve the arg min2 c n 1 XB(c) to obtain the optimal clustering number c* for data set X. The XB index integrates two properties, compactness and separation, of the data set and clusters. The numerator represents the compactness in each cluster and the denominator indicates the separation between clusters [30]. For obtaining the best performance of partitions, the value of the compactness in one cluster is as small as possible and the separation between the clusters is as high as possible. To avoid the problem that the XB index decreases monotonically when cluster number c is close to the cardinality of a data set, the authors recommend plotting the XB curve as a function of c and then selecting the starting point of the monotonic epoch as the maximum c (cmax) to be considered [31]. (d) Fukuyama and Sugeno (FS) index [32]: The FS index is another index that integrates the properties of compactness and separation of the data set and clusters, and is defined as follows: FS(c) = n c c n m 2 ( ik ) || xi vk || - ( ik ) 2 || vk v || , (12) i 1k 1 k 1 i 1 where 1 m , and v is the mean of the cluster cen- ters. In the right side of (12), the first term represents the compactness of the clusters and the second term indicates the separation between the clusters. A small value of FS index indicates a good fuzzy clustering result with good compactness in clusters and separation between the 41 clusters. The optimal cluster number of a clustering algorithm is obtained by solving arg min2 c n 1 FS(c). The above indexes have the common objective of finding a good estimate of cluster number c so that each one of the c clusters is compact or/and separated from the other clusters. But in real application systems, data domains are often affected by noises, but there has been little discussion in the literature about the influence of noise on validity indexes. In this paper, a novel robust validity index is proposed to indicate the goodness of partitions in a data set with noise. 3. A Novel Robust Cluster Validity for the FCM Algorithm In this paper, a novel validity index is proposed which integrates the properties of compactness and separation to indicate the fitness of the partition obtained by FCM algorithm. Let W wik | 0 wik 1 , 1 i n, 1 k c be a weighting matrix, where wik is a weighting of xi belonging to cluster k, and V={ v1 ,v2 ,…,vc } is a set of the cluster centers, where vk d . The wik is defined by: wik exp( || xi vk ||2 2 2 ), (13) where ||xi-vk || represents the Euclidean distance between xi and vk , and is the width of the Gaussian function. In the proposed method, , where is defined by (6). This paper defines a new compactness measure to indicate the optimal cluster number obtained by FCM. The proposed compactness measure is defined as (14): c n Com(c) = ( 1 || xi vk || wik ) 2 ik n k 1 i 1 (14) c and ik 1 . k 1 where the wik is a weight indicating the importance of distance between the data xi and cluster vk. The difference between wik and ik obtained by FCM algorithm is that there is no limitation to “the sum of wik =1, i, k ”. So, the wik has more representative than ik in reflecting correlation between data and clusters. Especially, the noises do not need to satisfy the limitation. Hence, the influence of noises is reduced by wik . In (14), if all data around cluster k is close to cluster center vk, then the value of the Com(c) is low, which indicates that all clusters are compactness; so, in order to find the best partition of the sampling data, the Com(c) value should be as low as possible. To measure the separation between clusters, this paper defines the separation function as: International Journal of Fuzzy Systems, Vol. 16, No. 1, March 2014 42 SE(c) 1 (15) c c || vi vk || i 1k 1 c(c 1) 2 1 , = SV MV mini j (|| vi v j ||) c c || vi vk || where SV = i 1k 1 , c(c 1) 2 and MV = mini j (|| vi v j ||) . In (15), the denominator indicates the separation between clusters. The first term SV of denominator represents the average distance between clusters. The second term of MV denominator represents the minimal distance between clusters. The SC algorithm adopts (7) to reduce the potential of data x around data xi* , where xi* is the ith selected cluster center. When the data set contains noise that far away from the cluster centers and a new cluster center is added at the noise point, the compactness of (14) is reduced, but the SV will be decreased to prevent a data forming a cluster. When the noise is close to the cluster centers and a new cluster center is located at the noise point, the minimal distance between cluster centers may be reduced. So, the SV and MV are used to prevent the noise becoming new cluster centers. To measure the separation between clusters and the compactness in each cluster, this paper combines Com(c) with SE(c) to find a good estimation of cluster number c* for the data clustering algorithm. This can be defined as: CS(c) = Com(c)×SE(c). (16) The low CS(c) value means that each of the c clusters is compact in each cluster and separated from the other clusters. In order to find the optimal cluster number c*, this paper tries to find the arg min2 c n 1 CS(c) to produce the optimal compactness and separation properties for the clusters generated by the SC algorithm. The above mentioned procedure can be concluded with the following algorithm: Step 1: Set the initial values of , , and cmax and max iteration. Step 2: Calculate the potential P(xi) of each data xi, 1 i n. Step 3: Set the initial number of c=2. Step 4: Evolve the maximal potential value pc* from the P(xi). According to pc* , find the data point xc, that is, vc = xc and xc’s potential value is pc* . Step 5: Update the potential according to (7). Step 6: Set V={ v1,v2 ,…,vc } as the initial cluster centers for FCM. Step 7: For k = 1 to max iteration Step 8: Calculate the k and v k by (5) and (4). Step 9: Next k Step 10: Calculate the CS(c) value by (13) ~ (16). Step 11: c=c+1, if c<=cmax, then go to Step 4. Step 12: Search the minimal value of CS(c). The related value c of the minimal value of CS(c) is the optimal number of clusters. 4. Experimental Results In this section, four experiments are illustrated and their results show that the validity index proposed in this paper outperforms other indexes proposed in the literature. The values of , are set by the definition of the potential function. Example 1: In Fig. 1(a), there is a data set made up of three clusters with noises/outliers. This data set is generated by four cluster centers at (x, y) = {(2, 2), (7, 7), (13, 13)} with Gaussian noise N(0, 0.2 ), and twelve noises/outliers denoted by circles are added. Intuitively, c = 3 is suitable for the data set. As shown in Fig. 1(b), the PE index indicates that the optimal cluster number is three and in Figs. 1(c) and (d), the XB and FS indexes indicate that optimal cluster numbers are two and four for this data set, respectively. In Fig. 1(e), the CS index indicates that the optimal cluster number is three for this data set. The results show that the PE and the CS index proposed by this paper obtained the correct cluster number of this data set. Example 2: In Fig. 2(a), there is a data set made up of four clusters with noises. This data set is generated by four cluster centers at (x, y) = {(1, 1), (7, 7), (13, 13), (19, 19)} with Gaussian noise N(0, 0.8 ), and twenty noises denoted by circles are added. Intuitively, c = 4 is suitable for the data set. As shown in Fig. 2(b), the PE index indicates that the optimal cluster number is four and in Figs. 2(c) and 2(d), the XB and FS indexes indicate that optimal cluster numbers are two and eight for this data set, respectively. In Fig. 2(e), the CS index indicates that the optimal cluster number is four for this data set. The results show that the PE and the CS index proposed by this paper obtained the correct cluster number of this data set. Horng-Lin Shieh: A Hybrid Fuzzy Clustering Method with a Robust Validity Index 25 1 20 0.9 0.8 15 0.7 PE y 10 5 0.6 0.5 0.4 0 0.3 -5 -10 -10 0.2 -5 0 5 10 15 20 0.1 25 2 4 6 Cluster number x (a) 8 10 (b) 3.5 -3000 43 Example 3: In this example, shown as Fig. 3(a), the data set was extended to eight clusters to test the performances of the validity indexes. As shown in Fig. 3(e) and 3(d), the CS and FS indexes correctly acquired the optimal number of clusters. The optimal number of cluster centers obtained by PE is two, as shown in Fig 3(b). In Fig. 3(c), the result of the XB index indicates that the optimal cluster number is four, but, in fact, the number eight is a secondary optimal choice of XB index. 3 -3500 2.5 40 0.7 -4000 30 FS XB 2 1.5 -4500 0.6 20 1 0.55 -5000 2 4 6 Cluster number 8 -5500 10 2 4 (c) 6 Cluster number 8 10 10 PE y 0.5 0 0.65 0.5 0.45 0 0.4 (d) -10 0.35 1800 -20 -20 1600 -10 0 1400 10 x 20 30 40 2 4 6 8 Cluster number (a) 10 12 10 12 (b) CS 1200 1000 4 2.5 800 0 x 10 -0.5 2 600 4 6 Cluster number 8 (e) 1 -4 0.9 0 0.8 20 PE 10 2 4 6 8 Cluster number 12 -4.5 2 4 6 8 Cluster number (d) 0.6 0.5 0 10 (c) 0.7 y -3.5 0.5 1 30 -2 -2.5 -3 Figure 1. A data set contains three clusters. 40 -1 -1.5 1.5 10 FS 2 XB 400 x 10 3 4 0.4 -10 -10 0 10 x 20 30 0.2 40 2 4 6 8 Cluster number (a) 10 (b) 7 -0.6 x 10 4 0.5 -0.8 5 -0.9 FS XB 3 -1.1 -1.3 -1.4 1 2 4 6 8 Cluster number 10 12 (e) Figure 3. A data set contains eight clusters with noises. -1.2 2 -1.5 2 4 6 8 Cluster number 10 12 -1.6 2 4 6 8 Cluster number (c) 10 (d) 8000 7000 6000 CS 0 0 -1 4 1.5 1 -0.7 6 2 12 CS -20 -20 2.5 0.3 5000 4000 3000 2000 2 4 6 8 Cluster number 10 12 (e) Figure 2. A data set contains four clusters with noises. 12 Example 4: As shown in Fig. 4(a), this example used a data set analogous to the data set proposed by Wu et al [33] to identify the above-mentioned indexes. In Fig. 4(e), the result indicated that the CS index proposed in this paper correctly acquired the optimal number of clusters. The results, shown in Fig. 4(c), also show that XB obtained the correct optimal cluster number, but PE and FS, shown in Fig. 4(b) and 4(d), indicated the optimal cluster number as 2 and 15, respectively. Fig 4(f) shows the sixteen cluster centers obtained by the fuzzy c-means algorithm. International Journal of Fuzzy Systems, Vol. 16, No. 1, March 2014 44 As shown by the above examples, the proposed CS index always outperformed the PE, XB and FS indexes. The PE index has good performance when the data set contains small amount clusters. When the clusters are increased, the PE cannot obtain correct cluster number for various data sets. The performances obtained by the CS index were better than the XB index in Examples 1, 2 and 3, and the performance obtained by the CS were also better than the FS index in Examples 1, 2 and 4. 40 1.3 since the PE index only considers the fuzzy membership degree of data belonging to each cluster without considering the structure of the clusters, it elicits a lower level of performance than proposed index, especially when the number of cluster is large. Therefore, the proposed validity index shows the greatest capability to estimate the optimal cluster number for various data sets. The results of examples have proven that the validity index proposed in this paper has better performance than the PE, XB and FS indexes. 1.2 30 Acknowledgment 1.1 20 1 0.9 PE y 10 0 This paper was supported by the National Science Council under contract number NSC 102-2622-E-129001-CC3. 0.8 0.7 -10 0.6 0.5 -20 -20 -10 0 10 x 20 30 0.4 40 2 4 6 8 (a) 10 12 Cluster number 14 16 18 20 References (b) 1.4 3 1.2 2 1 1 x 10 4 [1] 0 FS XB 0.8 0.6 -1 -2 0.4 -3 0.2 0 -4 2 4 6 8 10 12 Cluster number 14 16 18 -5 20 2 4 6 8 (c) 10 x 10 10 12 Cluster number 14 16 18 20 (d) 4 40 30 8 20 CS y 6 4 10 0 -10 2 -20 0 2 4 6 8 10 12 Cluster number 14 16 18 20 -20 -10 (e) 0 10 x 20 30 40 (f) Figure 4. A data set contains sixteen clusters. 5. Conclusion In this paper, a novel validity index for the FCM algorithm is proposed for evaluating the fitness of the resultant cluster number to a data set with noise. Firstly, the SC algorithm is adopted to generate the initial cluster centers for FCM algorithm, and then the FCM algorithm is adopted to reconstruct the cluster centers using the resultant centers obtained by SC algorithm as the initial cluster centers. A robust validity index is proposed in this paper for evaluating the fitness of clustering to data sets. The core idea in the proposed validity index was the combination of the compactness measure of each cluster and the separation property between clusters. From the experiments, both of the XB and FS indexes are affected by noises such that the optimal cluster number cannot be reached in three of the four simulation examples. Also, V. S. Tsen and C.-P. Kao, “A novel similarity-based fuzzy clustering algorithm by integrating PCM and mountain method,” IEEE Trans. on Fuzzy Systems, vol. 15, no. 6, pp. 1188-1196, 2007. [2] B. Bhattacharya and D. Kaller, “Reference set thinning for the k-nearest neighbor decision rule,” In Proc. of the Fourteenth International Conference on Pattern Recognition, vol. 1, pp. 238-242, 1998. [3] J. M. Keller, M. R. Gray, and J. A. Givens, “A fuzzy k-nearest neighbors algorithm,” IEEE Trans. on System, Man, and Cybernetics, vol. SMC-15, pp. 580-585,1985. [4] J. A. Hartigan and M. A. Wong, “A k-means clustering algorithm,” Applied Statistics, vol. 28, no. 1, pp. 100-108, 1979. [5] L. A. Zadeh, “Fuzzy sets,” Information Control, vol. 8, no. 3, pp. 338-353, 1965. [6] J. C. Bezdek, “A convergence theorem for the fuzzy ISODATA clustering algorithms,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. PAMI-2, no. 1, pp. 1-8, 1980. [7] X. Li, H.-S. Wong, and S. Wu, “A fuzzy minimax clustering model and its applications,” Information Sciences, vol. 186, no. 1, pp. 114-125, 2012. [8] E. Nadernejad and A. Barari, “A novel pixon-based image segmentation process using fuzzy filtering and fuzzy c-mean algorithm,” International Journal of Fuzzy Systems, vol. 13, no. 4, pp. 350-357, 2011. [9] P. Kaur, A. K. Soni, and A. Gosain, “A robust kernelized intuitionistic fuzzy c-means clustering algorithm in segmentation of noisy medical images,” Pattern Recognition Letters, vol. 34, no. 2, pp. 163-175, 2013. [10] S. Miyamoto, “Different objective functions in Horng-Lin Shieh: A Hybrid Fuzzy Clustering Method with a Robust Validity Index [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] fuzzy c-means algorithms and kernel-based clustering,” International Journal of Fuzzy Systems, vol. 13, no. 2, pp. 89-97, 2011. M. Sabzekar and M. Naghibzadeh, “Fuzzy c-means improvement using relaxed constraints support vector machines,” Applied Soft Computing, vol. 13, no. 2, pp. 881-890, 2013. A. Skabar and K. Abdalgader, “Clustering sentence-level text using a novel fuzzy relational clustering algorithm,” IEEE Trans. on Knowledge and Data Engineering, vol. 25, no. 1, pp. 62-75, 2013. Y. Hu, D. Wu, and A. Nucci, “Fuzzy-clusteringbased decision tree approach for large population speaker identification,” IEEE Trans. on Audio, Speech, and Language Processing, vol. 21, no. 4, pp. 762-774, 2013. C. Hu, N. Luo, X. Yan, and W. Shi, “Traffic flow data mining and evaluation based on fuzzy clustering techniques,” International Journal of Fuzzy Systems, vol. 13, no. 4, pp. 344-349, 2011. T.-P. Hong, G.-C. Lan, Y.-H. Lin, and S.-T. Pan, “An effective gradual data-reduction strategy for fuzzy itemset mining,” International Journal of Fuzzy Systems, vol. 15, no. 2, pp. 170-181, 2013. C.-T. Chen, P.-F. Pai, and W.-Z. Hung, “A new decision-making process for selecting project leader based on social network and knowledge map,” International Journal of Fuzzy Systems, vol. 15, no. 1, pp. 36-46, 2013. D. Yu, “Multi-criteria decision making based on generalized prioritized aggregation operators under intuitionistic fuzzy environment,” International Journal of Fuzzy Systems, vol. 15, no. 1, pp. 47-54, 2013. I.-J. Ding, “Fuzzy rule-based system for decision making support of hybrid SVM-GMM acoustic event detection,” International Journal of Fuzzy Systems, vol. 14, no. 1, pp. 118-130, 2012. Y.-J. Mon and C.-M. Lin, “Supervisory fuzzy Gaussian neural network design for mobile robot path control,” International Journal of Fuzzy Systems, vol. 15, no. 2, pp. 142-148, 2013. H. L. Shieh, Y.-K. Yang, P.-L. Chang, and J.-T. Jeng, “Robust neural-fuzzy method for function approximation,” Expert Systems with Applications, vol. 36, no. 3, pp. 6903-6913, 2009. R. R. Yager and D. P. Filev, “Approximate clustering via the mountain method,” IEEE Trans. on Systems, Man and Cybernetics, vol. 24, no. 8, pp. 1279-1284, 1994. S. Chiu, “Fuzzy model identification based on cluster estimation,” Journal of Intelligent and Fuzzy Systems, vol. 2, no. 3, pp. 267-278, 1994. R. N. Dav’e and R. Krishnapuram, “Robust clus- [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] 45 tering methods: a unified view,” IEEE Trans. on Fuzzy Systems, vol. 5, no. 2, pp. 270-293, 1997. M. P. Windham, “Cluster validity for the fuzzy c-means clustering algorithm,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. PAMI-4, no. 4, pp. 357-363, 1982. K. L. Wu and M. S. Yang, “A cluster validity index for fuzzy clustering,” Pattern Recognition Letters, vol. 26, no. 9, pp. 1275-1291, 2005. W. Wang and Y. Zhang, “On fuzzy cluster validity indices,” Fuzzy Sets and Systems, vol. 158, no. 19, pp. 2095-2117, 2007. R. N. Dav’e, “Validating fuzzy partition obtained through c-shells clustering,” Pattern Recognition Letter, vol. 17, no. 6, pp. 613-623, 1996. J. C. Bezdek, “Cluster validity with fuzzy sets,” Journal of Cybernetics. vol. 3, no. 3, pp. 58-73, 1974. X. L. Xie and G. Beni, “A validity measure for fuzzy clustering,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 13, no. 8, pp. 841-847, 1991. D.-W. Kim, K. H. Lee, and D. Lee, “On cluster validity index for estimation of the optimal number of fuzzy clusters,” Pattern Recognition, vol. 37, no. 10, pp. 2009-2025, 2004. N. R. Pal and J. C. Bezdek, “On cluster validity for the fuzzy c-means model,” IEEE Trans. on Fuzzy Systems, vol. 3, no. 3, pp. 370-379, 1995. Y. Fukuyama and M. Sugeno, “A new method of choosing the number of clusters for the fuzzy c-means method,” in Proc. of the 5th Fuzzy Systems Symposium, Japan, 1989, pp. 247-250. K. L. Wu, M. S. Yang, and J.-N. Hsieh, “Robust cluster validity indexes,” Pattern Recognition, vol. 42, no. 11, pp. 2541-2550, 2009. Horng-Lin Shieh was born in Chang Hua, Taiwan, September 20, 1965. He received his B.S., M.S., and Ph.D. degrees all in electrical engineering from National Taiwan University of Science and Technology in 1991, 1993 and 2006, respectively. He is currently a Professor in St. John’s University in Taiwan. His research interests include fuzzy clustering, intelligent systems, neural networks and RFID applications.
© Copyright 2024 ExpyDoc