A Survey Of Fuzzy Clustering - Old Dominion University

1y ago
18 Views
2 Downloads
1.29 MB
16 Pages
Last View : 13d ago
Last Download : 3m ago
Upload by : Genevieve Webb
Transcription

Mathl. Comput. Modelling Vol. 18, No. 11, pp. 1-16, 1993Printed in Great Britain. All rights reservedCopyright@0895-7177/93 6.00 0.001993 Pergamon Press LtdA Survey of Fuzzy ClusteringM.-S.YANGDepartment of MathematicsChung Yuan Christian UniversityChungli, Taiwan 32023, R.O.C.(Receivedand acceptedOctober1993)Abstract-Thispaper is a survey of fuzzy set theory applied in cluster analysis. These fuzzyclustering algorithms have been widely studied and applied in a variety of substantive areas. Theyalso become the major techniques in cluster analysis. In this paper, we give a survey of fuzzy clusteringin three categories. The first category is the fuzzy clustering based on fuzzy relation. The second oneis the fuzzy clustering based on objective function. Finally, we give an overview of a nonparametricclassifier. That is the fuzzy generalized k-nearest neighbor Yc-partitions,FUZZYrelation,FUZZYc-means,Fuzzy generalizedk-nearestneighborrule, Clustervalidity.1. INTRODUCTIONCluster analysis is one of the major techniques in pattern recognition. It is an approach tounsupervised learning. The importance of clustering in various areas such as taxonomy, medicine,geology, business, engineering systems and image processing, etc., is well documented in theliterature. See, for example, [l-3]. The conventional (hard) clustering methods restrict that eachpoint of the data set belongs to exactly one cluster. Fuzzy set theory proposed by Zadeh [4] in1965 gave an idea of uncertainty of belonging which was described by a membership function.The use of fuzzy sets provides imprecise class membership information. Applications of fuzzy settheory in cluster analysis were early proposed in the work of Bellman, Kalaba and Zadeh [5] andRuspini [S]. These papers open the door of research in fuzzy clustering. Now fuzzy clustering hasbeen widely studied and applied in a variety of substantive areas. See, for example, [7,8]. Thesemethods become the important tools to cluster analysis.The purpose of this paper is briefly to survey the fuzzy set theory applied in cluster analysis.We will focus on three categories: fuzzy clustering based on fuzzy relation, fuzzy clustering basedon objective functions, and the fuzzy generalized k-nearest neighbor rule-one of the powerfulnonparametric classifiers. We mention that this paper is a nonexhaustive survey of fuzzy settheory applied in cluster analysis. We may have overlooked the papers which might be importantin fuzzy clustering.In Section 2, we give a brief survey of fuzzy c-partitions, (fuzzy) similarity relation, and fuzzyclustering based on fuzzy relation. Section 3 covers a big part of fuzzy clustering based on theobjective functions. In the literature of fuzzy clustering, the fuzzy c-means (FCM) clusteringalgorithms defined by Dunn [9] and generated by Bezdek [7] are the well-known and powerfulmethods in cluster analysis. Several variations and generalizations of the FCM shall be discussedin this section. Finally, Section 4 contains an overview of a famous nonparametric classifierk-nearest neighbor rule in the fuzzy version.Typesetby AM-‘&X

M.-S.2FUZZY2.CLUSTERINGBASED ON FUZZYLetX be a subset of an s-dimensional1) . 11and let c be a e Ws with its ordinary EuclideannormA partitionof X into c clusters can beone.disjoint sets Bi, . . . , II, such that Bi U . . . U B, X or equivalentlyby 1, . . . , pe such that pi(z) 1 if z E Bi and pi(z) 0 if z ! Bi for allby the indicatorz E X and all i l,. ,c. In this case, X is said to have a hard c-partitionfunctions p (pi,. . . ,pc).The fuzzy set, first proposed by Zadeh [4] in 1965, is an extension to allow pi(x) to be a function(called a membershipfunction) assuming values in the interval [OJ]. Ruspini [6] introducedafuzzy c-partitionp ( 1, . . . , ,uc) by the extension to allow pi(z) to be functions assuming valuesin the interval (O,l] such that PI(Z) . . . pC(x) 1 since he first applied the fuzzy set in clusterrepresentedby mutuallythe indicatoranalysis.A (hard)functionrelationT in X is definedare said to have relationto be a functionif r(z, y) 1. A (hard)relationT : X x X - (0, 1) where x and y in Xr in X is called an equivalencerelationif and only if for all z, y E X,(1) R(z,z) 1 (reflexivity),(2) r(z, Y) T(Y, z) (symmetry),and(3) T(Z,Z) 7 /-z) 1 for some z E X T(z, y) 1 ( transitivity).Zadeh [4] defined a fuzzy relation T in X by an extension to allow the values of T in the interval[O,l], where (2, y) denotes the strength of the relation between z and y. In Zadeh [lo], he defineda similarity relation S in X if and only if S is a fuzzy relation and for all z, y, z E X, S(z, z) 1(reflexivity),S(z, y) S(y, z) (symmetry),and S(z, y) 2 VzEx(S(qz) A S(y, z)) (transitivity)where V and A denote max and min. It is obvious that the similarity relation S is essentiallyageneralizationof the equivalence relation.Let us consider a finite data set X (51,. . . , z,} of W”. Denote pi(zj) by / , i 1,. . . ,c,j l,. , n, and r(zj, zle) by rjrc, j, k 1,. . . , n. Let V,, be the usual vector space of realc x n matrices and let uij be the ijth element of U E V,,,. Following Bezdek [7], we define thefollowing:Then MC is exactly a (nondegenerate)hard c-partitionsspace for the finite data set X. DefineA 5 B if and only if aij bij V’i,j where A [aij] and B [bij] E V,,. Define RoR [rij] E V,,with rij VE , ( ik A Tkj)o LetR, {R E Vn, 1Tij E {O,l} Vi,j;Then R, is the set of all equivalence relationsmatrix R [rjk] in V,, be defined byTjk 1,{ 0,I 5 R;in X.R RT;R RoR).For any U [pgj] E n/i,, let the relationif bij /& 1 for some i,otherwise.Then R is obviously an equivalencerelation correspondingto the hard c-partitionsU, since itsatisfies reflexivity, symmetry,and transitivity.That is, for any U E MC there is a relationmatrix R in R, such that R is an equivalence relation correspondingto U.For a fuzzy extension of MC and R,, let

A Surveyof Fuzzy Clusteringand1 I R;Rfn {R E V,, ( rij E [0, l] Vi, j;R RTandR RoR).Then Mfc is a (nondegenerate) fuzzy c-partitions space for X, and Rfn is the set of all similarityrelations in X. Note that Bezdek and Harris [ll] showed that M, c MC, c MfC and that Mfc isthe convex hull of MC, where MC, is the set of matrices obtained by relaxing the last conditionof MC to Cj” i pij 0 Vi.Above we have described the so-called hard c-partitions and fuzzy c-partitions, also the equivalence relation and the similarity relation. These are main representations in cluster analysissince cluster analysis is just to partition a data set into c clusters for which there is always anequivalence relation correspondence to the partition. Next, we shall focus on the fuzzy clusteringmethods based on the fuzzy relation. These can be found in (10,12-161.Fuzzy clustering, based on fuzzy relations, was first proposed by Tamura et al. [12]. Theyproposed an N-step procedure by using the composition of fuzzy relations beginning with areflexive and symmetric fuzzy relation R in X. They showed that there is an n such thatI R 5 R2 5 . . . Rn Rn l . . . R” when X is a finite data set. Then R” is usedto define an equivalence relation Rx by the rule Rx(z, y) 1 wX 5 R”(z, y). In fact, R”is a similarity relation. Consequently, we can partition the data set into some clusters by theequivalence relation Rx. For 0 & 2 . . X2 5 X1 5 1, one can obtain a corresponding k-levelhierarchy of clusters, Di {equivalence clusters of Rx, in X}, i 1,. . . , k where Di refines Djfor i j. This hierarchy of clusters is just a single linkage hierarchy which is a well-known hardclustering method. Dunn [14] proposed a method of computing R” based on Prim’s minimalspanning tree algorithm. Here, we give a simple example to explain it.EXAMPLE.Let1I I0.80.3o’4R ’0.6010’1Therefore,x 0.29 {1,2,3,4},X 0.59 {1,2,3)u {4},x 0.79 *{1,3} u (2) u {4},x 0.81 (1) u (2) u (3) u (4).3.FUZZYCLUSTERINGBASEDON OBJECTIVEFUNCTIONSIn Section 2, we introduced fuzzy clustering based on fuzzy relation. This is one type of fuzzyclustering. But these methods are eventually the novel methods and nothing more than thesingle linkage hierarchy of (hard) agglomerative hierarchical clustering methods. Nothing comesout fuzzy there. This is why there is not so much research and very few results in this type offuzzy clustering. If we add the concept of fuzzy c-partitions in clustering, the situation shall betotally changed. Fuzzy charateristics could be represented by these fuzzy c-partitions. In thissection, we give a survey of this kind of fuzzy clustering.Let a distance (or dissimilarity) function d be given, where d is defined on a finite data setx {Xi,. ,Z,} byd:XxX IW with d(z:, z) 0 and d(z, y) d(y, cc) for all 2, y E X. Let

M.-S.4YANGdij d (xi, Zj) and D [dij] E Vn,. Ruspini [6] first proposed the fuzzy clustering baaed on anobjective function of J&U) to produce an optimal fuzzy c-partitionU in Mfcn, wherewith d E R. He derivedthe necessaryconditionsfor a minimizerU* of JR(U).Ruspini[17,18]continuedmaking more numerical experiments.Since JR is quite complicatedand difficult tointerpret,the Ruspini method was not so successful but he actually opened the door for furtherresearch, especially since he first put the idea of fuzzy c-partitionsin cluster analysis. Notice thatthe only given data in Ruspini’sobjectivefunctionJR(U) is the dissimilaritymatrixD, but notthe data set X even though the element of D could be obtained from the data set X.Based on the given dissimilaritymatrix D, Roubens [19] proposed the objective functionof which the qualityindex of the clusteringis the weightedsum of the total dissimilarity,JROwhereEven though Roubens’ objective function JRO is much more intuitive and simple than Ruspini’sJR, the proposed numerical algorithm to optimize JRO is unstable. Libert and Roubens (201 gavea modificationof it, but unless the substructureof D is quite distinct, useful results may not beobtained.To overcometheseproblemsin JRO, WindhamJAPnJAP(U,B)n[21] proposeda modifiedobjectivefunctionc (P;jb;k)d.jk,j lk li lwhere B [bik] E V,, with bik E [0, l] and CL 1 bik 1. Such bik’s are called prototype weights.Windham[21] presented the necessary conditionsfor minimizersU* and B* of J p(u, B) andalso discussed numerical convergence, its convergence rate, and computer storage. The Windhamprocedure is called the assignment-prototype(AP) algorithm.Hathaway et al. [22] and Bezdek et al. (231 gave a new type of objective function JRFCM, calledthe relational fuzzy c-mean (RFCM), wherecJRFCM( )(c,“,, c; ,P Jzdjk) ci l2 (CI" lPid"with m E [l, co). They gave an iterative algorithm by using a variation of the coordinatedecentmethod described in [24]. The objective function JRFCM seems to be hardly interpreted.Infact, it is just a simple variation of the well-known fuzzy clustering algorithms,called the fuzzyc-means (FCM). The FCM becomes the major part of fuzzy clustering.We will give a big part ofsurvey next. The numerical comparison between AP and RFCM algorithms was reported in [25].The dissimilaritydata matrix D is usually obtained indirectly and with difficultly. In general,the only given data may be the finite data set X {xl,. . . , z,}.For a given data set X zn},Dunn [9] first gave a fuzzy generalizationof the conventional(hard) c-means (more{Xl,.,popularly known as k-means) by combining the idea of Ruspini’s fuzzy c-partitions.He gave theobjective function JD (U, a), where

A Survey of Fuzzy Clustering5with U E Mfcn and a (al,. . . , a,) E (W”)” called the cluster centers. Bezdek [7] generalizedJD(U, a) to JFCM (U, a) withfor which m E (1, co) represents the degree of fuzziness. The necessary conditions for a minimizer(U’, a*) ofJFCM( ,a)areThe iterative algorithms for computing minimizers of Jpc (tY, 3) with these necessary conditionsare called the fuzzy c-means (FCM) clustering algorithms. These FCM clustering proceduresdefined by Dunn [9] and generalized by Bezdek [7] have been widely studied and applied in avariety of substantive areas. There are so many papers in the literature that concern themselveswith some aspects of the theory and applications of FCM. We give some references by the followingcategories. Interested readers may refer some of these:(4 Numerical Theorems 126-37)(b) Stochastic Theorems [34,38-411Cc)Methodologies [9,37,42-511(4 Applications(1)(2)(3)(4)Image Processing [45,52-581Engineering Systems [59-631Parameter estimation [38,64-681Others [69-74).Here we are interested in describing a variety of generalizations of FCM. These include sixgeneralizations and two variations.(A) GENERALIZATION1To add the effect of different cluster shapes, Gustafson and Kessel [47] generalized the FCMobjective function with fuzzy covariance matrices, wherej lwithU E Mfcn,a (al,. . . ,a,)i lE (R”)” and A (Al,. . . ,A,)definite s x s matrices with det(Ai) pi being fixed and ]]zj - ail]‘,for which Ai are positive (zj - oi)T Ai (zj - ai).The iterative algorithms use necessary conditions for a minimizer (U,‘a, A) of JFCM (U, a, A) asfollows:ai Pij Ai (pi det(Si))““SZ:i,where Si Cj” , jog (zj - ui) (zj - ai)‘.i l,.,c,j l,.,n.

M.-S.6YANG(B) GENERALIZATION 2JFCM (U, a) is well used to cluster the data set which has the hyper-sphericalshapes. Gustafsonand Kessel [4’7]generalized it to JFCM,, (U, 3, A) for improving the ability to detect different clustershapes (especially in different hyper-ellipsoidalshapes) in the given data set. Another attemptto improve the ability of JFCM(U, a) to detect nonhyper-ellipsoidallyshaped substructureswasproposed by Bezdek et al. [42,43], called the fuzzy c-varieties (FCV) clustering algorithms.Let us denotespannedthe linearvarietyof dimensionr(0 2 T s) in W throughthe pointa and{br, . . . , b,} byby the vectorsb,) Vr(%bl,.,yEW,y a &,&ElR.k l11If r 0, Vo {u} is a point in W; if T 1, VI L(a; b) {y E IIg’ 1y a tb, t E R} is a linein W; if T 2, Vz is a plane in W; if r 3, V, is a hyperplanein W; and V, W8. In general,the orthogonaldistance (in the A-norm on Ws for which A is a positive definite s x s matrix)from z E RS to V,, where the {bi} is an orthonormalbasis for VT, is112llX-ul 2,- ((x-a,bk)A)2DA(x,Vr) {d (x,y)} 7k lwhere (z, y)A xTAy,dA(Z,g) /Ix-Y\lA l]x]]: (&x)Aand(( - )* y))“ .functionand bk (blk,. . . , b&), k 1,. . , , T. Consider the objectiveerrors between theJFCV( ,a, by , . . . , b,) which is the total weighted sum of square orthogonallinear varieties, wheredata set X ( 1,. . . , xn} and c r-dimensionalLet a (al,., a,)JFCV(u, a, blb,) ,.)withf--j lr;Dfi,j l i rD, DA(“j , v,i) ,yE ‘]y oi tkb&,tkEll%j l,.k lThe FCV clusteringJFCV &Sfollows:algorithmsm Eare iterationswhere sik is the kth unit eigenvectorl,.,C,i Tthroughthe necessaryof the fuzzy scattermatrix,n,conditionsSi with71Si Al”c(correspondingto its k th largest2Pijp;(q- ai) (xj - ai)TA”2j leigenvalue,(Dij)2/ 1)andi l,.,c,j k i (Dkj)2’(m-1)l,.,n,k l,.,r.and[l, oo).andfor minimizersof

A Surveyof Fuzzy Clustering7(C) GENERALIZATION3Consider any probability distribution function G on W”. Let X (21,. . . , z,}be a randomsample of six n from G and let G, be the empirical distribution which puts mass l/n at eachand every point in the sample { 21, . . . , z,}. Then the FCM clustering procedures are to choosefuzzy c-partitions 1 ( 1,. . . , pc) and cluster center a to minimize the objective function GFCM('&, ) dh)Let us replace G, by G. Then we generalized the FCM with respect to G by the minimizationof JGFCM(G,cl,a),JGFCM(G,P, ) 2s ,zmb -ai(12G&).i lConsider a particular fuzzy c-partition of a in (lR”)c of the formIlz2dz,a) ij r 115-ui(121(m-l) -lflj1//2’(m-1)i l,.,c.’Define (a) (/JI(z, a), . . . , pc(z, 3)) and define that&(G,a) JGFCM(G,CL(;L), )Then, for any fuzzy c-partition p (PI(Z), . . . , p,.(z)) and a E (R ) ,JGFCA&,P( L), )5 GFCM(G,I ,; ).The reason for the last inequality is thatwhich can be seen from the simple fact thatci l for pk 0, Yk 0, k 1,. . . , C, plgradient of the Lagrangianc yyJ4j l l/Cm-l)Yj*.-772)Yi I 2PT Yii l p, 1. To claim the simple fact, we may take theand also check that the Hessian matrix of h with respect to pl, . . . ,p, will be positive definite.

M.-S.8YANGLet (/A*,a’) be a minimizer of JGFCM(G, p, a) and let b* be a minimizerJGFCM(G,p*,a*) LJGFCMof L,(G,a). Then,(G (h*),b*) L, (G,b*)I Lx (G,a*)JGFCM(G,p(a*),a*) JGFCM(G,P*, *).Thatis, JGFCM (G,p*, a*) L, (G,h*).problemTherefore,we have thatof JGFCM (G, p, a) over p and 3 is equivalentto solvingsolvingthe minimizationthe minimizationproblemof L, (G, a) over a with the specified fuzzy c-partitions (a). Based on this reduced objectivefunction L, (G, a), Yang et aZ. [39-411 created the existence and stochastic convergence propertiesThey derived the iterative algorithmsfor computingtheof the FCM clusteringprocedures.minimizerof L, (G, a) as follows:Set h (arc,. . , a,e).For k 2 1, let ak (ark,. . . , a&), where for i 1,. . . , c,with respect to anyThey extended FCM algorithmswith respect to G, to FCM algorithmsprobabilitydistributionfunction G, and also showed that the optimal cluster centers should bethe fixed points of these generalized FCM algorithms.The numerical convergence properties ofthese generalized FCM algorithms had been created in [36]. These include the global convergence,the local convergence,and its rate of convergence.(D) GENERALIZATION 4Boundarydetection is an importantstep in pattern recognitionand image processing.Thereare various methods for detecting line boundaries;see [75]. When the boundariesare curved,the detection becomes more complicatedand difficult. In this case, the Hough transform(HT)(see [76,77]) is th e only major technique available. Although the HT is a powerful technique fordetecting curve boundarieslike circles and ellipses, it requires large computer storage and highcomputingtime. The fuzzy c-shells (FCS) clustering algorithms,proposed by Dave [45,46], arefine new techniques for detecting curve boundaries,especially circular and elliptical.These FCSalgorithmshave major advantagesover the HT in the areas of computer storage requirementsand computationalrequirements.The FCS clusteringalgorithmsare used to choose a fuzzy c-partitionp ( 1,. . . , pc) andcluster centers a (al,. . . ,a,) and radius T ( 1,. . . , rc) to minimize the objective functionJFCS (p, a, T), which is the weighted sum of squared distance of data from a hyper-ellipsoidalshell prototypeJFCS(P,W) &?(lj)(11%-a&,-ri)2j li lamongall a E (IV)‘,/.LE W, and T E (lR )c with m E [l, co).The necessaryconditionsfor a

A Surveyof FuzzyClusteringminimizer( ,a*, r*) of JFCS are the followingequationssatisfiedfor a* and r*:1-withpf(zj)The FCS clustering(see [45,53]) appliedtigatedits numerical algorithmsare iterationsthrough thesethese FCS’s in digital image sary conditions.Dave et al.Bezdek and Hathaway[78] inveset al. [79) gave a new approachtothe FCS.(E) GENERALIZATION 5Trauwaertet al. [51] derived a generalizationof the FCM based on the maximumlikelihoodprinciple for the c members of mutivariatenormal distributions.Consider a log likelihood functionnormalsJFML (p, a, IV) of a random sample X {XI,. . . , 5%) from c numbers of mutivariateN(ai,Wi),i 1,. . . ,c, wherein which ml, m2, ma E [l, 00). To derive iterative algorithms for a maximizerthey further limited to the case where ml, m2, and ms are either 1 or 2.only the case of ml 2 and m2 m3 1 appeared as an acceptable basisrecognitionalgorithm.In the case of ml 2 and m2 ms 1, Trauwaertiterative algorithm as follows:’Bij (xj -.i)TW;l(xj -ai),Aij i In ]Wi],of JFML (p, a, IV),They claimed thatfor a membershipet al. [51] gave thewherei l,**.,c,j l,.,n.This general algorithmis called “fuzzy product” since it is an optimizationof the productfuzzy determinants.In the case of Ai I for all i the algorithm becomes the FCM.(F)ofGENERALIZATION 6Mixtures of distributionshave been used extensively as models in a wide variety of importantpractical situations where data can be viewed as arising from two or more subpopulationsmixed invaried proportions.Classificationmaximum likelihood (CML) procedure is a remarkable mixturemaximum likelihood approach to clustering, see [80]. Let the population of interest be known or beassumed to consist of c different subpopulations,or classes, where the density of an observationxfrom class i is fi(x; e), i 1,. . , , c for some parameter6. Let the proportionof individualsin

M.-S. YANG10the population which are from class i be denoted by oi with cri E (0,l) and x: ,1 oi 1. Letx (51,. . . , z,} be a random sample of size n from the population. The CML procedure is theoptimization problem by choosing a hard c-partition Jo* and a proportion cy* and an estimate 6’*to maximize the log likelihood I31 (CL,(Y,0) of X, whereB1(p,a,e) eC(xj;e)c71. lnwfiCCpi(q)lnaifij l(xj;e),k lin which ,Q(z) E (0,1) and pi( ) . (2) 1 for all 2 E X. Yang [37] gave a new type ofobjective function Bm , (11,Q, 0), where 1 for all z E X with the fixed constantsin which pi(z) E [0, l] and pi(z) . . . &x)m E [l,co) and w 2 0. Thus, B,,, (/J, a, e), called the fuzzy CML objective function, is a fuzzyextension of Bi (cl, (Y,0). Based on B,,, (P, (Y,e), we can derive a variety of hard and fuzzyclustering algorithms.Now we consider the fuzzy CML objective function B,,, (p, Q, 0) with the multivariate normalsubpopulation densities N (ai, Wi), i 1,. . . , c. Then, we have thatBased on B m,ul(p, a, IV), we extend the fuzzy clustering algorithms of Trauwaert et al. [51] to amore general setting by adding a penalty termw C,“ , xi , pyn(zj) In oi . The construction()of these penalized types of fuzzy clustering algorithms is similar to that of [51]. In the specialcaseofWi I,i l,.,c.Wehavethat JFCM( a) -W2 2 Py(Xj) In%.j li lSince JPFCM just adds the penalty term (-w &xk,,LL; (zj) In ai)to JFCM,JPFCMiscalled the penalized FCM objective function. We get the following necessary conditions for a

A Survey of Fuzzy Clustering11minimizer (p, a, a) of JPFCM:aj CYi )Pi(q) Thus, the penalized FCM clustering algorithms for computing the minimizer of JPFCM (p, a, CX)are iterations through these necessary conditions. By comparison between the penalized FCMand the FCM, the difference is the penalty term. This penalized FCM has been used in parameterestimation of the normal mixtures. The numerical comparison between the penalized FCM andEM algorithms for parameter estimation of the normal mixtures has been studied in [68].(G) OTHER VARIATIONSFrom (A) to (F), we give a survey of six generalizations of the FCM. These generalizations areall based on the Euclidean norm )I . )I ( i.e., Lz-norm ). In fact, we can extend these to L,-normJFCM, (uya) 22CL;&where dij (chsl(xjk - aik I”) “’ is the L,-norm of (zj - ai). When p 2 it is just the FCM.In Mathematics and its applications, there are two extreme L,-norms most concerned. These areLr-norm and L,-norm.Thus,with dij f:JFCM (U, ,) ce I;"dijj l i lnJFCM,(Vand(Sjk - a&[k lcwith dij kzy8) CCpc dijj 1 i ls ) sjk I 1aik 1.We cannot find the necessary conditions to guide numerical solutions of these nonlinear constrained objective functions JFCM and JFCM,.Bobrowski and Bezdek [44] gave a method tofind approximate minimizers of JFCM and JFCM, by using a basis exchange algorithm usedfor discriminate analysis with the perception criterion function by Bobrowski and Niemiro [Bl].Jajuga [49] also gave a method to find approximate minimizers of JFCM based on a method forregression analysis with the Li-norm by Bloomfield and Steiger 1821. Jajuga [49] gave anothersimple method to find approximate minimizers of J FCM, which we shall describe next.Recall thatJFCM W, with dij 2 2k ;dijj l i l)zjcjk -f&k).k lThen U may be a minimizer of JFCM for a being fixed only if2(dij)l -‘)k l(dkj) “(“ ‘)Pij For U being fixed, let wij ,zjkPfaik,.Then,cJFCM (w, a ci lMm 16:11-Bsck l2j l“ij(zjk -aik)‘.

M.-S. YANG12The aik may be a minimizerof JFCM (w, a) for WQ being fixed only ifaii c,“ l wiixikC&iJajuga[49] suggestedfollowingan algorithmfor findingl,.,c,k l;.approximate,s.minimizersof JFCM basedon thesteps:wij lXika,k z2Pij’“iji Pg- aikl’C,“ I“iixikEYE” ,wi.j’andi l,*.,c,(&)1-l)dij k r (dkj) llCna-l)4. FUZZYGENERALIn the conventionalnearest neighborpatternsand 0 representsthe labeling iZjk--aailc1,wherek-NEARESTj l,.,n,k l,.,s.k lNEIGHBORRULEpattern classificationproblem, X representsthe set ofvariable of c categories.We assume that a set of ncorrectly classified samples (ICI, &), (x2,6 ) , . . . , (x,, 19,) is given and follows some distributionF(z, Q), where the 5:s take values in a metric space (X,d) and the 0:s take values in the setis given, where only the measurement 0 is observableby{1,2 ,., c}. A new pair (Q,&)the statistician,and it is desired to estimate BO by utilizing the informationcontainedin theset of correctly classified points.We shall call Z; E {xl, 2,. . . ,z,} a nearest neighbor to zif d(zL,z) mind,2 ,,., n d(xi, x). The nearest neighbor rule (NNR) has been a well-knowndecision rule and perhaps the simplest nonparametricdecision rule. The NNR, first proposed byFix and Hodges [83], assigns x to the category 0; of its nearest neighbor &.Cover and Hart [84] proved that the large sample NN risk R is bounded above by twice the Bayesprobabilityof error R* under the 0- 1 loss function L, where L counts an error whenever a mistakein classificationis made. Cover [85] extended the finite-action(classification)problem of Coverand Hart [84] to the infinite-action(estimation)problem. He proved that the NN risk R is stillbounded above by twice the Bayes risk for both metric and squared-errorloss function.Moreover,Wagner [86] considered the posterior probabilityof error L, P [e; # eo ) (21, el), . . . , (xn, On)].Here L, is a random variable that is a function of the observed samples and, furthermore,0 and he proved that under some conditions, L, converges to R with probabilityEL, P[ :,# ],one as n tends to co. This obviously extended the results of Cover and Hart [84]. There are soThe k-NNR is a supervisedapproach to patternmany papers about k-NNR in the literature.recognition.Althoughk-NNR is not a clusteringalgorithm,it is always used to cluster anunlabelleddata set through the rule. In fact, Wong and Lane [87] combined clusteringwithk-NNR and developed the so-called k-NN clustering procedure.In 1982, Duin [88] first proposed strategies and examples in the use of continuouslabelingvariable.Joiwik [89] derived a fuzzy k-NNR in use of Duin’s idea and fuzzy sets. Bezdek etal. (901 generalizedk-NNR in conjunctionwith the FCM clustering.Keller et al. [91] proposedfuzzy k-NNR which assigns class membership to a sample vector rather than assigning the vectorto a particularclass. Moreover, BQreau and Dubuisson[92] gave a fuzzy extended k-NNR byBut in a sequence of study in the fuzzychoosing a particularexponentialweight function.version of k-NNR, there is no result about convergence properties.Recently, Yang and Chen [93]described a fuzzy generalized k-NN algorithm which is a unified approach to a variety of fuzzy

A Survey of Fuzzy Clusteringk-NNR’sthatand thenis, the strongcreatedthe stochasticconsistencyconvergenceof posteriorand Chen [94] gave its rate of convergence.[84], Cover [85], and Wagnerof the fuzzy generalizedrisk of the fuzzy generalizedNNR.Moreover,They showed that the rate of convergencerisk of the fuzzy generalized NNR is exponentiallycase of the fuzzy generalized Ic-NNR, the strongand Hartproperty13NNR,Yangof posteriorfast. Since the conventionalIc-NNR is a specialconsistencyin [93] extends the results of Cover[86].The fuzzy generalized Ic-NNR had been shown that it could improve the performancerate inmost instances.A lot of researchers gave numerical examples and their applications.The resultsin [93,94] just give us the theoretic foundation of the fuzzy generalized Ic-NNR. These also confirmthe value of their applications.5. CONCLUSIONSWe have already reviewed numerous fuzzy clustering algorithms.But it is necessary to preIn general, the number c should beassume the number c of clusters for all these algorithms.unknown.Therefore, the method to find optimal c is very important.This kind of problem isusually called cluster validity.If we use the objective function JFCM (or JFCV, JFC,S, etc.) as a criterion for cluster validity,it is clear that JFCM must decrease monotonicallywith c increasing.So, in some sense, if thesample of size n is really grouped into E compact, well-separatedclusters, one would expect tosee JFCM decrease rapidly until c 6, but it shall decrease much more slowly thereafter until itreaches zero at c n. This argument has been advanced for hierarchicalclustering procedures.But it is not suitable for clustering procedures based on the objective function.An effectivelyformal way to proceed is to devise some validity criteria such as a cluster separation measure or touse other techniques such as bootstrap technique.We do not give a survey here. But interestedreaders may be directed towards references [95-1031.Finally, we conclude that the fuzzy clusteringalgorithmshave obtainedgreat success in avariety of substantiveareas.Our survey may give a good extensive view of researcherswhohave concerns with applicationsin cluster analysis, or even encourage readers in the appliedmathematicscommunityto try to use these techniques of fuzzy clustering.REFERENCES1. R.O. Duda and P.E. Hart, Pattern Classification and Scene Analysis, Wiley, New York, (1973).2. P.A. Devijver and J. Kitttler, Pattern Recognition: A Statistical Approach, Prentice-Hall, Englewood Cliffs,NJ, (1982).3. A.K. Jain and R.

methods based on the fuzzy relation. These can be found in (10,12-161. Fuzzy clustering, based on fuzzy relations, was first proposed by Tamura et al. [12]. They proposed an N-step procedure by using the composition of fuzzy relations beginning with a reflexive and symmetric fuzzy relation R in X. They showed that there is an n such that

Related Documents:

with ellipsoidal shape. Then, a fuzzy clustering algorithm for relational data is described (Davé and Sen,2002) Fuzzy k-means algorithm The most known and used fuzzy clustering algorithm is the fuzzy k-means (FkM) (Bezdek,1981). The FkM algorithm aims at discovering the best fuzzy

ing fuzzy sets, fuzzy logic, and fuzzy inference. Fuzzy rules play a key role in representing expert control/modeling knowledge and experience and in linking the input variables of fuzzy controllers/models to output variable (or variables). Two major types of fuzzy rules exist, namely, Mamdani fuzzy rules and Takagi-Sugeno (TS, for short) fuzzy .

fuzzy controller that uses an adaptive neuro-fuzzy inference system. Fuzzy Inference system (FIS) is a popular computing framework and is based on the concept of fuzzy set theories, fuzzy if and then rules, and fuzzy reasoning. 1.2 LITERATURE REVIEW: Implementation of fuzzy logic technology for the development of sophisticated

Different types of fuzzy sets [17] are defined in order to clear the vagueness of the existing problems. D.Dubois and H.Prade has defined fuzzy number as a fuzzy subset of real line [8]. In literature, many type of fuzzy numbers like triangular fuzzy number, trapezoidal fuzzy number, pentagonal fuzzy number,

Caiado, J., Maharaj, E. A., and D’Urso, P. (2015) Time series clustering. In: Handbook of cluster analysis. Chapman and Hall/CRC. Andrés M. Alonso Time series clustering. Introduction Time series clustering by features Model based time series clustering Time series clustering by dependence Introduction to clustering

the data set. In graph-theoretic fuzzy clustering, the graph representing the data structure is a fuzzy graph and di erent notions of connectivity lead to di erent types of clusters. The idea of fuzzy graphs is rst mentioned in [10] whereby the fuzzy analogues of several basic graph-theoretic concepts

named Fuzzy DBSCAN subsumes the previous ones, thus allowing to generate clusters with both fuzzy cores and fuzzy overlapping borders. Our proposals are compared w.r.t. state of the art fuzzy clustering methods over real world datasets. 1 Introduction The advent of the big data era has

A Primer on Automotive EMC for Non-EMC Engineers The automotive industry has changed drastically in recent years. Advancements in technology paired with tighter federal fuel and emissions regulations have resulted in the need to place more electrical systems into vehicles. This in turn places a greater emphasis on keeping the Electromagnetic Interference (EMI) of these systems from interfering .