WJ-CH10-模式识别-聚类算法-05-英文版-教学课件.ppt
《WJ-CH10-模式识别-聚类算法-05-英文版-教学课件.ppt》由会员分享,可在线阅读,更多相关《WJ-CH10-模式识别-聚类算法-05-英文版-教学课件.ppt(98页珍藏版)》请在淘文阁 - 分享文档赚钱的网站上搜索。
1、(博士/教授/博导)1/模式识别Pattern Recognition Chapter 10(V)OTHER CLUSTERING ALGORITHMS4/10/20231vThe following types of algorithms will be considered:Graph theory based clustering algorithms.Competitive learning algorithms.Valley seeking clustering algorithms.Cost optimization clustering algorithms based on:B
2、ranch and bound approach.Simulated annealing methodology.Deterministic annealing.Genetic algorithms.Density-based clustering algorithms.Clustering algorithms for high dimensional data sets.OTHER CLUSTERING ALGORITHMSOTHER CLUSTERING ALGORITHMS2GRAPH THEORY BASED CLUSTERING GRAPH THEORY BASED CLUSTER
3、ING ALGORITHMSALGORITHMSvIn principle,such algorithms are capable of detecting clusters of various shapes,at least when they are well separated.In the sequel we discuss algorithms that are based on:The Minimum Spanning Tree(MST).Regions of influence.Directed trees.3vMinimum Spanning Tree(MST)algorit
4、hms Preliminaries:Let G be the complete graph,each node of which corresponds to a point of the data set X.e=(xi,xj)denote an edge of G connecting xi and xj.wed(xi,xj)denote the weight of the edge e.Definitions:Two edges e1 and e2 are k steps away from each other if the minimum path that connects a v
5、ertex of e1 and a vertex of e2 contains k-1 edges.A Spanning Tree of G is a connected graph that:Contains all the vertices of the graph.Has no loops.The weight of a Spanning Tree is the sum of weights of its edges.A Minimum Spanning Tree(MST)of G is a spanning tree with minimum weight(when all wes a
6、re different from each other,the MST is unique).4vMinimum Spanning Tree(MST)algorithms(cont)Example 1:For the MST in the figure and for k=2 and q=3 we have:For e0:we0=17,me0=2.3,e0=0.95.we0 lies 15.5*e0 away from me0,hence it is inconsistent.For e11:we11=3,me11=2.5,e11=2.12.we11 lies 0.24*e11 away f
7、rom me11,hence it is consistent.6vMinimum Spanning Tree(MST)algorithms(cont)Remarks:The algorithm depends on the choices of k and q.The algorithm is insensitive to the order in which the data points are considered.No initial conditions are required,no convergence aspects are involved.The algorithm w
8、orks well for many cases where the clusters are well separated.A problem may occur when a“large”edge e has another“large”edge as its neighbor.In this case,e is likely not to be characterized as inconsistent and the algorithm may fail to unravel the underlying clustering structure correctly.7vAlgorit
9、hms based on Regions of Influence(ROI)Definition:The region of influence of two distinct vectors xi,xjX is defined as:R(xi,xj)=x:cond(d(x,xi),d(x,xj),d(xi,xj),xi xj where cond(d(x,xi),d(x,xj),d(xi,xj)may be defined as:a)maxd(x,xi),d(x,xj)d(xi,xj),b)d2(x,xi)+d2(x,xj)d2(xi,xj),c)(d2(x,xi)+d2(x,xj)d2(x
10、i,xj)OR(mind(x,xi),d(x,xj)d(xi,xj),d)(maxd(x,xi),d(x,xj)d(xi,xj)OR(mind(x,xi),d(x,xj)d(xi,xj),where affects the size of the ROI defined by xi,xj and is called relative edge consistency.8vAlgorithms based on Regions of Influence(cont)Remarks:The algorithm is insensitive to the order in which the pair
11、s are considered.In the choices of cond in(c)and(d),must be chosen a priori.For the resulting graphs:-if the choice(a)is used for cond,they are called relative neighborhood graphs(RNGs)-if the choice(b)is used for cond,they are called Gabriel graphs(GGs)Several results show that better clusterings a
12、re produced when(c)and(d)conditions are used in the place of cond,instead of(a)and(b).10vAlgorithms based on Directed Trees Definitions:A directed graph is a graph whose edges are directed.A set of edges ei1,eiq constitute a directed path from a vertex A to a vertex B,if,A is the initial vertex of e
13、i1 B is the final vertex of eiq The destination vertex of the edge eij,j=1,q-1,is the departure vertex of the edge eij+1.(In figure(a)the sequence e1,e2,e3 constitute a directed path connecting the vertices A and B).11vAlgorithms based on Directed Trees(cont Clustering Algorithm based on Directed Tr
14、eesSet to a specific value.Determine ni,i=1,N.Compute gij,i,j=1,N,ij.For i=1 to NIf ni=0 then-xi is the root of a new directed tree.Else-Determine xr such that gir=maxxji()gij-If gir0 thenoxr is the parent of xi(there exists a directed edge from xi to xr).13vAlgorithms based on Directed Trees(cont)R
15、emarks:The root xi of a directed tree is the point in i()with the most dense neighborhood.The branch that handles the case gir=0 ensures that no circles occur.The algorithm is sensitive to the order of consideration of the data points.For proper choice of and large N,this scheme behaves as a mode-se
16、eking algorithm(see a later section).Example 2:In the figure below,the size of the edge of the grid is 1 and=1.1.The above algorithm gives the directed trees shown in the figure.15COMPETITIVE LEARNING ALGORITHMSCOMPETITIVE LEARNING ALGORITHMSvThe main ideaEmploy a set of representatives wj(in the se
17、quel we consider only point representatives).Move them to regions of the vector space that are“dense”in vectors of X.vCommentsIn general,representatives are updated each time a new vector xX is presented to the algorithm(pattern mode algorithms).These algorithms do not necessarily stem from the opti
18、mization of a cost function.vThe strategyFor a given vector xAll representatives compete to each otherThe winner(representative that lies closest to x)moves towards x.The losers(the rest of the representatives)either remain unchanged or move towards x but at a much slower rate.16Remarks:h(x,wi)is an
19、 appropriately defined function(see below).and are the learning rates controlling the updating of the winner and the losers,respectively(may differ from looser to looser).A threshold of similarity (carefully chosen)controls the similarity between x and a representative wj.-If d(x,wj),for some distan
20、ce measure,x and wj are considered as dissimilar.The termination criterion is|W(t)-W(t-1)|tmax)(max allowable no of iterations)Assign each xX to the cluster whose representative wj lies closest to x.-(*)d(.)may be the Euclidean distance or other distances(e.g.,Itakura-Saito distortion).In addition s
21、imilarity measures may be used(in this case min is replaced by max).(*)is the learning rate and takes values in 0,1.19vBasic Competitive Learning Algorithm(cont)Remarks:In this scheme losers remain unchanged.The winner,after the updating,lies in the line segment formed by wj(t-1)and x.A priori knowl
22、edge of the number of clusters is required.If a representative is initialized far away from the regions where the points of X lie,it will never win.Possible solution:Initialize all representatives using vectors of X.Versions of the algorithm with variable learning rate have also been studied.20vLeak
23、y Learning Algorithm The same with the Basic Competitive Learning Algorithm except part(D),the updating equation of the representatives,which becomes where w and l are the learning rates in 0,1 and wl.Remarks:All representatives move towards x but the losers move at a much slower rate than the winne
24、r does.The algorithm does not suffer from the problem of poor initialization of the representatives(why?).An algorithm in the same spirit is the“neural-gas”algorithm,where l varies from loser to loser and decays as the corresponding representatives lie away from x.This algorithm results from the opt
25、imization of a cost function.21vConscientious Competitive Learning Algorithms Main Idea:Discourage a representative from winning if it has won many times in the past.Do this by assigning a“conscience”to each representative.A simple implementationEquip each representative wj,j=1,m,with a counter fj t
- 配套讲稿:
如PPT文件的首页显示word图标,表示该PPT已包含配套word讲稿。双击word图标可打开word文档。
- 特殊限制:
部分文档作品中含有的国旗、国徽等图片,仅作为作品整体效果示例展示,禁止商用。设计者仅对作品中独创性部分享有著作权。
- 关 键 词:
- WJ CH10 模式识别 算法 05 英文 教学 课件
限制150内