期刊文献+

基于稀疏Parzen窗密度估计的快速自适应相似度聚类方法 被引量:6

Fast Adaptive Similarity-based Clustering Using Sparse Parzen Window Density Estimation
在线阅读 下载PDF
导出
摘要 相似度聚类方法(Similarity-based clustering method,SCM)因其简单易实现和具有鲁棒性而广受关注.但由于内含相似度聚类算法(Similarity clustering algorithm,SCA)的高时间复杂度和凝聚型层次聚类(Agglomerative hierarchicalclustering,AHC)的高空间复杂度,SCM不适用大数据集场合.本文首先发现了SCM和核密度估计问题的本质联系,并以此入手,通过快速压缩集密度估计器(Fast reduced set density estimator,FRSDE)和基于图的松弛聚类(Graph-based relaxedclustering,GRC)算法提出了快速自适应相似度聚类方法(Fast adaptive similarity-based clustering method,FASCM).相比于原SCM,该方法的主要优点是:1)其总体渐近时间复杂度与样本容量呈线性关系;2)不依赖于人工经验的干预,具有了自适应性.由此,FASCM适用于大数据集环境.该方法的有效性在图像分割应用中进行了验证. Similarity-based clustering method (SCM) has received much attention because it is robust and can be implemented simply and easily. However, because of its high time complexity of the embedded similarity clustering algorithm (SCA) and high space complexity of the embedded agglomerative hierarchical clustering (AHC), SCM is impractical for large data sets. In this paper, the relationship is revealed between SCM and the kernel density estimation of samples, a novel fast adaptive similarity-based clustering method (FASCM) is accordingly proposed by adopting fast reduced set density estimator (FRSDE) and graph-based relaxed clustering (GRC). The distinctive advantages of FMSSC over MSSC exist in: 1) its asymptotic linear time complexity with the data size; 2) independent on artificial experience and its adaptability. Thus, FASCM is practical for large datasets. Its effectiveness has also been demonstrated in image segmentation examples.
出处 《自动化学报》 EI CSCD 北大核心 2011年第2期179-187,共9页 Acta Automatica Sinica
基金 国家自然科学基金(60903100 60975027 60773206)资助~~
关键词 相似度聚类 密度估计 时间复杂度 图像分割 Similarity-based clustering, density estimator, time complexity, image segmentation
  • 相关文献

参考文献22

  • 1Yang M S, Wu K L. A similaxity-based robust clustering method. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2004, 26(4): 434--448.
  • 2Deng Z H, Chung F L, Wang S T. FRSDE: fast reduced set density estimator using minimal enclosing ball approximation. Pattern Recognition, 2008, 41(4): 1363-1372.
  • 3Chung F L, Deng Z H, Wang S T. From minimum enclosing ball to fast fuzzy inference system training on large datasets. IEEE Transactions on Fuzzy Systems, 2009, 17(1): 173-184.
  • 4Lee C H, Zaiane O, Park H H, Huang J Y, Greiner R. Clustering high dimensional data: a graph-based relaxed optimization approach. Information Sciences, 2008, 178(23): 4501-4511.
  • 5Girolami M, Chao H. Probability density estimation from optimally condensed data samples. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2003, 25(10): 1253--1264.
  • 6Frigui H, Krishnapuram R. Clustering by competitive agglomeration. Pattern Recognition, 1997, 30(7): 1109-1119.
  • 7Frigui H, Krishnapuram R. A robust competitive clustering algorithm with applications in computer vision. IEEE Transactions on Pattern Analysis and Machine Intelligence, 1999, 21(5): 450-465.
  • 8Krishnapuram R, Frigui H, Nasraoui O. Fuzzy and possibilistic shell clustering algorithms and their application to boundary detection and surface approximation. 1EEE Transactions on Fuzzy Systems, 1995, 3(1): 29-43.
  • 9Tsang I W H, Kwok J T Y, Zurada J A. Generalized core vector machines. IEEE Transactions on Neural Networks, 2006, 17(5): 1126-1140.
  • 10Tsang I W, Kwok J T, Cheung P M. Core vector machines: fast SVM training on very large data sets. The Journal of Machine Learning Research, 2005, 6(12): 363-392.

二级参考文献2

共引文献68

同被引文献46

  • 1毛尚勤,黄心汉,王敏.基于密度聚类的彩色图像分割方法[J].华中科技大学学报(自然科学版),2011,39(S2):116-119. 被引量:2
  • 2Deng Z H, Chung F L, Wang S T. FRSDE: fast reduced set density estimator using minimal enclosing ball approximation. Pattern Recognition, 2008, 41(4): 1363-1372.
  • 3Tsang I, Kwok J, Zurada J. Generalized core vector machines. IEEE Transactions on Neural Networks, 2006, 17(5): 1126-1140.
  • 4Badoiu M, Clarkson K L. Optimal core-sets for balls. Computational Geometry: Theory and Applications, 2008, 40(1): 14-22.
  • 5Badoiu M, Har-Peled S, Indyk P. Approximate clustering via core-sets. In: Proceedings of the 34th Annual ACM Symposium on Theory of Computing. Quebec, Canada: ACM, 2002. 250-257.
  • 6Tsang I, Kwok J, Cheung P. Core vector machines: fast SVM training on very large data sets. The Journal of Ma- chine Learning Research, 2005, 6:363-392.
  • 7Xu D X. Energy, Entropy and Information Potential for Neural Computation [Ph.D. dissertation], University of Florida, USA, 1998.
  • 8Maynou J, Gallardo-Chacon J J, Vallverdu M, Caminal P, Perera A. Computational detection of transcription factor binding sites through differential Renyi entropy. IEEE Transactions on Information Theory, 2010, 56(2): 734-741.
  • 9Jenssen R. Kernel entropy component analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2010, 32(5): 847-860.
  • 10Chen S, Hong X, Harris C J. Probability density estimation with tunable kernels using orthogonal forward regression. IEEE Transactions on Systems, Man, and Cybernetics, Part B: Cybernetics, 2010, 40(4): 1101-1114.

引证文献6

二级引证文献25

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部