期刊文献+

p范数正则化支持向量机分类算法 被引量:19

Classification Algorithm of Support Vector Machine via p-norm Regularization
在线阅读 下载PDF
导出
摘要 L2范数罚支持向量机(Support vector machine,SVM)是目前使用最广泛的分类器算法之一,同时实现特征选择和分类器构造的L1范数和L0范数罚SVM算法也已经提出.但是,这两个方法中,正则化阶次都是事先给定,预设p=2或p=1.而我们的实验研究显示,对于不同的数据,使用不同的正则化阶次,可以改进分类算法的预测准确率.本文提出p范数正则化SVM分类器算法设计新模式,正则化范数的阶次p可取范围为0<p≤2.使用网格法选择模型参数值,使用迭代再权方法求解分类器目标函数,找出最小分类预测误差的模型参数值.在实际数据集上的实验结果验证了提出算法能够同时实现分类预测和特征选择,性能优于L2范数罚SVM,L1范数罚SVM和L0范数罚SVM. The L2 penalty support vector machine (SVM) algorithm is one of the most widely used learning algorithms, meanwhile L1 norm and L0 norm penalty support vector machines have been devised, which achieve simultaneously feature selection and classifier construction. However, in both methods, the regularization parameter is predetermined, i.e., the default p = 2 or p = 1. Our experimental study shows that different data, using a different regularization of order, can improve prediction accuracy of the classification algorithm. In this paper, new classifier design pattern of SVM based on p-norm regularization is proposed, where 0 〈 p 〈 2. We design grid method to select parameter values of model, use the iterative reweighted method to solve classification object function then discover the right parameter values of model at the minimum prediction error. The performance of classification and feature selection on real datasets indicate that the devised algorithm is better than L2-norm, L1-norm, and L0-norm SVM.
出处 《自动化学报》 EI CSCD 北大核心 2012年第1期76-87,共12页 Acta Automatica Sinica
基金 国家自然科学基金(21006127 20976193) 中国石油大学(北京)基础学科研究基金项目资助~~
关键词 迭代再权方法 p范数(0<p≤2) 支持向量机 特征选择 稀疏化模型 高维小样本数据 Iterative reweighted method, p-norm (0 〈 p ≤ 2), support vector machine (SVM), feature selection, sparse model, high-dimensional small sample dataset
  • 相关文献

参考文献2

二级参考文献14

  • 1Li T, Zhang C L, Ogihara M. A comparative study of feature selection and multiclass classification methods for tissue classification based on gene expression. Bioinformatics, 2004, 20(15): 2429-2437.
  • 2Zhu J, Rosset S, Hastie T, Tibshirani R. 1-norm support vector machines. In: Proceedings of the Advances in Neural Information Processing Systems. Cambridge, USA: The MIT Press. 2004. 49--56.
  • 3Kujala J, Aho T, Elomaa T. A walk from 2-norm SVM to 1-norm SVM. In: Proceedings of the 9th IEEE International Conference on Data Mining. Miami, USA: IEEE, 2009. 836-841.
  • 4Zou H, Hastie T. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society Series B, 2005, 67(2): 301-320.
  • 5Lin D, Pitler E, Foster D P, Ungar L H. In defense of l. In: Proceedings of ICML Workshop on Sparse Optimization and Variable Selection. Helsinki, Finland: The Omnipress, 2008.
  • 6Guyon I, Elisseeff A. An introduction to variable and feature selection. The Journal of Machine Learning Research, 2003, 3:1157-1182.
  • 7Myung J I, Navarro D J, Pitt M A. Model selection by normalized maximum likelihood. Journal of Mathematical Psychology, 2006, 50(2): 167-179.
  • 8Fan J Q, Fan Y Y. High dimensional classification using features annealed independence rules. Annals of statistics, 2008, 36(6): 2605-2637.
  • 9Greenshtein E, Park J. Application of non parametric empirical Bayes estimation to high dimensional classification. The Journal of Machine Learning Research, 2009, 10:1687-1704.
  • 10Forman G. An extensive empirical study of feature selection metrics for text classification. The Journal of Machine Learning Research, 2003, 3:1289-1305.

共引文献2321

同被引文献201

引证文献19

二级引证文献113

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部