期刊文献+

不均衡数据情形的基于聚焦损失的CGAN的集成分类方法

Focused loss-based for imbalanced data scenarios integrated classification methods for CGAN
在线阅读 下载PDF
导出
摘要 针对非均衡数据的情形,基于条件生成对抗网络(conditional generative adversarial networks,CGAN),利用梯度提升树研究了聚焦损失的CGAN的集成分类方法.该方法首先通过CGAN降低不均衡率,通过聚焦损失的权值均衡结合GBDT算法,适当增加对少数类样本的关注度进而进一步提升分类器的分类性能.对方法的性质进行了研究,获得了若干理论成果.证明了:在一定条件下,由CGAN产生的经验条件分布收敛于相应总体的条件分布;聚集损失的CGAN方法其经验风险收敛到期望风险;该方法的估计量会收敛到使得期望风险最小化的函数.实验结果显示了聚焦损失的CGAN方法具有良好的表现. For the case of imbalanced data,an integrated classification method for CGAN-focal-loss was investigated based on conditional generative adversarial networks(CGAN)using gradient boosting trees.The method first reduces the imbalance rate by CGAN,and further improves the classification performance of the classifier by increasing the focus on a few classes of samples through the weight balancing of the focused loss combined with the GBDT algorithm.The properties of the method were investigated and several theoretical results were obtained.It was proved that the empirical conditional distribution generated by CGAN converges to the conditional distribution of the corresponding aggregate under certain conditions;that the empirical risk of the CGAN method with focused loss converges to the expected risk;and that the estimator of the method converges to the function that minimizes the expected risk.The experimental results show the good performance of the CGAN-focal-loss method.
作者 崔文泉 余厚莹 侯晓天 CUI Wenquan;YU Houying;HOU Xiaotian(Department of Statistics and Finance,School of Management,University of Science and of Technology of China,Hefei 230026,China)
出处 《中国科学技术大学学报》 CAS CSCD 北大核心 2020年第7期968-976,共9页 JUSTC
基金 国家自然科学基金(71873128)资助
关键词 非均衡数据 条件生成对抗网络 聚焦损失 集成学习 imbalanced data conditional generative adversarial networks(CGAN) focal loss ensemble learning
  • 相关文献

参考文献3

二级参考文献35

  • 1郑恩辉,李平,宋执环.代价敏感支持向量机[J].控制与决策,2006,21(4):473-476. 被引量:33
  • 2毛勇,周晓波,夏铮,尹征,孙优贤.特征选择算法研究综述[J].模式识别与人工智能,2007,20(2):211-218. 被引量:95
  • 3Searle S R. Linear models for unbalanced data[M]. Wiley, 1987.
  • 4Chawla N V, Bowyer K W, Hall L O. SMOTE: Synthetic minority over-sampling technique[J]. Artificial Intelli- gence Research, 2002, 16(3): 321-357.
  • 5Lee C Y, Lee Z J. A novel algorithm applied to classify unbalanced data[J]. Applied Soft Computing, 2012(12): 2481 2485.
  • 6Schapire R E. The strength of weak learnability[J]. Machine Learning, 1990, 5(2): 197-227.
  • 7Breiman L. Bagging predictors[J]. Machine Learning, 1996, 24(2): 123-140.
  • 8Galar M, Fern~ndez A, Barrenechea E. A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches[J]. IEEE Transactions on Systems, Man, and Cybernetics, Part C, 2012,42(4): 463-484.
  • 9Nitesh V C, Aleksandar L, Lawrence O H, et al. SMOTEBoost: Improving prediction of the minority class in boosting[C]// 7th European Conference on Principles and Practice of Knowledge Discovery in Databases, 2003, 107-119.
  • 10Zhou Z H, Liu X Y. Training cost-sensitive neural networks with methods addressing the class imbalance prob- lem[J]. IEEE Transactions on Knowledge and Data Engineering, 2006: 63-77.

共引文献89

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部