期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Semantic-aware knowledge distillation with parameter-free feature uniformization
1
作者 Guangyu Guo Longfei Han +2 位作者 Le Wang Dingwen Zhang Junwei Han 《Visual Intelligence》 2023年第1期346-357,共12页
Knowledge distillation aims to distill knowledge from teacher networks to train student networks.Distilling intermediate features has attracted much attention in recent years as it can beflexibly applied in variousfiel... Knowledge distillation aims to distill knowledge from teacher networks to train student networks.Distilling intermediate features has attracted much attention in recent years as it can beflexibly applied in variousfields such as image classification,object detection and semantic segmentation.A critical obstacle of feature-based knowledge distillation is the dimension gap between the intermediate features of teacher and student,and plenty of methods have been proposed to resolve this problem.However,these works usually implement feature uniformization in an unsupervised way,lacking guidance to help the student network learn meaningful mapping functions in the uniformization process.Moreover,the dimension uniformization process of the student and teacher network is usually not equivalent as the mapping functions are different.To this end,some factors of the feature are discarded during parametric feature alignment,or some factors are blended in some non-parametric operations.In this paper,we propose a novel semantic-aware knowledge distillation scheme to solve these problems.We build a standalone feature-based classification branch to extract semantic-aware knowledge for better guiding the learning process of the student network.In addition,to avoid the inequivalence of feature uniformization between teacher and student,we design a novel parameter-free self-attention operation that can convert features of different dimensions into vectors of the same length.Experimental results show that the proposed knowledge distillation scheme outperforms existing feature-based distillation methods on the widely used CIFAR-100 and CINIC-10 datasets. 展开更多
关键词 Knowledge distillation Feature-based semantic-aware Parameter-free Self-attention Feature uniformization
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部