Aiming at the problem of potential information noise introduced during the generation of ghost feature maps in GhostNet,this paper proposes a novel lightweight neural network model called ResghostNet.This model constr...Aiming at the problem of potential information noise introduced during the generation of ghost feature maps in GhostNet,this paper proposes a novel lightweight neural network model called ResghostNet.This model constructs the Resghost Module by combining residual connections and Adaptive-SE Blocks,which enhances the quality of generated feature maps through direct propagation of original input information and selection of important channels before cheap operations.Specifically,ResghostNet introduces residual connections on the basis of the Ghost Module to optimize the information flow,and designs a weight self-attention mechanism combined with SE blocks to enhance feature expression capabilities in cheap operations.Experimental results on the ImageNet dataset show that,compared to GhostNet,ResghostNet achieves higher accuracy while reducing the number of parameters by 52%.Although the computational complexity increases,by optimizing the usage strategy of GPU cachememory,themodel’s inference speed becomes faster.The ResghostNet is optimized in terms of classification accuracy and the number of model parameters,and shows great potential in edge computing devices.展开更多
基金funded by Science and Technology Innovation Project grant No.ZZKY20222304.
文摘Aiming at the problem of potential information noise introduced during the generation of ghost feature maps in GhostNet,this paper proposes a novel lightweight neural network model called ResghostNet.This model constructs the Resghost Module by combining residual connections and Adaptive-SE Blocks,which enhances the quality of generated feature maps through direct propagation of original input information and selection of important channels before cheap operations.Specifically,ResghostNet introduces residual connections on the basis of the Ghost Module to optimize the information flow,and designs a weight self-attention mechanism combined with SE blocks to enhance feature expression capabilities in cheap operations.Experimental results on the ImageNet dataset show that,compared to GhostNet,ResghostNet achieves higher accuracy while reducing the number of parameters by 52%.Although the computational complexity increases,by optimizing the usage strategy of GPU cachememory,themodel’s inference speed becomes faster.The ResghostNet is optimized in terms of classification accuracy and the number of model parameters,and shows great potential in edge computing devices.