期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Token Masked Pose Transformers Are Efficient Learners
1
作者 Xinyi Song Haixiang Zhang Shaohua Li 《Computers, Materials & Continua》 2025年第5期2735-2750,共16页
In recent years,Transformer has achieved remarkable results in the field of computer vision,with its built-in attention layers effectively modeling global dependencies in images by transforming image features into tok... In recent years,Transformer has achieved remarkable results in the field of computer vision,with its built-in attention layers effectively modeling global dependencies in images by transforming image features into token forms.However,Transformers often face high computational costs when processing large-scale image data,which limits their feasibility in real-time applications.To address this issue,we propose Token Masked Pose Transformers(TMPose),constructing an efficient Transformer network for pose estimation.This network applies semantic-level masking to tokens and employs three different masking strategies to optimize model performance,aiming to reduce computational complexity.Experimental results show that TMPose reduces computational complexity by 61.1%on the COCO validation dataset,with negligible loss in accuracy.Additionally,our performance on the MPII dataset is also competitive.This research not only enhances the accuracy of pose estimation but also significantly reduces the demand for computational resources,providing new directions for further studies in this field. 展开更多
关键词 Pattern recognition image processing neural network pose transformer
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部