This paper considers the problem of long-term target tracking in complex scenes when tracking failures are unavoidable due to illumination change,target deformation,scale change,motion blur,and other factors.More spec...This paper considers the problem of long-term target tracking in complex scenes when tracking failures are unavoidable due to illumination change,target deformation,scale change,motion blur,and other factors.More specifically,a target tracking algorithm,called re-detection multi-feature fusion,is proposed based on the fusion of scale-adaptive kernel correlation filtering and re-detection.The target tracking algorithm trains three kernel correlation filters based on the histogram of oriented gradients,colour name,and local binary pattern features and then obtains the fusion weight of response graphs corresponding to different features based on average peak correlation energy criterion and uses weighted average to complete the position estimation of the tracked target.In order to deal with the problem that the target is occluded and disappears in the tracking process,a random fern classifier is trained to perform re-detection when the target is occluded.After comparing the OTB-50 target tracking dataset,the experimental results show that the proposed tracker can track the target well in the occlusion attribute video sequence in the OTB-100 test dataset and has a certain improvement in tracking accuracy and success rate compared with the traditional correlation filter tracker.展开更多
In the past decades,substantial progress has been made in human action recognition.However,most existing studies and datasets for human action recognition utilise still images or videos as the primary modality.Image-b...In the past decades,substantial progress has been made in human action recognition.However,most existing studies and datasets for human action recognition utilise still images or videos as the primary modality.Image-based approaches can be easily impacted by adverse environmental conditions.In this paper,the authors propose combining RGB images and point clouds from LiDAR sensors for human action recognition.A dynamic lateral convolutional network(DLCN)is proposed to fuse features from multi-modalities.The RGB features and the geometric information from the point clouds closely interact with each other in the DLCN,which is complementary in action recognition.The experimental results on the JRDB-Act dataset demonstrate that the proposed DLCN outperforms the state-of-the-art approaches of human action recognition.The authors show the potential of the proposed DLCN in various complex scenarios,which is highly valuable in real-world applications.展开更多
基金International Cooperation and Exchange Program of Shaanxi Province,Grant/Award Number:2022KW‐04Natural Science Foundation of Shaanxi Province,Grant/Award Number:2018JM6120+1 种基金Xi'an Science and Technology Plan Project,Grant/Award Number:21XJZZ0072Major Science and Technology Projects of Xian Yang City,Grant/Award Number:2017k01‐25‐12。
文摘This paper considers the problem of long-term target tracking in complex scenes when tracking failures are unavoidable due to illumination change,target deformation,scale change,motion blur,and other factors.More specifically,a target tracking algorithm,called re-detection multi-feature fusion,is proposed based on the fusion of scale-adaptive kernel correlation filtering and re-detection.The target tracking algorithm trains three kernel correlation filters based on the histogram of oriented gradients,colour name,and local binary pattern features and then obtains the fusion weight of response graphs corresponding to different features based on average peak correlation energy criterion and uses weighted average to complete the position estimation of the tracked target.In order to deal with the problem that the target is occluded and disappears in the tracking process,a random fern classifier is trained to perform re-detection when the target is occluded.After comparing the OTB-50 target tracking dataset,the experimental results show that the proposed tracker can track the target well in the occlusion attribute video sequence in the OTB-100 test dataset and has a certain improvement in tracking accuracy and success rate compared with the traditional correlation filter tracker.
基金National Key Research and Development Program of China,Grant/Award Number:2022YEB4700400National Natural Science Foundation of China,Grant/Award Numbers:U22B2041,62173056Education Department of Hainan Province,Grant/Award Number:Hnky2024ZD‐19。
文摘In the past decades,substantial progress has been made in human action recognition.However,most existing studies and datasets for human action recognition utilise still images or videos as the primary modality.Image-based approaches can be easily impacted by adverse environmental conditions.In this paper,the authors propose combining RGB images and point clouds from LiDAR sensors for human action recognition.A dynamic lateral convolutional network(DLCN)is proposed to fuse features from multi-modalities.The RGB features and the geometric information from the point clouds closely interact with each other in the DLCN,which is complementary in action recognition.The experimental results on the JRDB-Act dataset demonstrate that the proposed DLCN outperforms the state-of-the-art approaches of human action recognition.The authors show the potential of the proposed DLCN in various complex scenarios,which is highly valuable in real-world applications.