期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Global Spatial-Temporal Information Encoder-Decoder Based Action Segmentation in Untrimmed Video 被引量:1
1
作者 Yichao Liu Yiyang Sun +2 位作者 Zhide Chen Chen Feng Kexin Zhu 《Tsinghua Science and Technology》 2025年第1期290-302,共13页
Action segmentation has made significant progress,but segmenting and recognizing actions from untrimmed long videos remains a challenging problem.Most state-of-the-art methods focus on designing models based on tempor... Action segmentation has made significant progress,but segmenting and recognizing actions from untrimmed long videos remains a challenging problem.Most state-of-the-art methods focus on designing models based on temporal convolution.However,the limitations of modeling long-term temporal dependencies and the inflexibility of temporal convolutions restrict the potential of these models.To address the issue of over-segmentation in existing action segmentation methods,which leads to classification errors and reduced segmentation quality,this paper proposes a global spatial-temporal information encoder-decoder based action segmentation method.The method proposed in this paper uses the global temporal information captured by refinement layer to assist the Encoder-Decoder(ED)structure in judging the action segmentation point more accurately and,at the same time,suppress the excessive segmentation phenomenon caused by the ED structure.The method proposed in this paper achieves 93%frame accuracy on the constructed real Tai Chi action dataset.The experimental results prove that this method can accurately and efficiently complete the long video action segmentation task. 展开更多
关键词 Encoder-Decoder(ED) Bidirectional Long Short-Term Memory(BiLSTM) Tai Chi action segmentation untrimmed video
原文传递
Prediction of Assembly Intent for Human-Robot Collaboration Based on Video Analytics and Hidden Markov Model
2
作者 Jing Qu Yanmei Li +2 位作者 Changrong Liu Wen Wang Weiping Fu 《Computers, Materials & Continua》 2025年第8期3787-3810,共24页
Despite the gradual transformation of traditional manufacturing by the Human-Robot Collaboration Assembly(HRCA),challenges remain in the robot’s ability to understand and predict human assembly intentions.This study ... Despite the gradual transformation of traditional manufacturing by the Human-Robot Collaboration Assembly(HRCA),challenges remain in the robot’s ability to understand and predict human assembly intentions.This study aims to enhance the robot’s comprehension and prediction capabilities of operator assembly intentions by capturing and analyzing operator behavior and movements.We propose a video feature extraction method based on the Temporal Shift Module Network(TSM-ResNet50)to extract spatiotemporal features from assembly videos and differentiate various assembly actions using feature differences between video frames.Furthermore,we construct an action recognition and segmentation model based on the Refined-Multi-Scale Temporal Convolutional Network(Refined-MS-TCN)to identify assembly action intervals and accurately acquire action categories.Experiments on our self-built reducer assembly action dataset demonstrate that our network can classify assembly actions frame by frame,achieving an accuracy rate of 83%.Additionally,we develop a HiddenMarkovModel(HMM)integrated with assembly task constraints to predict operator assembly intentions based on the probability transition matrix and assembly task constraints.The experimental results show that our method for predicting operator assembly intentions can achieve an accuracy of 90.6%,which is a 13.3%improvement over the HMM without task constraints. 展开更多
关键词 Human-robot collaboration assembly assembly intent prediction video feature extraction action recognition and segmentation HMM
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部