期刊文献+

基于流形学习的人体动作识别 被引量:30

Human action recognition based on manifold learning
原文传递
导出
摘要 目的提出了一个基于流形学习的动作识别框架,用来识别深度图像序列中的人体行为。方法从Kinect设备获得的深度信息中评估出人体的关节点信息,并用相对关节点位置差作为人体特征表达。在训练阶段,利用LE(Lalpacian eigenmaps)流形学习对高维空间下的训练集进行降维,得到低维隐空间下的运动模型。在识别阶段,用最近邻差值方法将测试序列映射到低维流形空间中去,然后进行匹配计算。在匹配过程中,通过使用改进的Hausdorff距离对低维空间下测试序列和训练运动集的吻合度和相似度进行度量。结果用Kinect设备捕获的数据进行了实验,取得了良好的效果;同时也在MSR Action3D数据库上进行了测试,结果表明在训练样本较多情况下,本文方法识别效果优于以往方法。结论实验结果表明本文方法适用于基于深度图像序列的人体动作识别。 Objective Human action recognition is a widely studied area in computer vision and machine learning; it has many potential applications including human computer interfaces, video surveillance, and health care. In the past decade, extensive research efforts focused on recognizing human action from monocular video sequences. Since human motion is articulated, capturing human joint characters accurately from video is a very difficult task. The recent introduction of real time depth cameras such as the Kinect sensor, give us the opportunity to use 3D depth data of a scene instead of pictures. In this paper, we present a manifold-based framework for human action recognition using depth image data captured from depth camera. Method With the recent release of Kinect sensor and the technology assessing skeleton joint position from depth image matured, recent research used 3D skeleton joint position information as human body representation and achieved good recognition performance. As we know, human action is composed of ordered posture set, and the difference between postures is only a few changes of 3D joints pairwise, most of the 3D information changes only little. In this paper, we estimated the 3D joint locations from Kinect depth images and use pairwise relative positions as the representation of human features. In the training phase, the LE ( Lalpacian eigenmaps ) is used to build action model in low dimensional space. In test phase, the nearest-neighbor interpolation technique is used to map test sequence to the manifold space, then measure the distance with the test sequence and the training data. A novel modified Hausdorff distance is used to measure similarity and fitness of the test sequence and the training data in the matching process. Result The recognition performance of the proposed method was evaluated from Kinect sensor dataset and the result confirmed the proposed method can work well in several experiments. We also tested the proposed method on the MSR Action3D dataset and achieved state of the art accuracy in our comparison with related work when the training set has many samples. Conclusion Manifold learning is an effective nonlinear dimensionality reduction method and low-dimensional motion models can be trained well when training sample size is large. We propose a novel human action recognition based on manifold learning in this paper. The experimental results show the effectiveness of the proposed method for human action recognition based on depth image sequence.
出处 《中国图象图形学报》 CSCD 北大核心 2014年第6期914-923,共10页 Journal of Image and Graphics
基金 国家自然科学基金项目(61303142 61173096 61103140) 浙江省自然科学基金项目(Y1110882 Y1110688 R1110679 LY13F020034) 教育部高等学校博士学科点专项科研基金项目(20113317110001) 浙江省教育厅一般科研项目(Y201330304)
关键词 KINECT SENSOR 人体动作识别 流形学习 HAUSDORFF距离 深度数据 Kinect sensor human action recognition manifold learning Hausdorff distance depth data
  • 相关文献

参考文献23

  • 1徐光祐,曹媛媛.动作识别与行为理解综述[J].中国图象图形学报,2009,14(2):189-195. 被引量:52
  • 2Plagemann C, Ganapathi V, Koller D, et al. Real-time identification and localization of body parts from depth images[C]//Proceedings of IEEE International Conference on Robotics and Automation. Anchorage, USA: IEEE, 2010: 3108-3113.
  • 3Schwarz L A, Mkhitaryan A, Mateus D, et al. Estimating human 3d pose from time-of-flight images based on geodesic distances and optical flow[C]//Proceedings of IEEE International Conference on Automatic Face & Gesture Recognition and Workshops. Santa Barbara, USA: IEEE, 2011: 700-706.
  • 4Shotton J, Sharp T, Kipman A, et al. Real-time human pose recognition in parts from single depth images[J]. Communications of the ACM, 2013, 56(1): 116-124.
  • 5Sim D G, Kwon O K, Park R H. Object matching algorithms using robust Hausdorff distance measures[J]. IEEE Transactions on Image Processing, 1999, 8(3): 425-429.
  • 6Laptev I. On space-time interest points[J]. International Journal of Computer Vision, 2005, 64(2-3): 107-123.
  • 7Efros A A, Berg A C, Mori G, et al. Recognizing action at a distance[C]//Proceedings of IEEE International Conference on Computer Vision. Nice, France: IEEE, 2003: 726-733.
  • 8Dalal N, Triggs B. Histograms of oriented gradients for human detection[C]//Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition. San Diego, USA: IEEE, 2005, 1: 886-893.
  • 9Yang X, Zhang C, Tian Y L. Recognizing actions using depth motion maps-based histograms of oriented gradients[C]//Proceedings of ACM International Conference on Multimedia. New York, USA: ACM, 2012: 1057-1060.
  • 10Oreifej O, Liu Z, Redmond W A. HON4D: Histogram of oriented 4D normals for activity recognition from depth sequences[C]//Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Portland USA: IEEE, 2013: 716-723.

二级参考文献62

  • 1刘玉栋,苏开娜,马丽.一种基于模型的步态识别方法[J].计算机工程与应用,2005,41(9):88-92. 被引量:16
  • 2苏菡,黄凤岗,洪文.一种基于步态分析的身份识别方法[J].系统仿真学报,2006,18(5):1292-1296. 被引量:6
  • 3何国辉,甘俊英.二维主元分析在人脸识别中的应用研究[J].计算机工程与设计,2006,27(24):4667-4669. 被引量:22
  • 4李和平,胡占义,吴毅红,吴福朝.基于半监督学习的行为建模与异常检测[J].软件学报,2007,18(3):527-537. 被引量:30
  • 5王科俊,侯本博.步态识别综述[J].中国图象图形学报,2007,12(7):1152-1160. 被引量:45
  • 6Zhang D, Gatica-Perez D, Bengio S, et al. Modeling individual group actions in meetings: a two-layer HMM framework[A]. In: Proceedings of IEEE CVPR Workshop on Detection and Recognition of Events in Video[ C ] , Washington, DC, USA,2004 : 117-125.
  • 7Olivier N, Horovitz E, Garg A. Layered representations for human activity recognition [ A ] . In: Proceedings of IEEE International Conference on Muhimodal Interfaces [ C ] , Pittsburgh, PA, USA, 2002 : 3-8.
  • 8Luo Y, Wu T D, Hwang J N. Object-based analysis and interpretation of human motion in sports video sequences by dynamic Bayesian networks [ J ] . Computer Vision and Image Understanding, 2003, 92(2-3) : 196-216.
  • 9Du Y T, Chen F, Xu W L, et al. Recognizing interaction activities using dynamic Bayesian network [ A ] . In: Proceedings of International Conference on Pattern Recognition [ C ], New York, USA. 2006: 618-621.
  • 10Buxton H, Gong S G. Advanced visual surveillance using Bayesian networks [ A ]. In: Proceedings of International Conference on Computer Vision [ C ] , Boston, MA, USA, 1995 : 111 - 123.

共引文献73

同被引文献211

引证文献30

二级引证文献165

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部