As an essential field of multimedia and computer vision,3D shape recognition has attracted much research attention in recent years.Multiview-based approaches have demonstrated their superiority in generating effective...As an essential field of multimedia and computer vision,3D shape recognition has attracted much research attention in recent years.Multiview-based approaches have demonstrated their superiority in generating effective 3D shape representations.Typical methods usually extract the multiview global features and aggregate them together to generate 3D shape descriptors.However,there exist two disadvantages:First,the mainstream methods ignore the comprehensive exploration of local information in each view.Second,many approaches roughly aggregate multiview features by adding or concatenating them together.The information loss for some discriminative characteristics limits the representation effectiveness.To address these problems,a novel architecture named region-based joint attention network(RJAN)was proposed.Specifically,the authors first design a hierarchical local information exploration module for view descriptor extraction.The region-to-region and channel-to-channel relationships from different granularities can be comprehensively explored and utilised to provide more discriminative characteristics for view feature learning.Subsequently,a novel relation-aware view aggregation module is designed to aggregate the multiview features for shape descriptor generation,considering the view-to-view relationships.Extensive experiments were conducted on three public databases:ModelNet40,ModelNet10,and ShapeNetCore55.RJAN achieves state-of-the-art performance in the tasks of 3D shape classification and 3D shape retrieval,which demonstrates the effectiveness of RJAN.The code has been released on https://github.com/slurrpp/RJAN.展开更多
Background Three-dimensional(3D)shape representation using mesh data is essential in various applications,such as virtual reality and simulation technologies.Current methods for extracting features from mesh edges or ...Background Three-dimensional(3D)shape representation using mesh data is essential in various applications,such as virtual reality and simulation technologies.Current methods for extracting features from mesh edges or faces struggle with complex 3D models because edge-based approaches miss global contexts and face-based methods overlook variations in adjacent areas,which affects the overall precision.To address these issues,we propose the Feature Discrimination and Context Propagation Network(FDCPNet),which is a novel approach that synergistically integrates local and global features in mesh datasets.Methods FDCPNet is composed of two modules:(1)the Feature Discrimination Module,which employs an attention mechanism to enhance the identification of key local features,and(2)the Context Propagation Module,which enriches key local features by integrating global contextual information,thereby facilitating a more detailed and comprehensive representation of crucial areas within the mesh model.Results Experiments on popular datasets validated the effectiveness of FDCPNet,showing an improvement in the classification accuracy over the baseline MeshNet.Furthermore,even with reduced mesh face numbers and limited training data,FDCPNet achieved promising results,demonstrating its robustness in scenarios of variable complexity.展开更多
目的:比较使用口内扫描法和传统硅橡胶法制取种植磨牙单冠的临床修复效果评价。方法:选择62例患者(共86颗后牙),随机分为试验组和对照组,每组43颗牙(其中前磨牙13颗,磨牙30颗)。临床分别采用硅橡胶法和口内扫描法制取印模,制作氧化锆全...目的:比较使用口内扫描法和传统硅橡胶法制取种植磨牙单冠的临床修复效果评价。方法:选择62例患者(共86颗后牙),随机分为试验组和对照组,每组43颗牙(其中前磨牙13颗,磨牙30颗)。临床分别采用硅橡胶法和口内扫描法制取印模,制作氧化锆全瓷修复体,于治疗后2周复诊,采用VAS评分量表及改良的美国公共卫生总署(United states public health service,USPHS)评价系统对修复效果及舒适度进行评价。结果:口内扫描法组的患者满意度更高,舒适度更好,两组USPHS评分中除口内扫描组修复体的边缘适合性A级比例比对照组高外,其余指标的A级比例差异无统计学意义(P>0.05)。结论:口内扫描技术应用在种植磨牙单冠的印模制作中,修复效果良好、精度高外,患者满意度高,值得临床推广应用。展开更多
针对室内环境中超宽带(Ultra-Wideband,UWB)信号易受障碍物遮挡导致非视距(Non Line of Sight,NLOS)误差的问题,本文提出了一种基于激光雷达(Light Detection And Ranging,LiDAR)点云识别UWB NLOS的融合定位方法,该方法利用LiDAR点云信...针对室内环境中超宽带(Ultra-Wideband,UWB)信号易受障碍物遮挡导致非视距(Non Line of Sight,NLOS)误差的问题,本文提出了一种基于激光雷达(Light Detection And Ranging,LiDAR)点云识别UWB NLOS的融合定位方法,该方法利用LiDAR点云信息辅助UWBNLOS识别,并通过UWB视距(LineofSight,LOS)测距值消除LiDAR同时定位与建图(Simultaneous Localization and Mapping,SLAM)过程中的累计误差,从而提高室内融合定位精度和鲁棒性。首先,采用八叉树对LiDAR点云进行处理,根据UWB基准站位置信息构建测距方向,并从LiDAR点云中提取测距方向上相关区域的点云数据。然后,通过3D Alpha Shape算法对所提取点云中可能阻碍UWB信号传播的障碍物进行轮廓提取。此外,根据分析提取的障碍物轮廓和UWB测距方向的空间关系,以此有效判定UWB信号是否存在NLOS测距情况。最后,剔除UWB测距过程中存在的NLOS测距值,通过紧组合方式,采用扩展卡尔曼滤波(EKF)将UWB LOS测距值和LiDAR SLAM的定位信息进行融合解算,消除LiDAR SLAM定位结果中的累积误差,以此提高融合定位精度和鲁棒性。为验证本文所提出的融合定位算法的有效性,通过搭建的融合定位实验平台在教学楼大厅进行了NLOS静态识别实验,在地下停车场进行了动态NLOS识别与动态定位实验。实验结果表明,该方法能够显著提高在室内复杂环境中的NLOS识别与定位的准确性,相较于单传感器定位与UWB原始测距值与LiDAR SLAM紧组合EKF的定位方法,NLOS识别准确率为93.22%,定位精度分别提高了49.24%、47.03%、96.13%,定位误差为0.067 m,实现了亚分米级室内定位。展开更多
Hand gesture recognition is a popular topic in computer vision and makes human-computer interaction more flexible and convenient.The representation of hand gestures is critical for recognition.In this paper,we propose...Hand gesture recognition is a popular topic in computer vision and makes human-computer interaction more flexible and convenient.The representation of hand gestures is critical for recognition.In this paper,we propose a new method to measure the similarity between hand gestures and exploit it for hand gesture recognition.The depth maps of hand gestures captured via the Kinect sensors are used in our method,where the 3D hand shapes can be segmented from the cluttered backgrounds.To extract the pattern of salient 3D shape features,we propose a new descriptor-3D Shape Context,for 3D hand gesture representation.The 3D Shape Context information of each 3D point is obtained in multiple scales because both local shape context and global shape distribution are necessary for recognition.The description of all the 3D points constructs the hand gesture representation,and hand gesture recognition is explored via dynamic time warping algorithm.Extensive experiments are conducted on multiple benchmark datasets.The experimental results verify that the proposed method is robust to noise,articulated variations,and rigid transformations.Our method outperforms state-of-the-art methods in the comparisons of accuracy and efficiency.展开更多
In the shape analysis community,decomposing a 3D shape intomeaningful parts has become a topic of interest.3D model segmentation is largely used in tasks such as shape deformation,shape partial matching,skeleton extra...In the shape analysis community,decomposing a 3D shape intomeaningful parts has become a topic of interest.3D model segmentation is largely used in tasks such as shape deformation,shape partial matching,skeleton extraction,shape correspondence,shape annotation and texture mapping.Numerous approaches have attempted to provide better segmentation solutions;however,the majority of the previous techniques used handcrafted features,which are usually focused on a particular attribute of 3Dobjects and so are difficult to generalize.In this paper,we propose a three-stage approach for using Multi-view recurrent neural network to automatically segment a 3D shape into visually meaningful sub-meshes.The first stage involves normalizing and scaling a 3D model to fit within the unit sphere and rendering the object into different views.Contrasting viewpoints,on the other hand,might not have been associated,and a 3D region could correlate into totally distinct outcomes depending on the viewpoint.To address this,we ran each view through(shared weights)CNN and Bolster block in order to create a probability boundary map.The Bolster block simulates the area relationships between different views,which helps to improve and refine the data.In stage two,the feature maps generated in the previous step are correlated using a Recurrent Neural network to obtain compatible fine detail responses for each view.Finally,a layer that is fully connected is used to return coherent edges,which are then back project to 3D objects to produce the final segmentation.Experiments on the Princeton Segmentation Benchmark dataset show that our proposed method is effective for mesh segmentation tasks.展开更多
基金the National Key Research and Development Program of China,Grant/Award Number:2020YFB1711704the National Natural Science Foundation of China,Grant/Award Number:62272337。
文摘As an essential field of multimedia and computer vision,3D shape recognition has attracted much research attention in recent years.Multiview-based approaches have demonstrated their superiority in generating effective 3D shape representations.Typical methods usually extract the multiview global features and aggregate them together to generate 3D shape descriptors.However,there exist two disadvantages:First,the mainstream methods ignore the comprehensive exploration of local information in each view.Second,many approaches roughly aggregate multiview features by adding or concatenating them together.The information loss for some discriminative characteristics limits the representation effectiveness.To address these problems,a novel architecture named region-based joint attention network(RJAN)was proposed.Specifically,the authors first design a hierarchical local information exploration module for view descriptor extraction.The region-to-region and channel-to-channel relationships from different granularities can be comprehensively explored and utilised to provide more discriminative characteristics for view feature learning.Subsequently,a novel relation-aware view aggregation module is designed to aggregate the multiview features for shape descriptor generation,considering the view-to-view relationships.Extensive experiments were conducted on three public databases:ModelNet40,ModelNet10,and ShapeNetCore55.RJAN achieves state-of-the-art performance in the tasks of 3D shape classification and 3D shape retrieval,which demonstrates the effectiveness of RJAN.The code has been released on https://github.com/slurrpp/RJAN.
基金Supported by the National Key R&D Program of China(2022YFC3803600).
文摘Background Three-dimensional(3D)shape representation using mesh data is essential in various applications,such as virtual reality and simulation technologies.Current methods for extracting features from mesh edges or faces struggle with complex 3D models because edge-based approaches miss global contexts and face-based methods overlook variations in adjacent areas,which affects the overall precision.To address these issues,we propose the Feature Discrimination and Context Propagation Network(FDCPNet),which is a novel approach that synergistically integrates local and global features in mesh datasets.Methods FDCPNet is composed of two modules:(1)the Feature Discrimination Module,which employs an attention mechanism to enhance the identification of key local features,and(2)the Context Propagation Module,which enriches key local features by integrating global contextual information,thereby facilitating a more detailed and comprehensive representation of crucial areas within the mesh model.Results Experiments on popular datasets validated the effectiveness of FDCPNet,showing an improvement in the classification accuracy over the baseline MeshNet.Furthermore,even with reduced mesh face numbers and limited training data,FDCPNet achieved promising results,demonstrating its robustness in scenarios of variable complexity.
文摘目的:比较使用口内扫描法和传统硅橡胶法制取种植磨牙单冠的临床修复效果评价。方法:选择62例患者(共86颗后牙),随机分为试验组和对照组,每组43颗牙(其中前磨牙13颗,磨牙30颗)。临床分别采用硅橡胶法和口内扫描法制取印模,制作氧化锆全瓷修复体,于治疗后2周复诊,采用VAS评分量表及改良的美国公共卫生总署(United states public health service,USPHS)评价系统对修复效果及舒适度进行评价。结果:口内扫描法组的患者满意度更高,舒适度更好,两组USPHS评分中除口内扫描组修复体的边缘适合性A级比例比对照组高外,其余指标的A级比例差异无统计学意义(P>0.05)。结论:口内扫描技术应用在种植磨牙单冠的印模制作中,修复效果良好、精度高外,患者满意度高,值得临床推广应用。
文摘针对室内环境中超宽带(Ultra-Wideband,UWB)信号易受障碍物遮挡导致非视距(Non Line of Sight,NLOS)误差的问题,本文提出了一种基于激光雷达(Light Detection And Ranging,LiDAR)点云识别UWB NLOS的融合定位方法,该方法利用LiDAR点云信息辅助UWBNLOS识别,并通过UWB视距(LineofSight,LOS)测距值消除LiDAR同时定位与建图(Simultaneous Localization and Mapping,SLAM)过程中的累计误差,从而提高室内融合定位精度和鲁棒性。首先,采用八叉树对LiDAR点云进行处理,根据UWB基准站位置信息构建测距方向,并从LiDAR点云中提取测距方向上相关区域的点云数据。然后,通过3D Alpha Shape算法对所提取点云中可能阻碍UWB信号传播的障碍物进行轮廓提取。此外,根据分析提取的障碍物轮廓和UWB测距方向的空间关系,以此有效判定UWB信号是否存在NLOS测距情况。最后,剔除UWB测距过程中存在的NLOS测距值,通过紧组合方式,采用扩展卡尔曼滤波(EKF)将UWB LOS测距值和LiDAR SLAM的定位信息进行融合解算,消除LiDAR SLAM定位结果中的累积误差,以此提高融合定位精度和鲁棒性。为验证本文所提出的融合定位算法的有效性,通过搭建的融合定位实验平台在教学楼大厅进行了NLOS静态识别实验,在地下停车场进行了动态NLOS识别与动态定位实验。实验结果表明,该方法能够显著提高在室内复杂环境中的NLOS识别与定位的准确性,相较于单传感器定位与UWB原始测距值与LiDAR SLAM紧组合EKF的定位方法,NLOS识别准确率为93.22%,定位精度分别提高了49.24%、47.03%、96.13%,定位误差为0.067 m,实现了亚分米级室内定位。
基金supported by the National Natural Science Foundation of China(61773272,61976191)the Six Talent Peaks Project of Jiangsu Province,China(XYDXX-053)Suzhou Research Project of Technical Innovation,Jiangsu,China(SYG201711)。
文摘Hand gesture recognition is a popular topic in computer vision and makes human-computer interaction more flexible and convenient.The representation of hand gestures is critical for recognition.In this paper,we propose a new method to measure the similarity between hand gestures and exploit it for hand gesture recognition.The depth maps of hand gestures captured via the Kinect sensors are used in our method,where the 3D hand shapes can be segmented from the cluttered backgrounds.To extract the pattern of salient 3D shape features,we propose a new descriptor-3D Shape Context,for 3D hand gesture representation.The 3D Shape Context information of each 3D point is obtained in multiple scales because both local shape context and global shape distribution are necessary for recognition.The description of all the 3D points constructs the hand gesture representation,and hand gesture recognition is explored via dynamic time warping algorithm.Extensive experiments are conducted on multiple benchmark datasets.The experimental results verify that the proposed method is robust to noise,articulated variations,and rigid transformations.Our method outperforms state-of-the-art methods in the comparisons of accuracy and efficiency.
基金supported by the National Natural Science Foundation of China (61671397).
文摘In the shape analysis community,decomposing a 3D shape intomeaningful parts has become a topic of interest.3D model segmentation is largely used in tasks such as shape deformation,shape partial matching,skeleton extraction,shape correspondence,shape annotation and texture mapping.Numerous approaches have attempted to provide better segmentation solutions;however,the majority of the previous techniques used handcrafted features,which are usually focused on a particular attribute of 3Dobjects and so are difficult to generalize.In this paper,we propose a three-stage approach for using Multi-view recurrent neural network to automatically segment a 3D shape into visually meaningful sub-meshes.The first stage involves normalizing and scaling a 3D model to fit within the unit sphere and rendering the object into different views.Contrasting viewpoints,on the other hand,might not have been associated,and a 3D region could correlate into totally distinct outcomes depending on the viewpoint.To address this,we ran each view through(shared weights)CNN and Bolster block in order to create a probability boundary map.The Bolster block simulates the area relationships between different views,which helps to improve and refine the data.In stage two,the feature maps generated in the previous step are correlated using a Recurrent Neural network to obtain compatible fine detail responses for each view.Finally,a layer that is fully connected is used to return coherent edges,which are then back project to 3D objects to produce the final segmentation.Experiments on the Princeton Segmentation Benchmark dataset show that our proposed method is effective for mesh segmentation tasks.