期刊文献+
共找到381篇文章
< 1 2 20 >
每页显示 20 50 100
Keypoint全功能肌电诱发电位故障维修1例
1
作者 晋虎 《医疗卫生装备》 CAS 2012年第8期145-146,共2页
Keypoint全功能肌电诱发电位是维迪公司的一款性能稳定可靠、使用便捷的台式肌电图,在使用中轻轻点击快速完成数据采集即可将患者从痛苦的检查中解放出来。Keypoint具有保留所有原始的波形提供给医师在报告时阅读分析参考诊断。
关键词 keypoint 肌电诱发电位 全功能 故障维修 数据采集 肌电图
暂未订购
基于Keypoint RCNN改进模型的物体抓取检测算法 被引量:14
2
作者 夏浩宇 索双富 +2 位作者 王洋 安琪 张妙恬 《仪器仪表学报》 EI CAS CSCD 北大核心 2021年第4期236-246,共11页
机器人抓取在工业中的应用有两个难点:如何准确地检测可抓取物体,以及如何从检测出的多个物体中选择最优抓取目标。本文在Keypoint RCNN模型中引入同方差不确定性学习各损失的权重,并在特征提取器中加入注意力模块,构成了Keypoint RCNN... 机器人抓取在工业中的应用有两个难点:如何准确地检测可抓取物体,以及如何从检测出的多个物体中选择最优抓取目标。本文在Keypoint RCNN模型中引入同方差不确定性学习各损失的权重,并在特征提取器中加入注意力模块,构成了Keypoint RCNN改进模型。基于改进模型提出了两阶段物体抓取检测算法,第一阶段用模型预测物体掩码和关键点,第二阶段用掩码和关键点计算物体的抓取描述和重合度,重合度表示抓取时的碰撞程度,根据重合度可以从多个可抓取物体中选择最优抓取目标。对照实验证明,相较原模型,Keypoint RCNN改进模型在目标检测、实例分割、关键点检测上的性能均有提高,在自建数据集上的平均精度分别为85.15%、79.66%、86.63%,机器人抓取实验证明抓取检测算法能够准确计算物体的抓取描述、选择最优抓取,引导机器人无碰撞地抓取目标。 展开更多
关键词 抓取检测 keypoint RCNN改进模型 损失权重 注意力模块 抓取描述 重合度 最优抓取
原文传递
Keypoint Description Using Statistical Descriptor with Similarity-Invariant Regions 被引量:2
3
作者 Ibrahim El rube Sameer Alsharif 《Computer Systems Science & Engineering》 SCIE EI 2022年第7期407-421,共15页
This article presents a method for the description of key points using simple statistics for regions controlled by neighboring key points to remedy the gap in existing descriptors.Usually,the existent descriptors such... This article presents a method for the description of key points using simple statistics for regions controlled by neighboring key points to remedy the gap in existing descriptors.Usually,the existent descriptors such as speeded up robust features(SURF),Kaze,binary robust invariant scalable keypoints(BRISK),features from accelerated segment test(FAST),and oriented FAST and rotated BRIEF(ORB)can competently detect,describe,and match images in the presence of some artifacts such as blur,compression,and illumination.However,the performance and reliability of these descriptors decrease for some imaging variations such as point of view,zoom(scale),and rotation.The intro-duced description method improves image matching in the event of such distor-tions.It utilizes a contourlet-based detector to detect the strongest key points within a specified window size.The selected key points and their neighbors con-trol the size and orientation of the surrounding regions,which are mapped on rec-tangular shapes using polar transformation.The resulting rectangular matrices are subjected to two-directional statistical operations that involve calculating the mean and standard deviation.Consequently,the descriptor obtained is invariant(translation,rotation,and scale)because of the two methods;the extraction of the region and the polar transformation techniques used in this paper.The descrip-tion method introduced in this article is tested against well-established and well-known descriptors,such as SURF,Kaze,BRISK,FAST,and ORB,techniques using the standard OXFORD dataset.The presented methodology demonstrated its ability to improve the match between distorted images compared to other descriptors in the literature. 展开更多
关键词 keypoint detection DESCRIPTORS neighbor region similarity invariance
在线阅读 下载PDF
Copy-Move Forgeries Detection and Localization Using Two Levels of Keypoints Extraction 被引量:1
4
作者 Soad Samir Eid Emary +1 位作者 Khaled Elsayed Hoda Onsi 《Journal of Computer and Communications》 2019年第9期1-18,共18页
Copy-move offense is considerably used to conceal or hide several data in the digital image for specific aim, and onto this offense some portion of the genuine image is reduplicated and pasted in the same image. There... Copy-move offense is considerably used to conceal or hide several data in the digital image for specific aim, and onto this offense some portion of the genuine image is reduplicated and pasted in the same image. Therefore, Copy-Move forgery is a very significant problem and active research area to check the confirmation of the image. In this paper, a system for Copy Move Forgery detection is proposed. The proposed system is composed of two stages: one is called the detection stages and the second is called the refine detection stage. The detection stage is executed using Speeded-Up Robust Feature (SURF) and Binary Robust Invariant Scalable Keypoints (BRISK) for feature detection and in the refine detection stage, image registration using non-linear transformation is used to enhance detection efficiency. Initially, the genuine image is picked, and then both SURF and BRISK feature extractions are used in parallel to detect the interest keypoints. This gives an appropriate number of interest points and gives the assurance for finding the majority of the manipulated regions. RANSAC is employed to find the superior group of matches to differentiate the manipulated parts. Then, non-linear transformation between the best-matched sets from both extraction features is used as an optimization to get the best-matched set and detect the copied regions. A number of numerical experiments performed using many benchmark datasets such as, the CASIA v2.0, MICC-220, MICC-F600 and MICC-F2000 datasets. With the proposed algorithm, an overall average detection accuracy of 95.33% is obtained for evaluation carried out with the aforementioned databases. Forgery detection achieved True Positive Rate of 97.4% for tampered images with object translation, different degree of rotation and enlargement. Thus, results from different datasets have been set, proving that the proposed algorithm can individuate the altered areas, with high reliability and dealing with multiple cloning. 展开更多
关键词 COPY MOVE FORGERY DETECTION keypoint Based Methods SURF BRISK Bi-Cubic Interpolation
暂未订购
Keypoints and Descriptors Based on Cross-Modality Information Fusion for Camera Localization
5
作者 MA Shuo GAO Yongbin+ +4 位作者 TIAN Fangzheng LU Junxin HUANG Bo GU Jia ZHOU Yilong 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2021年第2期128-136,共9页
To address the problem that traditional keypoint detection methods are susceptible to complex backgrounds and local similarity of images resulting in inaccurate descriptor matching and bias in visual localization, key... To address the problem that traditional keypoint detection methods are susceptible to complex backgrounds and local similarity of images resulting in inaccurate descriptor matching and bias in visual localization, keypoints and descriptors based on cross-modality fusion are proposed and applied to the study of camera motion estimation. A convolutional neural network is used to detect the positions of keypoints and generate the corresponding descriptors, and the pyramid convolution is used to extract multi-scale features in the network. The problem of local similarity of images is solved by capturing local and global feature information and fusing the geometric position information of keypoints to generate descriptors. According to our experiments, the repeatability of our method is improved by 3.7%, and the homography estimation is improved by 1.6%. To demonstrate the practicability of the method, the visual odometry part of simultaneous localization and mapping is constructed and our method is 35% higher positioning accuracy than the traditional method. 展开更多
关键词 keypoints DESCRIPTORS cross-modality information global feature visual odometry
原文传递
Two-Fold and Symmetric Repeatability Rates for Comparing Keypoint Detectors
6
作者 Ibrahim El rube’ 《Computers, Materials & Continua》 SCIE EI 2022年第12期6495-6511,共17页
The repeatability rate is an important measure for evaluating and comparing the performance of keypoint detectors.Several repeatability rate measurementswere used in the literature to assess the effectiveness of keypo... The repeatability rate is an important measure for evaluating and comparing the performance of keypoint detectors.Several repeatability rate measurementswere used in the literature to assess the effectiveness of keypoint detectors.While these repeatability rates are calculated for pairs of images,the general assumption is that the reference image is often known and unchanging compared to other images in the same dataset.So,these rates are asymmetrical as they require calculations in only one direction.In addition,the image domain in which these computations take place substantially affects their values.The presented scatter diagram plots illustrate how these directional repeatability rates vary in relation to the size of the neighboring region in each pair of images.Therefore,both directional repeatability rates for the same image pair must be included when comparing different keypoint detectors.This paper,firstly,examines several commonly utilized repeatability rate measures for keypoint detector evaluations.The researcher then suggests computing a two-fold repeatability rate to assess keypoint detector performance on similar scene images.Next,the symmetric mean repeatability rate metric is computed using the given two-fold repeatability rates.Finally,these measurements are validated using well-known keypoint detectors on different image groups with various geometric and photometric attributes. 展开更多
关键词 Repeatability rate keypoint detector symmetric measure geometric transformation scatter diagram
在线阅读 下载PDF
Multi-Level Feature Aggregation-Based Joint Keypoint Detection and Description
7
作者 Jun Li Xiang Li +2 位作者 Yifei Wei Mei Song Xiaojun Wang 《Computers, Materials & Continua》 SCIE EI 2022年第11期2529-2540,共12页
Image keypoint detection and description is a popular method to find pixel-level connections between images,which is a basic and critical step in many computer vision tasks.The existing methods are far from optimal in... Image keypoint detection and description is a popular method to find pixel-level connections between images,which is a basic and critical step in many computer vision tasks.The existing methods are far from optimal in terms of keypoint positioning accuracy and generation of robust and discriminative descriptors.This paper proposes a new end-to-end selfsupervised training deep learning network.The network uses a backbone feature encoder to extract multi-level feature maps,then performs joint image keypoint detection and description in a forward pass.On the one hand,in order to enhance the localization accuracy of keypoints and restore the local shape structure,the detector detects keypoints on feature maps of the same resolution as the original image.On the other hand,in order to enhance the ability to percept local shape details,the network utilizes multi-level features to generate robust feature descriptors with rich local shape information.A detailed comparison with traditional feature-based methods Scale Invariant Feature Transform(SIFT),Speeded Up Robust Features(SURF)and deep learning methods on HPatches proves the effectiveness and robustness of the method proposed in this paper. 展开更多
关键词 Multi-scale information keypoint detection and description artificial intelligence
在线阅读 下载PDF
Clothes Keypoints Detection with Cascaded Pyramid Network
8
作者 LI Chao ZHAO Mingbo 《Journal of Donghua University(English Edition)》 EI CAS 2020年第3期232-237,共6页
With the development of the society,people's requirements for clothing matching are constantly increasing when developing clothing recommendation system.This requires that the algorithm for understanding the cloth... With the development of the society,people's requirements for clothing matching are constantly increasing when developing clothing recommendation system.This requires that the algorithm for understanding the clothing images should be sufficiently efficient and robust.Therefore,we detect the keypoints in clothing accurately to capture the details of clothing images.Since the joint points of the garment are similar to those of the human body,this paper utilizes a kind of deep neural network called cascaded pyramid network(CPN)about estimating the posture of human body to solve the problem of keypoints detection in clothing.In this paper,we first introduce the structure and characteristic of this neural network when detecting keypoints.Then we evaluate the results of the experiments and verify effectiveness of detecting keypoints of clothing with CPN,with normalized error about 5%7%.Finally,we analyze the influence of different backbones when detecting keypoints in this network. 展开更多
关键词 deep learning keypoints estimation convolutional neural network
在线阅读 下载PDF
Skeleton Keypoints Extraction Method Combined with Object Detection
9
作者 Jiabao Shi Zhao Qiu +4 位作者 Tao Chen Jiale Lin Hancheng Huang Yunlong He d Yu Yang 《Journal of New Media》 2022年第2期97-106,共10页
Big data is a comprehensive result of the development of the Internet of Things and information systems.Computer vision requires a lot of data as the basis for research.Because skeleton data can adapt well to dynamic ... Big data is a comprehensive result of the development of the Internet of Things and information systems.Computer vision requires a lot of data as the basis for research.Because skeleton data can adapt well to dynamic environment and complex background,it is used in action recognition tasks.In recent years,skeleton-based action recognition has received more and more attention in the field of computer vision.Therefore,the keypoints of human skeletons are essential for describing the pose estimation of human and predicting the action recognition of the human.This paper proposes a skeleton point extraction method combined with object detection,which can focus on the extraction of skeleton keypoints.After a large number of experiments,our model can be combined with object detection for skeleton points extraction,and the detection efficiency is improved. 展开更多
关键词 Big data object decetion skeleton keypoints lightweight openpose
在线阅读 下载PDF
A keypoint-based method for detecting weed growth points in corn field environments
10
作者 Mochen Liu Xiaoli Xu +4 位作者 Tingdong Tian Mingrui Shang Zhanhua Song Fuyang Tian Yinfa Yan 《Plant Phenomics》 2025年第3期103-116,共14页
Weed growth significantly impacts corn yield.With the continuous development of weed control technologies,achieving more effective and precise weed management has become a major challenge in corn production.To achieve... Weed growth significantly impacts corn yield.With the continuous development of weed control technologies,achieving more effective and precise weed management has become a major challenge in corn production.To achieve precise weed suppression,this study proposes a growth point detection method based on a keypoint pose estimation model capable of effectively detecting various weeds and locating various weed growth points during the 2nd-5th leaf stage of corn development.To address the complex working environment of precision weeding machines in corn fields,including occlusion,dense growth,and variable lighting conditions,we design a dilation-wise residual module(DWRM)for the detector and a separation and enhancement attention module(SEAM)for pose estimation to adapt to these challenges.Furthermore,owing to the limited computational re-sources in field settings,we introduced the RepViT block(RVB)to achieve model lightweighting.The proposed method was evaluated on the constructed corn field dataset.The experimental results demonstrated that SRD-YOLO achieved an mAPkpt of 96.5%,an Fl score of 94%,and an FPS of 169,while reducing the model pa-rameters by 8.7M.SRD-YOLO effectively meets the requirements for growth point localization under challenging conditions,providing robust technical support for real-time and precise weed control in corn fields. 展开更多
关键词 Weed detection keypoints Corn seedlings Growth points Precision weeding
原文传递
基于多尺度信息的生成式人体姿态估计
11
作者 陈俊芬 冯武山 +1 位作者 郝旭阳 谢博鋆 《计算机工程与应用》 北大核心 2026年第3期265-276,共12页
针对人体姿态估计中遮挡带来的缺乏图像低级特征指导和预测姿势与人体生理结构的不一致性问题,提出了一种新颖的生成式人体姿态估计方法(generative human pose estimation,GenPose)。该模型使用多尺度信息融合和条件生成模块解决了严... 针对人体姿态估计中遮挡带来的缺乏图像低级特征指导和预测姿势与人体生理结构的不一致性问题,提出了一种新颖的生成式人体姿态估计方法(generative human pose estimation,GenPose)。该模型使用多尺度信息融合和条件生成模块解决了严重遮挡问题。多尺度模块从尺度和通道上细粒度融合图像特征,能捕捉到更多肢体细节,从而推理出遮挡关键点的特征信息。条件生成模块通过建模遮挡场景与姿态间的对应关系,根据标记编码器特征动态调整生成姿态,在保证可见点准确率的同时,在一定程度上减少了遮挡对非遮挡的干扰,提升了对遮挡姿态的生成效果。在公开的COCO和MPII数据集上,同以往方法相比,有了更好的结果,同时在CrowdPose、OCHuman以及SyncOCC数据集上验证了泛化能力。该模型在一定程度上能够解决严重遮挡下的姿态估计问题,提高了预测姿态的合理性,取得了更加优异的效果。 展开更多
关键词 人体姿态估计 不可见关键点 严重遮挡 注意力机制 变分编码器
在线阅读 下载PDF
Application of the Improved PF-Flow-Style-VTON in Virtual Try-On
12
作者 TIAN Jiajia HUANG Rong +1 位作者 DONG Aihua WANG Zhijie 《Journal of Donghua University(English Edition)》 2026年第1期104-117,共14页
During the image generation phase,the parserfree Flow-Style-VTON model(PF-Flow-Style-VTON),which utilizes distilled appearance flows,faces two main challenges:blurring,deformation,occlusion,or loss of the arm or palm ... During the image generation phase,the parserfree Flow-Style-VTON model(PF-Flow-Style-VTON),which utilizes distilled appearance flows,faces two main challenges:blurring,deformation,occlusion,or loss of the arm or palm regions in the generated image when these regions of the person occlude the garment;blurring and deformation in the generated image when the person performs large pose movements and the target garment is complex with detailed patterns.To solve these two problems,an improved virtual try-on network model,denoted as IPF-Flow-Style-VTON,is proposed.Firstly,a target warped garment mask refinement module(M-RM)is introduced to refine the warped garment mask and remove erroneous information in the arm and palm regions,thereby improving the quality of subsequent image generation.Secondly,an improved global attention module(GAM)is integrated into the original image generation network,enhancing the ResUNet’s understanding of global context and optimizing the fusion of local features and global information,thereby further improving image generation quality.Finally,the UniPose model is used to provide the pose keypoint information of the target person image,guiding the task execution during the image generation phase.Experiments conducted on the VITON dataset show that the proposed method outperforms the original method,Flow-Style-VTON,by 5.4%,0.3%,6.7%,and 2.2%in Frchet inception distance(FID),structural similarity index measure(SSIM),learned perceptual image patch similarity(LPIPS),and peak signal-to-noise ratio(PSNR),respectively.Overall,the proposed method effectively improves upon the shortcomings of the original network and achieves better visual results. 展开更多
关键词 virtual try-on image generation network pose keypoint deep learning
在线阅读 下载PDF
Intelligent Human Interaction Recognition with Multi-Modal Feature Extraction and Bidirectional LSTM
13
作者 Muhammad Hamdan Azhar Yanfeng Wu +4 位作者 Nouf Abdullah Almujally Shuaa S.Alharbi Asaad Algarni Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2026年第4期1632-1649,共18页
Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationall... Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationally intensive,sensitive to video resolution changes and often fail in crowded scenes.We propose a novel hybrid system that is computationally efficient,robust to degraded video quality and able to filter out irrelevant individuals,making it suitable for real-life use.The system leverages multi-modal handcrafted features for interaction representation and a deep learning classifier for capturing complex dependencies.Using Mask R-CNN and YOLO11-Pose,we extract grayscale silhouettes and keypoint coordinates of interacting individuals,while filtering out irrelevant individuals using a proposed algorithm.From these,we extract silhouette-based features(local ternary pattern and histogram of optical flow)and keypoint-based features(distances,angles and velocities)that capture distinct spatial and temporal information.A Bidirectional Long Short-Term Memory network(BiLSTM)then classifies the interactions.Extensive experiments on the UT Interaction,SBU Kinect Interaction and the ISR-UOL 3D social activity datasets demonstrate that our system achieves competitive accuracy.They also validate the effectiveness of the chosen features and classifier,along with the proposed system’s computational efficiency and robustness to occlusion. 展开更多
关键词 Human interaction recognition keypoint coordinates grayscale silhouettes bidirectional long shortterm memory network
在线阅读 下载PDF
基于视觉的无标记运动学分析方法
14
作者 黄高华 李玉榕 +1 位作者 姜海燕 陈建国 《仪器仪表学报》 北大核心 2026年第1期312-324,共13页
针对基于标记的光学系统成本高、耗时长、专业性强的问题,提出了一种采用两个视角的视觉无标记运动学分析方法,旨在实现便捷、低成本的运动学评估。首先,集成Swin Transformer的全局上下文建模能力、坐标注意力的精准位置感知能力、双... 针对基于标记的光学系统成本高、耗时长、专业性强的问题,提出了一种采用两个视角的视觉无标记运动学分析方法,旨在实现便捷、低成本的运动学评估。首先,集成Swin Transformer的全局上下文建模能力、坐标注意力的精准位置感知能力、双向特征金字塔网络的多尺度特征融合能力,构建二维特征提取架构,克服自遮挡、关键点小目标检测问题,有效提取二维特征。其次,提出以关键点位置合理性和肢体长度一致性为关节上下文约束的三角测量方法,并结合人体参数化模型对三维关键点进行重构,提高关键点估计精度。最后,构建关键点增强模型,获取解剖标记集并结合肌骨模型进行运动学分析。公开数据集上的运动学评估实验表明,所提方法的平均关节角度误差为8.59°,平均关节位置误差为42.02 mm,优于现有的高性能方法。同时,为验证方法在真实场景下的适用性,以商用动作捕捉系统Xsens作为评估标准,并与当前主流方法OpenCap进行比较,分别对肩关节和步态运动学展开分析。实验结果表明,在肩关节和步态运动学上,所提方法与Xsens的相关系数分别为0.92和0.86,较OpenCap的相关系数分别提高9.52%和7.40%;角度误差分别为13.97°和3.12°,较OpenCap的误差分别下降27.01%和25.18%。综上所述,在公开数据集和真实场景下,所提方法可实现比当前主流方法更准确的运动学分析,对促进运动学分析相关应用的推广具有重要意义。 展开更多
关键词 视觉 无标记 三维关键点估计 关键点增强 肌骨模型 运动学分析
原文传递
基于YOLO-FMC-pose的中华绒螯蟹头胸甲关键点检测方法
15
作者 张哲 于合龙 +3 位作者 杨信廷 罗娜 李珊珊 孙传恒 《农业工程学报》 北大核心 2026年第1期210-221,共12页
中华绒螯蟹(Eriocheir sinensis)的头胸甲形态在同一物种的不同个体之间表现出明显差异,这一特征可作为产地溯源和个体识别的重要依据。其中,头胸甲关键点的精准检测是实现个体识别与表型分析等任务的基础环节。然而,传统的人工检测方... 中华绒螯蟹(Eriocheir sinensis)的头胸甲形态在同一物种的不同个体之间表现出明显差异,这一特征可作为产地溯源和个体识别的重要依据。其中,头胸甲关键点的精准检测是实现个体识别与表型分析等任务的基础环节。然而,传统的人工检测方法依赖经验性判断,存在效率低、重复性差等问题,难以满足规模化水产处理的实际需求。为此,该研究提出了一种基于YOLO-FMC-pose的中华绒螯蟹头胸甲关键点自动检测方法,以实现高精度、自动化的特征提取。首先,构建了一个包含大量中华绒螯蟹头胸甲图像的自建数据集,并选取具有代表性的35个地标关键点进行精确标注,同时通过数据增强提升模型的训练效果。其次,该研究基于改进的YOLO11n-pose框架设计了中华绒螯蟹头胸甲关键点检测模型YOLO-FMC-pose。模型中引入了融合频率动态卷积(FDConv)的C3K2FD模块、混合聚合网络(MANet)模块以及CBAM注意力机制,从频域响应、特征融合与空间关注等层面对结构进行了优化。结果表明,所提出的YOLO-FMC-pose模型在关键点检测精度方面均优于现有主流方法,准确率、召回率、mAP_(0.5)和m AP_(0.5:0.95)分别为97.98%、97.00%、98.27%和73.28%,相较于原始YOLO11n-pose,准确率、召回率、mAP_(0.5)和mAP_(0.5:0.95)分别提高了3.33、2.33、2.94和13.08个百分点,标准化平均误差(normalized mean error,NME)降低至3.835%,单帧图片推理时间为7.5 ms,具备良好的实际应用潜力。该研究为中华绒螯蟹的个体智能识别、产地溯源与防伪管控提供了关键技术支撑,也为水产品精细化特征检测提供了路径。 展开更多
关键词 中华绒螯蟹 关键点检测 表型特征识别 深度学习 图像处理
在线阅读 下载PDF
基于扩散模型多模态提示的电力人员行为图像生成
16
作者 朱志航 闫云凤 齐冬莲 《浙江大学学报(工学版)》 北大核心 2026年第1期43-51,70,共10页
电力人员行为的特殊性与复杂性导致其图像数据稀缺,给数据驱动下的行为识别带来了挑战.在稳定扩散模型的基础上,充分融合人体骨架、掩膜以及文本描述信息,加入关键点损失函数,建立多模态条件控制的电力人员行为图像生成模型PoseNet,该... 电力人员行为的特殊性与复杂性导致其图像数据稀缺,给数据驱动下的行为识别带来了挑战.在稳定扩散模型的基础上,充分融合人体骨架、掩膜以及文本描述信息,加入关键点损失函数,建立多模态条件控制的电力人员行为图像生成模型PoseNet,该模型可以生成高质量的可控人体图像.设计基于关键点相似度的图像滤波器,以去除错误、低质量的生成图像;采用双阶段训练策略,在通用数据上对模型进行预训练,并在私有数据上微调,提升模型性能;针对电力人员行为特点,设计集通用、专用评价指标于一体的生成图像评价指标集,分析不同评价指标下的图像生成效果.实验结果表明,与主流人体生成模型ControlNet、HumanSD相比,该模型的生成结果更精准、真实、效果更优. 展开更多
关键词 条件图像生成模型 数据扩充 人体关键点 图像分割 扩散模型 深度学习
在线阅读 下载PDF
基于深度学习的服装关键点实时检测模型
17
作者 冯采伶 于施佳 韩曙光 《纺织学报》 北大核心 2026年第1期196-206,共11页
针对复杂场景下服装关键点检测模型的准确性与实时性难以兼得的问题,以提高检测准确性并保持实时性能为目的,提出了一种基于深度学习的服装关键点实时检测模型。该模型以实时多人姿态估计架构为基础,首先构建中值增强的通道与空间注意... 针对复杂场景下服装关键点检测模型的准确性与实时性难以兼得的问题,以提高检测准确性并保持实时性能为目的,提出了一种基于深度学习的服装关键点实时检测模型。该模型以实时多人姿态估计架构为基础,首先构建中值增强的通道与空间注意力模块,通过并行执行全局平均池化、最大池化与中值池化,融合生成注意力权重,增强服装关键部位的特征表示;其次设计跨尺度特征融合模块,将骨干网络中不同层级的特征图进行上采样、拼接与交叉卷积融合,构建兼具细节信息与语义特征的金字塔结构;进一步建立自注意力特征增强模块,通过计算特征点间相似性动态生成注意力图,自适应调整各区域特征权重;最终实施分类别微调策略,针对6类典型服装分别建立专用模型以优化整体性能。结果表明:该方法在DeepFashion2和DeepFashion数据集上分别达到了65.1%与68.0%的检测准确度,同时保持140.0帧/s和142.3帧/s的实时处理速度。该模型提升了复杂场景下服装关键点检测的综合性能,未来可应用于服装智能制造和虚拟试衣等领域。 展开更多
关键词 服装关键点 实时检测 深度学习 自注意力机制 跨尺度特征融合
在线阅读 下载PDF
基于多视角融合的机器人抓取位姿估计模型
18
作者 杨红玲 代路 解仑 《科学技术与工程》 北大核心 2026年第9期3616-3625,共10页
为解决机械臂抓取时因物体对称性和遮挡等因素引发的位姿估计难题,通过融合多视角RGB-D数据的方法研究了目标物体的六自由度位姿估计问题。该方法基于单视角位姿估计框架,利用三维关键点投票网络对各视角场景中的三维关键点进行独立预测... 为解决机械臂抓取时因物体对称性和遮挡等因素引发的位姿估计难题,通过融合多视角RGB-D数据的方法研究了目标物体的六自由度位姿估计问题。该方法基于单视角位姿估计框架,利用三维关键点投票网络对各视角场景中的三维关键点进行独立预测,并通过置信度评分、加权融合以及均值偏移聚类实现多视角信息的有效整合,最终完成目标物体位姿的精确求解。结果表明:该方法在对称与遮挡场景下能够显著提升位姿估计精度;多视角融合策略有效增强了关键点的预测稳定性与鲁棒性;在桌面环境下的自主抓取实验中显著提高了抓取成功率与执行效果。实验证明,所提出的方法在复杂抓取场景中具有良好的适应性和应用价值,能有效提高在非结构化场景下的作业自主性,还为解决复杂工况下的智能装配作业提供了关键技术支撑,具有广阔的产业应用前景。 展开更多
关键词 场景融合 位姿估计 均值偏移 关键点 多视角 机器人抓取
在线阅读 下载PDF
基于改进YOLOv8n-pose的巨峰葡萄采摘定位方法
19
作者 陈馨 吴子炜 +1 位作者 周素茵 夏芳 《华南农业大学学报》 北大核心 2026年第1期118-127,共10页
【目的】对巨峰葡萄进行精准高效地采摘定位,以有效降低果实损伤。【方法】提出一种基于改进YOLOv8n-pose的葡萄采摘定位方法。首先,利用改进YOLOv8n-pose检测葡萄果梗和顶部易损果粒的关键点,基于关键点的坐标构建果实上界位姿的表征向... 【目的】对巨峰葡萄进行精准高效地采摘定位,以有效降低果实损伤。【方法】提出一种基于改进YOLOv8n-pose的葡萄采摘定位方法。首先,利用改进YOLOv8n-pose检测葡萄果梗和顶部易损果粒的关键点,基于关键点的坐标构建果实上界位姿的表征向量;然后,利用此向量计算出最优采摘角度;最后,通过将采摘点与采摘角协同,确定最佳采摘位置。【结果】试验结果表明,改进后YOLOv8n-pose的P、R、mAP@0.50、mAP@0.50~0.95较原模型分别提升了1.7、0.7、0.9、1.7个百分点,较YOLOv12s-pose分别提升了0.4、0.1、0.6、2.7个百分点,同时模型参数量比YOLOv8n-pose减少了5.8%。应用本文方法的葡萄采摘定位成功率为90.8%,相较于不使用采摘角的定位方法,提升了9.2个百分点。【结论】研究为巨峰葡萄采摘机器人提供了一种低损定位方法。 展开更多
关键词 巨峰葡萄 YOLOv8-pose 关键点检测 采摘定位 采摘角度
在线阅读 下载PDF
基于改进RTMPose3D模型的番茄三维关键点估计方法
20
作者 王蓬勃 刘宇 +1 位作者 赵胜辉 傅毅凯 《农业机械学报》 北大核心 2026年第5期149-158,共10页
针对温室环境下串番茄枝叶遮挡严重、光照干扰强,导致自动采摘机器人难以稳定获取果实三维位姿的问题,提出了一种改进的串番茄三维关键点估计模型TomatoPose3D。该模型在训练阶段引入RGB图像与三维真值关键点的联合约束,增强结构一致性... 针对温室环境下串番茄枝叶遮挡严重、光照干扰强,导致自动采摘机器人难以稳定获取果实三维位姿的问题,提出了一种改进的串番茄三维关键点估计模型TomatoPose3D。该模型在训练阶段引入RGB图像与三维真值关键点的联合约束,增强结构一致性与泛化能力;在推理阶段,仅输入单幅RGB图像即可端到端回归三维关键点坐标,避免了因点云缺失或稀疏导致的定位失败。改进模型以RTMPose3D为基线,引入全局结构感知模块MobileVit Block与分布感知解码策略(DARK),在保持轻量化的同时提升了定位精度。温室场景对比实验表明,TomatoPose3D的PCK@0.05指标较RTMPose3D和SimpleBaseline3D分别提升5.18、9.98个百分点;在无深度信息辅助下,其定位精度与基于RGB-D投影的方法相当,且鲁棒性更优。此外,模型经TensorRT加速部署于工业级嵌入式平台,端到端推理速度达37 f/s,满足采摘机器人实时空间视觉感知需求。 展开更多
关键词 温室番茄 采摘机器人 三维关键点估计 轻量化网络 RTMPose3D
在线阅读 下载PDF
上一页 1 2 20 下一页 到第
使用帮助 返回顶部