期刊文献+
共找到365篇文章
< 1 2 19 >
每页显示 20 50 100
Keypoint全功能肌电诱发电位故障维修1例
1
作者 晋虎 《医疗卫生装备》 CAS 2012年第8期145-146,共2页
Keypoint全功能肌电诱发电位是维迪公司的一款性能稳定可靠、使用便捷的台式肌电图,在使用中轻轻点击快速完成数据采集即可将患者从痛苦的检查中解放出来。Keypoint具有保留所有原始的波形提供给医师在报告时阅读分析参考诊断。
关键词 keypoint 肌电诱发电位 全功能 故障维修 数据采集 肌电图
暂未订购
基于Keypoint RCNN改进模型的物体抓取检测算法 被引量:14
2
作者 夏浩宇 索双富 +2 位作者 王洋 安琪 张妙恬 《仪器仪表学报》 EI CAS CSCD 北大核心 2021年第4期236-246,共11页
机器人抓取在工业中的应用有两个难点:如何准确地检测可抓取物体,以及如何从检测出的多个物体中选择最优抓取目标。本文在Keypoint RCNN模型中引入同方差不确定性学习各损失的权重,并在特征提取器中加入注意力模块,构成了Keypoint RCNN... 机器人抓取在工业中的应用有两个难点:如何准确地检测可抓取物体,以及如何从检测出的多个物体中选择最优抓取目标。本文在Keypoint RCNN模型中引入同方差不确定性学习各损失的权重,并在特征提取器中加入注意力模块,构成了Keypoint RCNN改进模型。基于改进模型提出了两阶段物体抓取检测算法,第一阶段用模型预测物体掩码和关键点,第二阶段用掩码和关键点计算物体的抓取描述和重合度,重合度表示抓取时的碰撞程度,根据重合度可以从多个可抓取物体中选择最优抓取目标。对照实验证明,相较原模型,Keypoint RCNN改进模型在目标检测、实例分割、关键点检测上的性能均有提高,在自建数据集上的平均精度分别为85.15%、79.66%、86.63%,机器人抓取实验证明抓取检测算法能够准确计算物体的抓取描述、选择最优抓取,引导机器人无碰撞地抓取目标。 展开更多
关键词 抓取检测 keypoint RCNN改进模型 损失权重 注意力模块 抓取描述 重合度 最优抓取
原文传递
Keypoint Description Using Statistical Descriptor with Similarity-Invariant Regions 被引量:2
3
作者 Ibrahim El rube Sameer Alsharif 《Computer Systems Science & Engineering》 SCIE EI 2022年第7期407-421,共15页
This article presents a method for the description of key points using simple statistics for regions controlled by neighboring key points to remedy the gap in existing descriptors.Usually,the existent descriptors such... This article presents a method for the description of key points using simple statistics for regions controlled by neighboring key points to remedy the gap in existing descriptors.Usually,the existent descriptors such as speeded up robust features(SURF),Kaze,binary robust invariant scalable keypoints(BRISK),features from accelerated segment test(FAST),and oriented FAST and rotated BRIEF(ORB)can competently detect,describe,and match images in the presence of some artifacts such as blur,compression,and illumination.However,the performance and reliability of these descriptors decrease for some imaging variations such as point of view,zoom(scale),and rotation.The intro-duced description method improves image matching in the event of such distor-tions.It utilizes a contourlet-based detector to detect the strongest key points within a specified window size.The selected key points and their neighbors con-trol the size and orientation of the surrounding regions,which are mapped on rec-tangular shapes using polar transformation.The resulting rectangular matrices are subjected to two-directional statistical operations that involve calculating the mean and standard deviation.Consequently,the descriptor obtained is invariant(translation,rotation,and scale)because of the two methods;the extraction of the region and the polar transformation techniques used in this paper.The descrip-tion method introduced in this article is tested against well-established and well-known descriptors,such as SURF,Kaze,BRISK,FAST,and ORB,techniques using the standard OXFORD dataset.The presented methodology demonstrated its ability to improve the match between distorted images compared to other descriptors in the literature. 展开更多
关键词 keypoint detection DESCRIPTORS neighbor region similarity invariance
在线阅读 下载PDF
Copy-Move Forgeries Detection and Localization Using Two Levels of Keypoints Extraction 被引量:1
4
作者 Soad Samir Eid Emary +1 位作者 Khaled Elsayed Hoda Onsi 《Journal of Computer and Communications》 2019年第9期1-18,共18页
Copy-move offense is considerably used to conceal or hide several data in the digital image for specific aim, and onto this offense some portion of the genuine image is reduplicated and pasted in the same image. There... Copy-move offense is considerably used to conceal or hide several data in the digital image for specific aim, and onto this offense some portion of the genuine image is reduplicated and pasted in the same image. Therefore, Copy-Move forgery is a very significant problem and active research area to check the confirmation of the image. In this paper, a system for Copy Move Forgery detection is proposed. The proposed system is composed of two stages: one is called the detection stages and the second is called the refine detection stage. The detection stage is executed using Speeded-Up Robust Feature (SURF) and Binary Robust Invariant Scalable Keypoints (BRISK) for feature detection and in the refine detection stage, image registration using non-linear transformation is used to enhance detection efficiency. Initially, the genuine image is picked, and then both SURF and BRISK feature extractions are used in parallel to detect the interest keypoints. This gives an appropriate number of interest points and gives the assurance for finding the majority of the manipulated regions. RANSAC is employed to find the superior group of matches to differentiate the manipulated parts. Then, non-linear transformation between the best-matched sets from both extraction features is used as an optimization to get the best-matched set and detect the copied regions. A number of numerical experiments performed using many benchmark datasets such as, the CASIA v2.0, MICC-220, MICC-F600 and MICC-F2000 datasets. With the proposed algorithm, an overall average detection accuracy of 95.33% is obtained for evaluation carried out with the aforementioned databases. Forgery detection achieved True Positive Rate of 97.4% for tampered images with object translation, different degree of rotation and enlargement. Thus, results from different datasets have been set, proving that the proposed algorithm can individuate the altered areas, with high reliability and dealing with multiple cloning. 展开更多
关键词 COPY MOVE FORGERY DETECTION keypoint Based Methods SURF BRISK Bi-Cubic Interpolation
暂未订购
Keypoints and Descriptors Based on Cross-Modality Information Fusion for Camera Localization
5
作者 MA Shuo GAO Yongbin+ +4 位作者 TIAN Fangzheng LU Junxin HUANG Bo GU Jia ZHOU Yilong 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2021年第2期128-136,共9页
To address the problem that traditional keypoint detection methods are susceptible to complex backgrounds and local similarity of images resulting in inaccurate descriptor matching and bias in visual localization, key... To address the problem that traditional keypoint detection methods are susceptible to complex backgrounds and local similarity of images resulting in inaccurate descriptor matching and bias in visual localization, keypoints and descriptors based on cross-modality fusion are proposed and applied to the study of camera motion estimation. A convolutional neural network is used to detect the positions of keypoints and generate the corresponding descriptors, and the pyramid convolution is used to extract multi-scale features in the network. The problem of local similarity of images is solved by capturing local and global feature information and fusing the geometric position information of keypoints to generate descriptors. According to our experiments, the repeatability of our method is improved by 3.7%, and the homography estimation is improved by 1.6%. To demonstrate the practicability of the method, the visual odometry part of simultaneous localization and mapping is constructed and our method is 35% higher positioning accuracy than the traditional method. 展开更多
关键词 keypoints DESCRIPTORS cross-modality information global feature visual odometry
原文传递
Two-Fold and Symmetric Repeatability Rates for Comparing Keypoint Detectors
6
作者 Ibrahim El rube’ 《Computers, Materials & Continua》 SCIE EI 2022年第12期6495-6511,共17页
The repeatability rate is an important measure for evaluating and comparing the performance of keypoint detectors.Several repeatability rate measurementswere used in the literature to assess the effectiveness of keypo... The repeatability rate is an important measure for evaluating and comparing the performance of keypoint detectors.Several repeatability rate measurementswere used in the literature to assess the effectiveness of keypoint detectors.While these repeatability rates are calculated for pairs of images,the general assumption is that the reference image is often known and unchanging compared to other images in the same dataset.So,these rates are asymmetrical as they require calculations in only one direction.In addition,the image domain in which these computations take place substantially affects their values.The presented scatter diagram plots illustrate how these directional repeatability rates vary in relation to the size of the neighboring region in each pair of images.Therefore,both directional repeatability rates for the same image pair must be included when comparing different keypoint detectors.This paper,firstly,examines several commonly utilized repeatability rate measures for keypoint detector evaluations.The researcher then suggests computing a two-fold repeatability rate to assess keypoint detector performance on similar scene images.Next,the symmetric mean repeatability rate metric is computed using the given two-fold repeatability rates.Finally,these measurements are validated using well-known keypoint detectors on different image groups with various geometric and photometric attributes. 展开更多
关键词 Repeatability rate keypoint detector symmetric measure geometric transformation scatter diagram
在线阅读 下载PDF
Multi-Level Feature Aggregation-Based Joint Keypoint Detection and Description
7
作者 Jun Li Xiang Li +2 位作者 Yifei Wei Mei Song Xiaojun Wang 《Computers, Materials & Continua》 SCIE EI 2022年第11期2529-2540,共12页
Image keypoint detection and description is a popular method to find pixel-level connections between images,which is a basic and critical step in many computer vision tasks.The existing methods are far from optimal in... Image keypoint detection and description is a popular method to find pixel-level connections between images,which is a basic and critical step in many computer vision tasks.The existing methods are far from optimal in terms of keypoint positioning accuracy and generation of robust and discriminative descriptors.This paper proposes a new end-to-end selfsupervised training deep learning network.The network uses a backbone feature encoder to extract multi-level feature maps,then performs joint image keypoint detection and description in a forward pass.On the one hand,in order to enhance the localization accuracy of keypoints and restore the local shape structure,the detector detects keypoints on feature maps of the same resolution as the original image.On the other hand,in order to enhance the ability to percept local shape details,the network utilizes multi-level features to generate robust feature descriptors with rich local shape information.A detailed comparison with traditional feature-based methods Scale Invariant Feature Transform(SIFT),Speeded Up Robust Features(SURF)and deep learning methods on HPatches proves the effectiveness and robustness of the method proposed in this paper. 展开更多
关键词 Multi-scale information keypoint detection and description artificial intelligence
在线阅读 下载PDF
Clothes Keypoints Detection with Cascaded Pyramid Network
8
作者 LI Chao ZHAO Mingbo 《Journal of Donghua University(English Edition)》 EI CAS 2020年第3期232-237,共6页
With the development of the society,people's requirements for clothing matching are constantly increasing when developing clothing recommendation system.This requires that the algorithm for understanding the cloth... With the development of the society,people's requirements for clothing matching are constantly increasing when developing clothing recommendation system.This requires that the algorithm for understanding the clothing images should be sufficiently efficient and robust.Therefore,we detect the keypoints in clothing accurately to capture the details of clothing images.Since the joint points of the garment are similar to those of the human body,this paper utilizes a kind of deep neural network called cascaded pyramid network(CPN)about estimating the posture of human body to solve the problem of keypoints detection in clothing.In this paper,we first introduce the structure and characteristic of this neural network when detecting keypoints.Then we evaluate the results of the experiments and verify effectiveness of detecting keypoints of clothing with CPN,with normalized error about 5%7%.Finally,we analyze the influence of different backbones when detecting keypoints in this network. 展开更多
关键词 deep learning keypoints estimation convolutional neural network
在线阅读 下载PDF
Skeleton Keypoints Extraction Method Combined with Object Detection
9
作者 Jiabao Shi Zhao Qiu +4 位作者 Tao Chen Jiale Lin Hancheng Huang Yunlong He d Yu Yang 《Journal of New Media》 2022年第2期97-106,共10页
Big data is a comprehensive result of the development of the Internet of Things and information systems.Computer vision requires a lot of data as the basis for research.Because skeleton data can adapt well to dynamic ... Big data is a comprehensive result of the development of the Internet of Things and information systems.Computer vision requires a lot of data as the basis for research.Because skeleton data can adapt well to dynamic environment and complex background,it is used in action recognition tasks.In recent years,skeleton-based action recognition has received more and more attention in the field of computer vision.Therefore,the keypoints of human skeletons are essential for describing the pose estimation of human and predicting the action recognition of the human.This paper proposes a skeleton point extraction method combined with object detection,which can focus on the extraction of skeleton keypoints.After a large number of experiments,our model can be combined with object detection for skeleton points extraction,and the detection efficiency is improved. 展开更多
关键词 Big data object decetion skeleton keypoints lightweight openpose
在线阅读 下载PDF
A keypoint-based method for detecting weed growth points in corn field environments
10
作者 Mochen Liu Xiaoli Xu +4 位作者 Tingdong Tian Mingrui Shang Zhanhua Song Fuyang Tian Yinfa Yan 《Plant Phenomics》 2025年第3期103-116,共14页
Weed growth significantly impacts corn yield.With the continuous development of weed control technologies,achieving more effective and precise weed management has become a major challenge in corn production.To achieve... Weed growth significantly impacts corn yield.With the continuous development of weed control technologies,achieving more effective and precise weed management has become a major challenge in corn production.To achieve precise weed suppression,this study proposes a growth point detection method based on a keypoint pose estimation model capable of effectively detecting various weeds and locating various weed growth points during the 2nd-5th leaf stage of corn development.To address the complex working environment of precision weeding machines in corn fields,including occlusion,dense growth,and variable lighting conditions,we design a dilation-wise residual module(DWRM)for the detector and a separation and enhancement attention module(SEAM)for pose estimation to adapt to these challenges.Furthermore,owing to the limited computational re-sources in field settings,we introduced the RepViT block(RVB)to achieve model lightweighting.The proposed method was evaluated on the constructed corn field dataset.The experimental results demonstrated that SRD-YOLO achieved an mAPkpt of 96.5%,an Fl score of 94%,and an FPS of 169,while reducing the model pa-rameters by 8.7M.SRD-YOLO effectively meets the requirements for growth point localization under challenging conditions,providing robust technical support for real-time and precise weed control in corn fields. 展开更多
关键词 Weed detection keypoints Corn seedlings Growth points Precision weeding
原文传递
基于多尺度信息的生成式人体姿态估计
11
作者 陈俊芬 冯武山 +1 位作者 郝旭阳 谢博鋆 《计算机工程与应用》 北大核心 2026年第3期265-276,共12页
针对人体姿态估计中遮挡带来的缺乏图像低级特征指导和预测姿势与人体生理结构的不一致性问题,提出了一种新颖的生成式人体姿态估计方法(generative human pose estimation,GenPose)。该模型使用多尺度信息融合和条件生成模块解决了严... 针对人体姿态估计中遮挡带来的缺乏图像低级特征指导和预测姿势与人体生理结构的不一致性问题,提出了一种新颖的生成式人体姿态估计方法(generative human pose estimation,GenPose)。该模型使用多尺度信息融合和条件生成模块解决了严重遮挡问题。多尺度模块从尺度和通道上细粒度融合图像特征,能捕捉到更多肢体细节,从而推理出遮挡关键点的特征信息。条件生成模块通过建模遮挡场景与姿态间的对应关系,根据标记编码器特征动态调整生成姿态,在保证可见点准确率的同时,在一定程度上减少了遮挡对非遮挡的干扰,提升了对遮挡姿态的生成效果。在公开的COCO和MPII数据集上,同以往方法相比,有了更好的结果,同时在CrowdPose、OCHuman以及SyncOCC数据集上验证了泛化能力。该模型在一定程度上能够解决严重遮挡下的姿态估计问题,提高了预测姿态的合理性,取得了更加优异的效果。 展开更多
关键词 人体姿态估计 不可见关键点 严重遮挡 注意力机制 变分编码器
在线阅读 下载PDF
Application of the Improved PF-Flow-Style-VTON in Virtual Try-On
12
作者 TIAN Jiajia HUANG Rong +1 位作者 DONG Aihua WANG Zhijie 《Journal of Donghua University(English Edition)》 2026年第1期104-117,共14页
During the image generation phase,the parserfree Flow-Style-VTON model(PF-Flow-Style-VTON),which utilizes distilled appearance flows,faces two main challenges:blurring,deformation,occlusion,or loss of the arm or palm ... During the image generation phase,the parserfree Flow-Style-VTON model(PF-Flow-Style-VTON),which utilizes distilled appearance flows,faces two main challenges:blurring,deformation,occlusion,or loss of the arm or palm regions in the generated image when these regions of the person occlude the garment;blurring and deformation in the generated image when the person performs large pose movements and the target garment is complex with detailed patterns.To solve these two problems,an improved virtual try-on network model,denoted as IPF-Flow-Style-VTON,is proposed.Firstly,a target warped garment mask refinement module(M-RM)is introduced to refine the warped garment mask and remove erroneous information in the arm and palm regions,thereby improving the quality of subsequent image generation.Secondly,an improved global attention module(GAM)is integrated into the original image generation network,enhancing the ResUNet’s understanding of global context and optimizing the fusion of local features and global information,thereby further improving image generation quality.Finally,the UniPose model is used to provide the pose keypoint information of the target person image,guiding the task execution during the image generation phase.Experiments conducted on the VITON dataset show that the proposed method outperforms the original method,Flow-Style-VTON,by 5.4%,0.3%,6.7%,and 2.2%in Frchet inception distance(FID),structural similarity index measure(SSIM),learned perceptual image patch similarity(LPIPS),and peak signal-to-noise ratio(PSNR),respectively.Overall,the proposed method effectively improves upon the shortcomings of the original network and achieves better visual results. 展开更多
关键词 virtual try-on image generation network pose keypoint deep learning
在线阅读 下载PDF
Intelligent Human Interaction Recognition with Multi-Modal Feature Extraction and Bidirectional LSTM
13
作者 Muhammad Hamdan Azhar Yanfeng Wu +4 位作者 Nouf Abdullah Almujally Shuaa S.Alharbi Asaad Algarni Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2026年第4期1632-1649,共18页
Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationall... Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationally intensive,sensitive to video resolution changes and often fail in crowded scenes.We propose a novel hybrid system that is computationally efficient,robust to degraded video quality and able to filter out irrelevant individuals,making it suitable for real-life use.The system leverages multi-modal handcrafted features for interaction representation and a deep learning classifier for capturing complex dependencies.Using Mask R-CNN and YOLO11-Pose,we extract grayscale silhouettes and keypoint coordinates of interacting individuals,while filtering out irrelevant individuals using a proposed algorithm.From these,we extract silhouette-based features(local ternary pattern and histogram of optical flow)and keypoint-based features(distances,angles and velocities)that capture distinct spatial and temporal information.A Bidirectional Long Short-Term Memory network(BiLSTM)then classifies the interactions.Extensive experiments on the UT Interaction,SBU Kinect Interaction and the ISR-UOL 3D social activity datasets demonstrate that our system achieves competitive accuracy.They also validate the effectiveness of the chosen features and classifier,along with the proposed system’s computational efficiency and robustness to occlusion. 展开更多
关键词 Human interaction recognition keypoint coordinates grayscale silhouettes bidirectional long shortterm memory network
在线阅读 下载PDF
基于YOLO-FMC-pose的中华绒螯蟹头胸甲关键点检测方法
14
作者 张哲 于合龙 +3 位作者 杨信廷 罗娜 李珊珊 孙传恒 《农业工程学报》 北大核心 2026年第1期210-221,共12页
中华绒螯蟹(Eriocheir sinensis)的头胸甲形态在同一物种的不同个体之间表现出明显差异,这一特征可作为产地溯源和个体识别的重要依据。其中,头胸甲关键点的精准检测是实现个体识别与表型分析等任务的基础环节。然而,传统的人工检测方... 中华绒螯蟹(Eriocheir sinensis)的头胸甲形态在同一物种的不同个体之间表现出明显差异,这一特征可作为产地溯源和个体识别的重要依据。其中,头胸甲关键点的精准检测是实现个体识别与表型分析等任务的基础环节。然而,传统的人工检测方法依赖经验性判断,存在效率低、重复性差等问题,难以满足规模化水产处理的实际需求。为此,该研究提出了一种基于YOLO-FMC-pose的中华绒螯蟹头胸甲关键点自动检测方法,以实现高精度、自动化的特征提取。首先,构建了一个包含大量中华绒螯蟹头胸甲图像的自建数据集,并选取具有代表性的35个地标关键点进行精确标注,同时通过数据增强提升模型的训练效果。其次,该研究基于改进的YOLO11n-pose框架设计了中华绒螯蟹头胸甲关键点检测模型YOLO-FMC-pose。模型中引入了融合频率动态卷积(FDConv)的C3K2FD模块、混合聚合网络(MANet)模块以及CBAM注意力机制,从频域响应、特征融合与空间关注等层面对结构进行了优化。结果表明,所提出的YOLO-FMC-pose模型在关键点检测精度方面均优于现有主流方法,准确率、召回率、mAP_(0.5)和m AP_(0.5:0.95)分别为97.98%、97.00%、98.27%和73.28%,相较于原始YOLO11n-pose,准确率、召回率、mAP_(0.5)和mAP_(0.5:0.95)分别提高了3.33、2.33、2.94和13.08个百分点,标准化平均误差(normalized mean error,NME)降低至3.835%,单帧图片推理时间为7.5 ms,具备良好的实际应用潜力。该研究为中华绒螯蟹的个体智能识别、产地溯源与防伪管控提供了关键技术支撑,也为水产品精细化特征检测提供了路径。 展开更多
关键词 中华绒螯蟹 关键点检测 表型特征识别 深度学习 图像处理
在线阅读 下载PDF
基于扩散模型多模态提示的电力人员行为图像生成
15
作者 朱志航 闫云凤 齐冬莲 《浙江大学学报(工学版)》 北大核心 2026年第1期43-51,70,共10页
电力人员行为的特殊性与复杂性导致其图像数据稀缺,给数据驱动下的行为识别带来了挑战.在稳定扩散模型的基础上,充分融合人体骨架、掩膜以及文本描述信息,加入关键点损失函数,建立多模态条件控制的电力人员行为图像生成模型PoseNet,该... 电力人员行为的特殊性与复杂性导致其图像数据稀缺,给数据驱动下的行为识别带来了挑战.在稳定扩散模型的基础上,充分融合人体骨架、掩膜以及文本描述信息,加入关键点损失函数,建立多模态条件控制的电力人员行为图像生成模型PoseNet,该模型可以生成高质量的可控人体图像.设计基于关键点相似度的图像滤波器,以去除错误、低质量的生成图像;采用双阶段训练策略,在通用数据上对模型进行预训练,并在私有数据上微调,提升模型性能;针对电力人员行为特点,设计集通用、专用评价指标于一体的生成图像评价指标集,分析不同评价指标下的图像生成效果.实验结果表明,与主流人体生成模型ControlNet、HumanSD相比,该模型的生成结果更精准、真实、效果更优. 展开更多
关键词 条件图像生成模型 数据扩充 人体关键点 图像分割 扩散模型 深度学习
在线阅读 下载PDF
基于深度学习的服装关键点实时检测模型
16
作者 冯采伶 于施佳 韩曙光 《纺织学报》 北大核心 2026年第1期196-206,共11页
针对复杂场景下服装关键点检测模型的准确性与实时性难以兼得的问题,以提高检测准确性并保持实时性能为目的,提出了一种基于深度学习的服装关键点实时检测模型。该模型以实时多人姿态估计架构为基础,首先构建中值增强的通道与空间注意... 针对复杂场景下服装关键点检测模型的准确性与实时性难以兼得的问题,以提高检测准确性并保持实时性能为目的,提出了一种基于深度学习的服装关键点实时检测模型。该模型以实时多人姿态估计架构为基础,首先构建中值增强的通道与空间注意力模块,通过并行执行全局平均池化、最大池化与中值池化,融合生成注意力权重,增强服装关键部位的特征表示;其次设计跨尺度特征融合模块,将骨干网络中不同层级的特征图进行上采样、拼接与交叉卷积融合,构建兼具细节信息与语义特征的金字塔结构;进一步建立自注意力特征增强模块,通过计算特征点间相似性动态生成注意力图,自适应调整各区域特征权重;最终实施分类别微调策略,针对6类典型服装分别建立专用模型以优化整体性能。结果表明:该方法在DeepFashion2和DeepFashion数据集上分别达到了65.1%与68.0%的检测准确度,同时保持140.0帧/s和142.3帧/s的实时处理速度。该模型提升了复杂场景下服装关键点检测的综合性能,未来可应用于服装智能制造和虚拟试衣等领域。 展开更多
关键词 服装关键点 实时检测 深度学习 自注意力机制 跨尺度特征融合
在线阅读 下载PDF
基于改进YOLOv8n-pose的巨峰葡萄采摘定位方法
17
作者 陈馨 吴子炜 +1 位作者 周素茵 夏芳 《华南农业大学学报》 北大核心 2026年第1期118-127,共10页
【目的】对巨峰葡萄进行精准高效地采摘定位,以有效降低果实损伤。【方法】提出一种基于改进YOLOv8n-pose的葡萄采摘定位方法。首先,利用改进YOLOv8n-pose检测葡萄果梗和顶部易损果粒的关键点,基于关键点的坐标构建果实上界位姿的表征向... 【目的】对巨峰葡萄进行精准高效地采摘定位,以有效降低果实损伤。【方法】提出一种基于改进YOLOv8n-pose的葡萄采摘定位方法。首先,利用改进YOLOv8n-pose检测葡萄果梗和顶部易损果粒的关键点,基于关键点的坐标构建果实上界位姿的表征向量;然后,利用此向量计算出最优采摘角度;最后,通过将采摘点与采摘角协同,确定最佳采摘位置。【结果】试验结果表明,改进后YOLOv8n-pose的P、R、mAP@0.50、mAP@0.50~0.95较原模型分别提升了1.7、0.7、0.9、1.7个百分点,较YOLOv12s-pose分别提升了0.4、0.1、0.6、2.7个百分点,同时模型参数量比YOLOv8n-pose减少了5.8%。应用本文方法的葡萄采摘定位成功率为90.8%,相较于不使用采摘角的定位方法,提升了9.2个百分点。【结论】研究为巨峰葡萄采摘机器人提供了一种低损定位方法。 展开更多
关键词 巨峰葡萄 YOLOv8-pose 关键点检测 采摘定位 采摘角度
在线阅读 下载PDF
实例分割与空间解析融合的番茄实时三维位姿估计方法
18
作者 苟豪 赵国瑞 +3 位作者 董适 吕生华 林晨 文剑 《农业工程学报》 北大核心 2026年第2期185-194,共10页
针对复杂种植环境中现有果实位姿估计方法精度低、实时性差等问题,该研究提出一种融合轻量化实例分割与空间解析的番茄实时三维位姿估计方法。通过构建改进的YOLOv7-M1轻量化网络,实现果实掩膜高精度提取与关键点感兴趣区域快速定位;设... 针对复杂种植环境中现有果实位姿估计方法精度低、实时性差等问题,该研究提出一种融合轻量化实例分割与空间解析的番茄实时三维位姿估计方法。通过构建改进的YOLOv7-M1轻量化网络,实现果实掩膜高精度提取与关键点感兴趣区域快速定位;设计HRNet-ECA嵌入高通量注意力机制提升检测准确率;搭建多模态数据融合框架,结合深度图与感兴趣区域,经点云滤波处理和空间几何计算实时获取果实三维位姿参数。试验结果表明,改进后的YOLOv7-M1掩膜分割平均精度为95.56%,召回率93.52%,准确率96.17%;改进后的HRNet-ECA关键点相似度为96.61%,位姿估计准确率95.00%,三维姿态角平均误差9.40°,关键点平均定位误差4.13 mm,关键点在X、Y、Z方向上的平均误差分别为3.41、2.95和1.02 mm。单果处理平均耗时0.063 s。该方法构建了轻量化实例分割网络与改进关键点检测模型的级联结构,结合点云空间解析,在保证精度指标同时兼顾实时效率,实现了番茄果实的高精度实时位姿估计,可为复杂农业场景下果蔬精准自动化采收提供高效的解决方案。 展开更多
关键词 机器人 位姿估计 实例分割 关键点检测 点云滤波
在线阅读 下载PDF
基于局部消失点的车道线检测方法研究
19
作者 秦文清 赵尔敦 王永强 《计算机工程与应用》 北大核心 2026年第1期356-368,共13页
现存的车道线检测方法常为提升检测精度而使用计算复杂、内存占用多的算法,往往忽视检测速度以及部署难度。为此,利用车道线互相平行的先验信息设计了一种新的车道线表示方法,该方法使车道线共享同一组局部消失点,不仅大幅降低了参数量... 现存的车道线检测方法常为提升检测精度而使用计算复杂、内存占用多的算法,往往忽视检测速度以及部署难度。为此,利用车道线互相平行的先验信息设计了一种新的车道线表示方法,该方法使车道线共享同一组局部消失点,不仅大幅降低了参数量,在部分遮挡情况下也能准确恢复车道线形状。在此基础上提出一种3D车道线检测模型——LVPDepth,并为训练适配了消失点标签转换算法、改进了KL散度损失函数。该模型的特点如下:设计了深度检测模块,从而通过相机内参矩阵和车道线深度就能获得车道线三维坐标;为训练过程定义一种匹配标签和预测结果的准则,可以预测任意条车道线;针对车道线细长的形状,引入动态蛇形卷积提升检测精度;利用车道线天然的深度信息,加入预设相对深度向量,使训练更快收敛、结果更准确稳定。模型在校正后的ONCE-3DLanes数据集上进行训练与验证,在检测速度达到132 FPS的同时精度损失甚微。 展开更多
关键词 消失点 透视学 关键点检测 车道线检测
在线阅读 下载PDF
轻量且高精度的飞行器关键点检测改进网络GMD-YOLO
20
作者 刘鹏飞 孙世岩 +1 位作者 李池 张瑜 《海军工程大学学报》 北大核心 2026年第1期76-84,共9页
针对空中飞行器关键点检测中存在的实时性要求高、低分辨率、多尺度分布及部分遮挡等挑战,本文提出了一种基于YOLOv11n-pose架构的轻量化高精度检测算法GMD-YOLO。首先,设计了双门控融合网络,通过中值增强通道注意力与动态门控瓶颈卷积... 针对空中飞行器关键点检测中存在的实时性要求高、低分辨率、多尺度分布及部分遮挡等挑战,本文提出了一种基于YOLOv11n-pose架构的轻量化高精度检测算法GMD-YOLO。首先,设计了双门控融合网络,通过中值增强通道注意力与动态门控瓶颈卷积双分支协同机制,增强复杂光照下的特征鲁棒性;其次,构建轻量动态特征融合模块,采用双阶段注意力实现跨层特征自适应加权,缓解多尺度目标错位问题;再次,引入可变形卷积增强的C2PSA模块,通过动态采样网格提升形变关键点建模能力;最后,提出自适应图卷积姿态头,显式编码关键点间刚体约束以优化空间一致性。在自建的飞行器仿真数据集上的实验结果表明:GMD-YOLO仅以3.50 MB参数量实现91.9%均值平均精度P_(mA)@0.5与81.7%的P_(mA)@0.5∶0.95,较基准模型分别提升了6.0%与5.3%,在复杂场景下展现出显著精度优势与工程应用潜力。 展开更多
关键词 关键点检测 固定翼飞行器 YOLOv11 可变形卷积 图卷积网络
在线阅读 下载PDF
上一页 1 2 19 下一页 到第
使用帮助 返回顶部