期刊文献+
共找到129篇文章
< 1 2 7 >
每页显示 20 50 100
EHDC-YOLO: Enhancing Object Detection for UAV Imagery via Multi-Scale Edge and Detail Capture
1
作者 Zhiyong Deng Yanchen Ye Jiangling Guo 《Computers, Materials & Continua》 2026年第1期1665-1682,共18页
With the rapid expansion of drone applications,accurate detection of objects in aerial imagery has become crucial for intelligent transportation,urban management,and emergency rescue missions.However,existing methods ... With the rapid expansion of drone applications,accurate detection of objects in aerial imagery has become crucial for intelligent transportation,urban management,and emergency rescue missions.However,existing methods face numerous challenges in practical deployment,including scale variation handling,feature degradation,and complex backgrounds.To address these issues,we propose Edge-enhanced and Detail-Capturing You Only Look Once(EHDC-YOLO),a novel framework for object detection in Unmanned Aerial Vehicle(UAV)imagery.Based on the You Only Look Once version 11 nano(YOLOv11n)baseline,EHDC-YOLO systematically introduces several architectural enhancements:(1)a Multi-Scale Edge Enhancement(MSEE)module that leverages multi-scale pooling and edge information to enhance boundary feature extraction;(2)an Enhanced Feature Pyramid Network(EFPN)that integrates P2-level features with Cross Stage Partial(CSP)structures and OmniKernel convolutions for better fine-grained representation;and(3)Dynamic Head(DyHead)with multi-dimensional attention mechanisms for enhanced cross-scale modeling and perspective adaptability.Comprehensive experiments on the Vision meets Drones for Detection(VisDrone-DET)2019 dataset demonstrate that EHDC-YOLO achieves significant improvements,increasing mean Average Precision(mAP)@0.5 from 33.2%to 46.1%(an absolute improvement of 12.9 percentage points)and mAP@0.5:0.95 from 19.5%to 28.0%(an absolute improvement of 8.5 percentage points)compared with the YOLOv11n baseline,while maintaining a reasonable parameter count(2.81 M vs the baseline’s 2.58 M).Further ablation studies confirm the effectiveness of each proposed component,while visualization results highlight EHDC-YOLO’s superior performance in detecting objects and handling occlusions in complex drone scenarios. 展开更多
关键词 UAV imagery object detection multi-scale feature fusion edge enhancement detail preservation YOLO feature pyramid network attention mechanism
在线阅读 下载PDF
MGD-YOLO:An Enhanced Road Defect Detection Algorithm Based on Multi-Scale Attention Feature Fusion
2
作者 Zhengji Li Fazhan Xiong +6 位作者 Boyun Huang Meihui Li Xi Xiao Yingrui Ji Jiacheng Xie Aokun Liang Hao Xu 《Computers, Materials & Continua》 2025年第9期5613-5635,共23页
Accurate and real-time road defect detection is essential for ensuring traffic safety and infrastructure maintenance.However,existing vision-based methods often struggle with small,sparse,and low-resolution defects un... Accurate and real-time road defect detection is essential for ensuring traffic safety and infrastructure maintenance.However,existing vision-based methods often struggle with small,sparse,and low-resolution defects under complex road conditions.To address these limitations,we propose Multi-Scale Guided Detection YOLO(MGD-YOLO),a novel lightweight and high-performance object detector built upon You Only Look Once Version 5(YOLOv5).The proposed model integrates three key components:(1)a Multi-Scale Dilated Attention(MSDA)module to enhance semantic feature extraction across varying receptive fields;(2)Depthwise Separable Convolution(DSC)to reduce computational cost and improve model generalization;and(3)a Visual Global Attention Upsampling(VGAU)module that leverages high-level contextual information to refine low-level features for precise localization.Extensive experiments on three public road defect benchmarks demonstrate that MGD-YOLO outperforms state-of-the-art models in both detection accuracy and efficiency.Notably,our model achieves 87.9%accuracy in crack detection,88.3%overall precision on TD-RD dataset,while maintaining fast inference speed and a compact architecture.These results highlight the potential of MGD-YOLO for deployment in real-time,resource-constrained scenarios,paving the way for practical and scalable intelligent road maintenance systems. 展开更多
关键词 YOLO road damage detection object detection computer vision deep learning
在线阅读 下载PDF
Infrared road object detection algorithm based on spatial depth channel attention network and improved YOLOv8
3
作者 LI Song SHI Tao +1 位作者 JING Fangke CUI Jie 《Optoelectronics Letters》 2025年第8期491-498,共8页
Aiming at the problems of low detection accuracy and large model size of existing object detection algorithms applied to complex road scenes,an improved you only look once version 8(YOLOv8)object detection algorithm f... Aiming at the problems of low detection accuracy and large model size of existing object detection algorithms applied to complex road scenes,an improved you only look once version 8(YOLOv8)object detection algorithm for infrared images,F-YOLOv8,is proposed.First,a spatial-to-depth network replaces the traditional backbone network's strided convolution or pooling layer.At the same time,it combines with the channel attention mechanism so that the neural network focuses on the channels with large weight values to better extract low-resolution image feature information;then an improved feature pyramid network of lightweight bidirectional feature pyramid network(L-BiFPN)is proposed,which can efficiently fuse features of different scales.In addition,a loss function of insertion of union based on the minimum point distance(MPDIoU)is introduced for bounding box regression,which obtains faster convergence speed and more accurate regression results.Experimental results on the FLIR dataset show that the improved algorithm can accurately detect infrared road targets in real time with 3%and 2.2%enhancement in mean average precision at 50%IoU(mAP50)and mean average precision at 50%—95%IoU(mAP50-95),respectively,and 38.1%,37.3%and 16.9%reduction in the number of model parameters,the model weight,and floating-point operations per second(FLOPs),respectively.To further demonstrate the detection capability of the improved algorithm,it is tested on the public dataset PASCAL VOC,and the results show that F-YOLO has excellent generalized detection performance. 展开更多
关键词 feature pyramid network infrared road object detection infrared imagesf yolov backbone networks channel attention mechanism spatial depth channel attention network object detection improved YOLOv
原文传递
YOLO-MFD:Remote Sensing Image Object Detection with Multi-Scale Fusion Dynamic Head
4
作者 Zhongyuan Zhang Wenqiu Zhu 《Computers, Materials & Continua》 SCIE EI 2024年第5期2547-2563,共17页
Remote sensing imagery,due to its high altitude,presents inherent challenges characterized by multiple scales,limited target areas,and intricate backgrounds.These inherent traits often lead to increased miss and false... Remote sensing imagery,due to its high altitude,presents inherent challenges characterized by multiple scales,limited target areas,and intricate backgrounds.These inherent traits often lead to increased miss and false detection rates when applying object recognition algorithms tailored for remote sensing imagery.Additionally,these complexities contribute to inaccuracies in target localization and hinder precise target categorization.This paper addresses these challenges by proposing a solution:The YOLO-MFD model(YOLO-MFD:Remote Sensing Image Object Detection withMulti-scale Fusion Dynamic Head).Before presenting our method,we delve into the prevalent issues faced in remote sensing imagery analysis.Specifically,we emphasize the struggles of existing object recognition algorithms in comprehensively capturing critical image features amidst varying scales and complex backgrounds.To resolve these issues,we introduce a novel approach.First,we propose the implementation of a lightweight multi-scale module called CEF.This module significantly improves the model’s ability to comprehensively capture important image features by merging multi-scale feature information.It effectively addresses the issues of missed detection and mistaken alarms that are common in remote sensing imagery.Second,an additional layer of small target detection heads is added,and a residual link is established with the higher-level feature extraction module in the backbone section.This allows the model to incorporate shallower information,significantly improving the accuracy of target localization in remotely sensed images.Finally,a dynamic head attentionmechanism is introduced.This allows themodel to exhibit greater flexibility and accuracy in recognizing shapes and targets of different sizes.Consequently,the precision of object detection is significantly improved.The trial results show that the YOLO-MFD model shows improvements of 6.3%,3.5%,and 2.5%over the original YOLOv8 model in Precision,map@0.5 and map@0.5:0.95,separately.These results illustrate the clear advantages of the method. 展开更多
关键词 object detection YOLOv8 multi-scale attention mechanism dynamic detection head
在线阅读 下载PDF
MSC-YOLO:Improved YOLOv7 Based on Multi-Scale Spatial Context for Small Object Detection in UAV-View
5
作者 Xiangyan Tang Chengchun Ruan +2 位作者 Xiulai Li Binbin Li Cebin Fu 《Computers, Materials & Continua》 SCIE EI 2024年第4期983-1003,共21页
Accurately identifying small objects in high-resolution aerial images presents a complex and crucial task in thefield of small object detection on unmanned aerial vehicles(UAVs).This task is challenging due to variati... Accurately identifying small objects in high-resolution aerial images presents a complex and crucial task in thefield of small object detection on unmanned aerial vehicles(UAVs).This task is challenging due to variations inUAV flight altitude,differences in object scales,as well as factors like flight speed and motion blur.To enhancethe detection efficacy of small targets in drone aerial imagery,we propose an enhanced You Only Look Onceversion 7(YOLOv7)algorithm based on multi-scale spatial context.We build the MSC-YOLO model,whichincorporates an additional prediction head,denoted as P2,to improve adaptability for small objects.We replaceconventional downsampling with a Spatial-to-Depth Convolutional Combination(CSPDC)module to mitigatethe loss of intricate feature details related to small objects.Furthermore,we propose a Spatial Context Pyramidwith Multi-Scale Attention(SCPMA)module,which captures spatial and channel-dependent features of smalltargets acrossmultiple scales.This module enhances the perception of spatial contextual features and the utilizationof multiscale feature information.On the Visdrone2023 and UAVDT datasets,MSC-YOLO achieves remarkableresults,outperforming the baseline method YOLOv7 by 3.0%in terms ofmean average precision(mAP).The MSCYOLOalgorithm proposed in this paper has demonstrated satisfactory performance in detecting small targets inUAV aerial photography,providing strong support for practical applications. 展开更多
关键词 Small object detection YOLOv7 multi-scale attention spatial context
在线阅读 下载PDF
Hybrid receptive field network for small object detection on drone view 被引量:1
6
作者 Zhaodong CHEN Hongbing JI +2 位作者 Yongquan ZHANG Wenke LIU Zhigang ZHU 《Chinese Journal of Aeronautics》 2025年第2期322-338,共17页
Drone-based small object detection is of great significance in practical applications such as military actions, disaster rescue, transportation, etc. However, the severe scale differences in objects captured by drones... Drone-based small object detection is of great significance in practical applications such as military actions, disaster rescue, transportation, etc. However, the severe scale differences in objects captured by drones and lack of detail information for small-scale objects make drone-based small object detection a formidable challenge. To address these issues, we first develop a mathematical model to explore how changing receptive fields impacts the polynomial fitting results. Subsequently, based on the obtained conclusions, we propose a simple but effective Hybrid Receptive Field Network (HRFNet), whose modules include Hybrid Feature Augmentation (HFA), Hybrid Feature Pyramid (HFP) and Dual Scale Head (DSH). Specifically, HFA employs parallel dilated convolution kernels of different sizes to extend shallow features with different receptive fields, committed to improving the multi-scale adaptability of the network;HFP enhances the perception of small objects by capturing contextual information across layers, while DSH reconstructs the original prediction head utilizing a set of high-resolution features and ultrahigh-resolution features. In addition, in order to train HRFNet, the corresponding dual-scale loss function is designed. Finally, comprehensive evaluation results on public benchmarks such as VisDrone-DET and TinyPerson demonstrate the robustness of the proposed method. Most impressively, the proposed HRFNet achieves a mAP of 51.0 on VisDrone-DET with 29.3 M parameters, which outperforms the extant state-of-the-art detectors. HRFNet also performs excellently in complex scenarios captured by drones, achieving the best performance on the CS-Drone dataset we built. 展开更多
关键词 Drone remote sensing object detection on drone view Small object detector Hybrid receptive field Feature pyramid network Feature augmentation multi-scale object detection
原文传递
DDFNet:real-time salient object detection with dual-branch decoding fusion for steel plate surface defects
7
作者 Tao Wang Wang-zhe Du +5 位作者 Xu-wei Li Hua-xin Liu Yuan-ming Liu Xiao-miao Niu Ya-xing Liu Tao Wang 《Journal of Iron and Steel Research International》 2025年第8期2421-2433,共13页
A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decod... A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decoder architecture.DDFNet integrates three key innovations:first,we introduce a novel,lightweight multi-scale progressive aggregation residual network that effectively suppresses background interference and refines defect details,enabling efficient salient feature extraction.Then,we propose an innovative dual-branch decoding fusion structure,comprising the refined defect representation branch and the enhanced defect representation branch,which enhance accuracy in defect region identification and feature representation.Additionally,to further improve the detection of small and complex defects,we incorporate a multi-scale attention fusion module.Experimental results on the public ESDIs-SOD dataset show that DDFNet,with only 3.69 million parameters,achieves detection performance comparable to current state-of-the-art models,demonstrating its potential for real-time industrial applications.Furthermore,our DDFNet-L variant consistently outperforms leading methods in detection performance.The code is available at https://github.com/13140W/DDFNet. 展开更多
关键词 Steel plate surface defect Real-time detection Salient object detection Dual-branch decoder multi-scale attention fusion multi-scale residual fusion
原文传递
SPD-YOLO:A Novel Lightweight YOLO Modelfor Road Information Detection
8
作者 Guoliang Li Xianxin Ke +1 位作者 Tao Xue Xiangyu Liao 《Journal of Beijing Institute of Technology》 2025年第5期482-495,共14页
Rapid and high-precision speed bump detection is critical for autonomous driving and road safety,yet it faces challenges from non-standard appearances and complex environments.To address this issue,this study proposes... Rapid and high-precision speed bump detection is critical for autonomous driving and road safety,yet it faces challenges from non-standard appearances and complex environments.To address this issue,this study proposes a you only look once(YOLO)algorithm for speed bump detection(SPD-YOLO),a lightweight model based on YOLO11s that integrates three core innova-tive modules to balance detection precision and computational efficiency:it replaces YOLO11s’original backbone with StarNet,which uses‘star operations’to map features into high-dimensional nonlinear spaces for enhanced feature representation while maintaining computational efficiency;its neck incorporates context feature calibration(CFC)and spatial feature calibration(SFC)to improve detection performance without significant computational overhead;and its detection head adopts a lightweight shared convolutional detection(LSCD)structure combined with GroupNorm,minimizing computational complexity while preserving multi-scale feature fusion efficacy.Experi-ments on a custom speed bump dataset show SPD-YOLO achieves a mean average precision(mAP)of 79.9%,surpassing YOLO11s by 1.3%and YOLO12s by 1.2%while reducing parameters by 26.3%and floating-point operations per second(FLOPs)by 29.5%,enabling real-time deploy-ment on resource-constrained platforms. 展开更多
关键词 LIGHTWEIGHT object detection road speed bump detection YOLO11 algorithm
在线阅读 下载PDF
MADF-YOLOv8:A Lightweight Model for Road Distress Detection Based on Adaptive Multiscale Feature Fusion
9
作者 Tao OuYang Haohui Yu +3 位作者 Guanlin Pan Yan Cui Qingling Chang Xiulong Fu 《Journal of Electronic Research and Application》 2025年第6期96-104,共9页
Efficient road distress detection is crucial for transportation safety.To address the challenge of balancing detection accuracy,efficiency,and multi-scale feature fusion in existing methods,this paper proposes a light... Efficient road distress detection is crucial for transportation safety.To address the challenge of balancing detection accuracy,efficiency,and multi-scale feature fusion in existing methods,this paper proposes a lightweight model named MADF-YOLOv8.The model enhances multi-scale feature extraction capability by introducing the Multi-Scale Ghost Residual Convolution(MSGRConv)and the Multiscale Adaptive Feature Processing Module(MAFP).Furthermore,it constructs a Multi-scale Dynamic sampling Bidirectional Feature Pyramid Network(MD-BiFPN)and incorporates the C2f-Faster module to optimize feature fusion efficiency.Experiments on the RDD2022 dataset demonstrate that the proposed model achieves a mean Average Precision at 0.5 Intersection over Union(mAP@0.5)of 88.6%with only 2.312 million parameters.Its overall performance surpasses various mainstream detectors,achieving an exceptional balance between accuracy and efficiency. 展开更多
关键词 road distress detection multi-scale feature fusion YOLOv8
在线阅读 下载PDF
Coupling the Power of YOLOv9 with Transformer for Small Object Detection in Remote-Sensing Images
10
作者 Mohammad Barr 《Computer Modeling in Engineering & Sciences》 2025年第4期593-616,共24页
Recent years have seen a surge in interest in object detection on remote sensing images for applications such as surveillance andmanagement.However,challenges like small object detection,scale variation,and the presen... Recent years have seen a surge in interest in object detection on remote sensing images for applications such as surveillance andmanagement.However,challenges like small object detection,scale variation,and the presence of closely packed objects in these images hinder accurate detection.Additionally,the motion blur effect further complicates the identification of such objects.To address these issues,we propose enhanced YOLOv9 with a transformer head(YOLOv9-TH).The model introduces an additional prediction head for detecting objects of varying sizes and swaps the original prediction heads for transformer heads to leverage self-attention mechanisms.We further improve YOLOv9-TH using several strategies,including data augmentation,multi-scale testing,multi-model integration,and the introduction of an additional classifier.The cross-stage partial(CSP)method and the ghost convolution hierarchical graph(GCHG)are combined to improve detection accuracy by better utilizing feature maps,widening the receptive field,and precisely extracting multi-scale objects.Additionally,we incorporate the E-SimAM attention mechanism to address low-resolution feature loss.Extensive experiments on the VisDrone2021 and DIOR datasets demonstrate the effectiveness of YOLOv9-TH,showing good improvement in mAP compared to the best existing methods.The YOLOv9-TH-e achieved 54.2% of mAP50 on the VisDrone2021 dataset and 92.3% of mAP on the DIOR dataset.The results confirmthemodel’s robustness and suitability for real-world applications,particularly for small object detection in remote sensing images. 展开更多
关键词 Remote sensing images YOLOv9-TH multi-scale object detection transformer heads VisDrone2021 dataset
在线阅读 下载PDF
Speed and Accuracy Tradeoff for LiDAR Data Based Road Boundary Detection 被引量:6
11
作者 Guojun Wang Jian Wu +1 位作者 Rui He Bin Tian 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第6期1210-1220,共11页
Road boundary detection is essential for autonomous vehicle localization and decision-making,especially under GPS signal loss and lane discontinuities.For road boundary detection in structural environments,obstacle oc... Road boundary detection is essential for autonomous vehicle localization and decision-making,especially under GPS signal loss and lane discontinuities.For road boundary detection in structural environments,obstacle occlusions and large road curvature are two significant challenges.However,an effective and fast solution for these problems has remained elusive.To solve these problems,a speed and accuracy tradeoff method for LiDAR-based road boundary detection in structured environments is proposed.The proposed method consists of three main stages:1)a multi-feature based method is applied to extract feature points;2)a road-segmentation-line-based method is proposed for classifying left and right feature points;3)an iterative Gaussian Process Regression(GPR)is employed for filtering out false points and extracting boundary points.To demonstrate the effectiveness of the proposed method,KITTI datasets is used for comprehensive experiments,and the performance of our approach is tested under different road conditions.Comprehensive experiments show the roadsegmentation-line-based method can classify left,and right feature points on structured curved roads,and the proposed iterative Gaussian Process Regression can extract road boundary points on varied road shapes and traffic conditions.Meanwhile,the proposed road boundary detection method can achieve real-time performance with an average of 70.5 ms per frame. 展开更多
关键词 3D-LiDAR autonomous vehicle object detection point cloud road boundary
在线阅读 下载PDF
Zero-DCE++Inspired Object Detection in Less Illuminated Environment Using Improved YOLOv5 被引量:3
12
作者 Ananthakrishnan Balasundaram Anshuman Mohanty +3 位作者 Ayesha Shaik Krishnadoss Pradeep Kedalu Poornachary Vijayakumar Muthu Subash Kavitha 《Computers, Materials & Continua》 SCIE EI 2023年第12期2751-2769,共19页
Automated object detection has received the most attention over the years.Use cases ranging from autonomous driving applications to military surveillance systems,require robust detection of objects in different illumi... Automated object detection has received the most attention over the years.Use cases ranging from autonomous driving applications to military surveillance systems,require robust detection of objects in different illumination conditions.State-of-the-art object detectors tend to fare well in object detection during daytime conditions.However,their performance is severely hampered in night light conditions due to poor illumination.To address this challenge,the manuscript proposes an improved YOLOv5-based object detection framework for effective detection in unevenly illuminated nighttime conditions.Firstly,the preprocessing strategies involve using the Zero-DCE++approach to enhance lowlight images.It is followed by optimizing the existing YOLOv5 architecture by integrating the Convolutional Block Attention Module(CBAM)in the backbone network to boost model learning capability and Depthwise Convolutional module(DWConv)in the neck network for efficient compression of network parameters.The Night Object Detection(NOD)and Exclusively Dark(ExDARK)dataset has been used for this work.The proposed framework detects classes like humans,bicycles,and cars.Experiments demonstrate that the proposed architecture achieved a higher Mean Average Precision(mAP)along with a reduction in model size and total parameters,respectively.The proposed model is lighter by 11.24%in terms of model size and 12.38%in terms of parameters when compared to baseline YOLOv5. 展开更多
关键词 object detection deep learning nighttime road scenes YOLOv5 DWConv Zero-DCE++ CBAM
在线阅读 下载PDF
Neighborhood fusion-based hierarchical parallel feature pyramid network for object detection 被引量:3
13
作者 Mo Lingfei Hu Shuming 《Journal of Southeast University(English Edition)》 EI CAS 2020年第3期252-263,共12页
In order to improve the detection accuracy of small objects,a neighborhood fusion-based hierarchical parallel feature pyramid network(NFPN)is proposed.Unlike the layer-by-layer structure adopted in the feature pyramid... In order to improve the detection accuracy of small objects,a neighborhood fusion-based hierarchical parallel feature pyramid network(NFPN)is proposed.Unlike the layer-by-layer structure adopted in the feature pyramid network(FPN)and deconvolutional single shot detector(DSSD),where the bottom layer of the feature pyramid network relies on the top layer,NFPN builds the feature pyramid network with no connections between the upper and lower layers.That is,it only fuses shallow features on similar scales.NFPN is highly portable and can be embedded in many models to further boost performance.Extensive experiments on PASCAL VOC 2007,2012,and COCO datasets demonstrate that the NFPN-based SSD without intricate tricks can exceed the DSSD model in terms of detection accuracy and inference speed,especially for small objects,e.g.,4%to 5%higher mAP(mean average precision)than SSD,and 2%to 3%higher mAP than DSSD.On VOC 2007 test set,the NFPN-based SSD with 300×300 input reaches 79.4%mAP at 34.6 frame/s,and the mAP can raise to 82.9%after using the multi-scale testing strategy. 展开更多
关键词 computer vision deep convolutional neural network object detection hierarchical parallel feature pyramid network multi-scale feature fusion
在线阅读 下载PDF
Bidirectional parallel multi-branch convolution feature pyramid network for target detection in aerial images of swarm UAVs 被引量:4
14
作者 Lei Fu Wen-bin Gu +3 位作者 Wei Li Liang Chen Yong-bao Ai Hua-lei Wang 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2021年第4期1531-1541,共11页
In this paper,based on a bidirectional parallel multi-branch feature pyramid network(BPMFPN),a novel one-stage object detector called BPMFPN Det is proposed for real-time detection of ground multi-scale targets by swa... In this paper,based on a bidirectional parallel multi-branch feature pyramid network(BPMFPN),a novel one-stage object detector called BPMFPN Det is proposed for real-time detection of ground multi-scale targets by swarm unmanned aerial vehicles(UAVs).First,the bidirectional parallel multi-branch convolution modules are used to construct the feature pyramid to enhance the feature expression abilities of different scale feature layers.Next,the feature pyramid is integrated into the single-stage object detection framework to ensure real-time performance.In order to validate the effectiveness of the proposed algorithm,experiments are conducted on four datasets.For the PASCAL VOC dataset,the proposed algorithm achieves the mean average precision(mAP)of 85.4 on the VOC 2007 test set.With regard to the detection in optical remote sensing(DIOR)dataset,the proposed algorithm achieves 73.9 mAP.For vehicle detection in aerial imagery(VEDAI)dataset,the detection accuracy of small land vehicle(slv)targets reaches 97.4 mAP.For unmanned aerial vehicle detection and tracking(UAVDT)dataset,the proposed BPMFPN Det achieves the mAP of 48.75.Compared with the previous state-of-the-art methods,the results obtained by the proposed algorithm are more competitive.The experimental results demonstrate that the proposed algorithm can effectively solve the problem of real-time detection of ground multi-scale targets in aerial images of swarm UAVs. 展开更多
关键词 Aerial images object detection Feature pyramid networks multi-scale feature fusion Swarm UAVs
在线阅读 下载PDF
Industrial Fusion Cascade Detection of Solder Joint
15
作者 Chunyuan Li Peng Zhang +2 位作者 Shuangming Wang Lie Liu Mingquan Shi 《Computers, Materials & Continua》 SCIE EI 2024年第10期1197-1214,共18页
With the remarkable advancements in machine vision research and its ever-expanding applications,scholars have increasingly focused on harnessing various vision methodologies within the industrial realm.Specifically,de... With the remarkable advancements in machine vision research and its ever-expanding applications,scholars have increasingly focused on harnessing various vision methodologies within the industrial realm.Specifically,detecting vehicle floor welding points poses unique challenges,including high operational costs and limited portability in practical settings.To address these challenges,this paper innovatively integrates template matching and the Faster RCNN algorithm,presenting an industrial fusion cascaded solder joint detection algorithm that seamlessly blends template matching with deep learning techniques.This algorithm meticulously weights and fuses the optimized features of both methodologies,enhancing the overall detection capabilities.Furthermore,it introduces an optimized multi-scale and multi-template matching approach,leveraging a diverse array of templates and image pyramid algorithms to bolster the accuracy and resilience of object detection.By integrating deep learning algorithms with this multi-scale and multi-template matching strategy,the cascaded target matching algorithm effectively accurately identifies solder joint types and positions.A comprehensive welding point dataset,labeled by experts specifically for vehicle detection,was constructed based on images from authentic industrial environments to validate the algorithm’s performance.Experiments demonstrate the algorithm’s compelling performance in industrial scenarios,outperforming the single-template matching algorithm by 21.3%,the multi-scale and multitemplate matching algorithm by 3.4%,the Faster RCNN algorithm by 19.7%,and the YOLOv9 algorithm by 17.3%in terms of solder joint detection accuracy.This optimized algorithm exhibits remarkable robustness and portability,ideally suited for detecting solder joints across diverse vehicle workpieces.Notably,this study’s dataset and feature fusion approach can be a valuable resource for other algorithms seeking to enhance their solder joint detection capabilities.This work thus not only presents a novel and effective solution for industrial solder joint detection but lays the groundwork for future advancements in this critical area. 展开更多
关键词 Cascade object detection deep learning feature fusion multi-scale and multi-template matching solder joint dataset
在线阅读 下载PDF
MFF‒YOLO:多尺度特征融合的轻量级道路缺陷检测算法
16
作者 侯涛 张田明 牛宏侠 《工程科学与技术》 北大核心 2026年第1期303-312,共10页
道路缺陷检测是道路养护的前提与基础,对道路安全保障非常重要。针对现有道路缺陷检测算法未能有效平衡检测精度与运算复杂度,从而难以应用于移动终端设备的问题,本文在YOLOv7-tiny的基础上提出了一种多尺度特征融合(multiscale feature... 道路缺陷检测是道路养护的前提与基础,对道路安全保障非常重要。针对现有道路缺陷检测算法未能有效平衡检测精度与运算复杂度,从而难以应用于移动终端设备的问题,本文在YOLOv7-tiny的基础上提出了一种多尺度特征融合(multiscale feature fusion,MFF)的轻量级道路缺陷检测算法MFF‒YOLO。首先,设计多尺度特征融合模块(MFFBlock)与下采样模块(DSB),并在此基础上构建高效的骨干多尺度特征提取网络(MFENet),以增强多尺度特征的提取能力。然后,颈部特征融合网络采用Slim-Neck设计范式,即利用GSConv与VoV‒GSCSPC模块实现颈部特征的聚合,降低计算复杂度,轻量化网络。同时,颈部特征融合网络采用FSF‒PAFPN结构以实现对多尺度特征的高效融合,提升算法对道路缺陷的定位与分类能力。最后,利用K-Means算法对道路缺陷数据集进行聚类,获取更符合道路缺陷目标形状特点的先验框,降低算法的训练难度,提升检测精度。在数据集RDD2022上的实验结果表明,相较于YOLOv7-tiny,MFF‒YOLO轻量化显著,参数量与运算量分别降低了约25.1%与约25.8%。此外,MFF‒YOLO的平均精确率均值达到了60.1%,较YOLOv7-tiny提升了2.3个百分点,为多种对比算法中的最高值。同时,MFF‒YOLO在检测效果方面表现出色,能够对道路缺陷进行精准定位并分类,且其检测帧率达到81帧/s,表现出较高的实时性。MFF‒YOLO实现了检测精度与计算复杂度的有效平衡,为在移动终端设备上实现道路缺陷检测提供参考。 展开更多
关键词 道路缺陷 目标检测 YOLOv7-tiny 轻量化
在线阅读 下载PDF
Efficient and accurate road crack detection technology based on YOLOv8-ES
17
作者 Kaili Zeng Rui Fan Xiaoyu Tang 《Autonomous Intelligent Systems》 2025年第1期339-349,共11页
Road damage detection is an important aspect of road maintenance.Traditional manual inspections are laborious and imprecise.With the rise of deep learning technology,pavement detection methods employing deep neural ne... Road damage detection is an important aspect of road maintenance.Traditional manual inspections are laborious and imprecise.With the rise of deep learning technology,pavement detection methods employing deep neural networks give an efficient and accurate solution.However,due to background diversity,limited resolution,and fracture similarity,it is tough to detect road cracks with high accuracy.In this study,we offer a unique,efficient and accurate road crack damage detection,namely YOLOv8-ES.We present a novel dynamic convolutional layer(EDCM)that successfully increases the feature extraction capabilities for small fractures.At the same time,we also present a new attention mechanism(SGAM).It can effectively retain crucial information and increase the network feature extraction capacity.The Wise-IoU technique contains a dynamic,non-monotonic focusing mechanism designed to return to the goal-bounding box more precisely,especially for low-quality samples.We validate our method on both RDD2022 and VOC2007 datasets.The experimental results suggest that YOLOv8-ES performs well.This unique approach provides great support for the development of intelligent road maintenance systems and is projected to achieve further advances in future applications. 展开更多
关键词 road crack detection object detection Attention mechanism Dynamic convolutional layer
原文传递
Adversarial Attack on Object Detection via Object Feature-Wise Attention and Perturbation Extraction
18
作者 Wei Xue Xiaoyan Xia +2 位作者 Pengcheng Wan Ping Zhong Xiao Zheng 《Tsinghua Science and Technology》 2025年第3期1174-1189,共16页
Deep neural networks are commonly used in computer vision tasks,but they are vulnerable to adversarial samples,resulting in poor recognition accuracy.Although traditional algorithms that craft adversarial samples have... Deep neural networks are commonly used in computer vision tasks,but they are vulnerable to adversarial samples,resulting in poor recognition accuracy.Although traditional algorithms that craft adversarial samples have been effective in attacking classification models,the attacking performance degrades when facing object detection models with more complex structures.To address this issue better,in this paper we first analyze the mechanism of multi-scale feature extraction of object detection models,and then by constructing the object feature-wise attention module and the perturbation extraction module,a novel adversarial sample generation algorithm for attacking detection models is proposed.Specifically,in the first module,based on the multi-scale feature map,we reduce the range of perturbation and improve the stealthiness of adversarial samples by computing the noise distribution in the object region.Then in the second module,we feed the noise distribution into the generative adversarial networks to generate adversarial perturbation with strong attack transferability.By doing so,the proposed approach possesses the ability to better confuse the judgment of detection models.Experiments carried out on the DroneVehicle dataset show that our method is computationally efficient and works well in attacking detection models measured by qualitative analysis and quantitative analysis. 展开更多
关键词 adversarial attack transfer attack object detection generative adversarial networks multi-scale feature map
原文传递
Real-time detection network for tiny traffic sign using multi-scale attention module 被引量:16
19
作者 YANG TingTing TONG Chao 《Science China(Technological Sciences)》 SCIE EI CAS CSCD 2022年第2期396-406,共11页
As one of the key technologies of intelligent vehicles, traffic sign detection is still a challenging task because of the tiny size of its target object. To address the challenge, we present a novel detection network ... As one of the key technologies of intelligent vehicles, traffic sign detection is still a challenging task because of the tiny size of its target object. To address the challenge, we present a novel detection network improved from yolo-v3 for the tiny traffic sign with high precision in real-time. First, a visual multi-scale attention module(MSAM), a light-weight yet effective module, is devised to fuse the multi-scale feature maps with channel weights and spatial masks. It increases the representation power of the network by emphasizing useful features and suppressing unnecessary ones. Second, we exploit effectively fine-grained features about tiny objects from the shallower layers through modifying backbone Darknet-53 and adding one prediction head to yolo-v3. Finally, a receptive field block is added into the neck of the network to broaden the receptive field. Experiments prove the effectiveness of our network in both quantitative and qualitative aspects. The m AP@0.5 of our network reaches 0.965 and its detection speed is55.56 FPS for 512 × 512 images on the challenging Tsinghua-Tencent 100 k(TT100 k) dataset. 展开更多
关键词 tiny object detection traffic sign detection multi-scale attention module REAL-TIME
原文传递
VCANet:Vanishing-Point-Guided Context-Aware Network for Small Road Object Detection 被引量:4
20
作者 Guang Chen Kai Chen +2 位作者 Lijun Zhang Liming Zhang Alois Knoll 《Automotive Innovation》 EI CSCD 2021年第4期400-412,共13页
Advanced deep learning technology has made great progress in generic object detection of autonomous driving,yet it is still challenging to detect small road hazards in a long distance owing to lack of large-scale smal... Advanced deep learning technology has made great progress in generic object detection of autonomous driving,yet it is still challenging to detect small road hazards in a long distance owing to lack of large-scale small-object datasets and dedicated methods.This work addresses the challenge from two aspects.Firstly,a self-collected long-distance road object dataset(TJ-LDRO)is introduced,which consists of 109,337 images and is the largest dataset so far for the small road object detection research.Secondly,a vanishing-point-guided context-aware network(VCANet)is proposed,which utilizes the vanishing point prediction block and the context-aware center detection block to obtain semantic information.The multi-scale feature fusion pipeline and the upsampling block in VCANet are introduced to enhance the region of interest(ROI)feature.The experimental results with TJ-LDRO dataset show that the proposed method achieves better performance than the representative generic object detection methods.This work fills a critical capability gap in small road hazards detection for high-speed autonomous vehicles. 展开更多
关键词 Autonomous driving road hazard object detection Deep learning Vanishing point
原文传递
上一页 1 2 7 下一页 到第
使用帮助 返回顶部