期刊文献+
共找到138篇文章
< 1 2 7 >
每页显示 20 50 100
EHDC-YOLO: Enhancing Object Detection for UAV Imagery via Multi-Scale Edge and Detail Capture
1
作者 Zhiyong Deng Yanchen Ye Jiangling Guo 《Computers, Materials & Continua》 2026年第1期1665-1682,共18页
With the rapid expansion of drone applications,accurate detection of objects in aerial imagery has become crucial for intelligent transportation,urban management,and emergency rescue missions.However,existing methods ... With the rapid expansion of drone applications,accurate detection of objects in aerial imagery has become crucial for intelligent transportation,urban management,and emergency rescue missions.However,existing methods face numerous challenges in practical deployment,including scale variation handling,feature degradation,and complex backgrounds.To address these issues,we propose Edge-enhanced and Detail-Capturing You Only Look Once(EHDC-YOLO),a novel framework for object detection in Unmanned Aerial Vehicle(UAV)imagery.Based on the You Only Look Once version 11 nano(YOLOv11n)baseline,EHDC-YOLO systematically introduces several architectural enhancements:(1)a Multi-Scale Edge Enhancement(MSEE)module that leverages multi-scale pooling and edge information to enhance boundary feature extraction;(2)an Enhanced Feature Pyramid Network(EFPN)that integrates P2-level features with Cross Stage Partial(CSP)structures and OmniKernel convolutions for better fine-grained representation;and(3)Dynamic Head(DyHead)with multi-dimensional attention mechanisms for enhanced cross-scale modeling and perspective adaptability.Comprehensive experiments on the Vision meets Drones for Detection(VisDrone-DET)2019 dataset demonstrate that EHDC-YOLO achieves significant improvements,increasing mean Average Precision(mAP)@0.5 from 33.2%to 46.1%(an absolute improvement of 12.9 percentage points)and mAP@0.5:0.95 from 19.5%to 28.0%(an absolute improvement of 8.5 percentage points)compared with the YOLOv11n baseline,while maintaining a reasonable parameter count(2.81 M vs the baseline’s 2.58 M).Further ablation studies confirm the effectiveness of each proposed component,while visualization results highlight EHDC-YOLO’s superior performance in detecting objects and handling occlusions in complex drone scenarios. 展开更多
关键词 UAV imagery object detection multi-scale feature fusion edge enhancement detail preservation YOLO feature pyramid network attention mechanism
在线阅读 下载PDF
SIM-Net:A Multi-Scale Attention-Guided Deep Learning Framework for High-Precision PCB Defect Detection
2
作者 Ping Fang Mengjun Tong 《Computers, Materials & Continua》 2026年第4期1754-1770,共17页
Defect detection in printed circuit boards(PCB)remains challenging due to the difficulty of identifying small-scale defects,the inefficiency of conventional approaches,and the interference from complex backgrounds.To ... Defect detection in printed circuit boards(PCB)remains challenging due to the difficulty of identifying small-scale defects,the inefficiency of conventional approaches,and the interference from complex backgrounds.To address these issues,this paper proposes SIM-Net,an enhanced detection framework derived from YOLOv11.The model integrates SPDConv to preserve fine-grained features for small object detection,introduces a novel convolutional partial attention module(C2PAM)to suppress redundant background information and highlight salient regions,and employs a multi-scale fusion network(MFN)with a multi-grain contextual module(MGCT)to strengthen contextual representation and accelerate inference.Experimental evaluations demonstrate that SIM-Net achieves 92.4%mAP,92%accuracy,and 89.4%recall with an inference speed of 75.1 FPS,outperforming existing state-of-the-art methods.These results confirm the robustness and real-time applicability of SIM-Net for PCB defect inspection. 展开更多
关键词 Deep learning small object detection PCB defect detection attention mechanism multi-scale fusion network
在线阅读 下载PDF
Research on Camouflage Target Detection Method Based on Edge Guidance and Multi-Scale Feature Fusion
3
作者 Tianze Yu Jianxun Zhang Hongji Chen 《Computers, Materials & Continua》 2026年第4期1676-1697,共22页
Camouflaged Object Detection(COD)aims to identify objects that share highly similar patterns—such as texture,intensity,and color—with their surrounding environment.Due to their intrinsic resemblance to the backgroun... Camouflaged Object Detection(COD)aims to identify objects that share highly similar patterns—such as texture,intensity,and color—with their surrounding environment.Due to their intrinsic resemblance to the background,camouflaged objects often exhibit vague boundaries and varying scales,making it challenging to accurately locate targets and delineate their indistinct edges.To address this,we propose a novel camouflaged object detection network called Edge-Guided and Multi-scale Fusion Network(EGMFNet),which leverages edge-guided multi-scale integration for enhanced performance.The model incorporates two innovative components:a Multi-scale Fusion Module(MSFM)and an Edge-Guided Attention Module(EGA).These designs exploit multi-scale features to uncover subtle cues between candidate objects and the background while emphasizing camouflaged object boundaries.Moreover,recognizing the rich contextual information in fused features,we introduce a Dual-Branch Global Context Module(DGCM)to refine features using extensive global context,thereby generatingmore informative representations.Experimental results on four benchmark datasets demonstrate that EGMFNet outperforms state-of-the-art methods across five evaluation metrics.Specifically,on COD10K,our EGMFNet-P improves F_(β)by 4.8 points and reduces mean absolute error(MAE)by 0.006 compared with ZoomNeXt;on NC4K,it achieves a 3.6-point increase in F_(β).OnCAMO and CHAMELEON,it obtains 4.5-point increases in F_(β),respectively.These consistent gains substantiate the superiority and robustness of EGMFNet. 展开更多
关键词 Camouflaged object detection multi-scale feature fusion edge-guided image segmentation
在线阅读 下载PDF
MGD-YOLO:An Enhanced Road Defect Detection Algorithm Based on Multi-Scale Attention Feature Fusion
4
作者 Zhengji Li Fazhan Xiong +6 位作者 Boyun Huang Meihui Li Xi Xiao Yingrui Ji Jiacheng Xie Aokun Liang Hao Xu 《Computers, Materials & Continua》 2025年第9期5613-5635,共23页
Accurate and real-time road defect detection is essential for ensuring traffic safety and infrastructure maintenance.However,existing vision-based methods often struggle with small,sparse,and low-resolution defects un... Accurate and real-time road defect detection is essential for ensuring traffic safety and infrastructure maintenance.However,existing vision-based methods often struggle with small,sparse,and low-resolution defects under complex road conditions.To address these limitations,we propose Multi-Scale Guided Detection YOLO(MGD-YOLO),a novel lightweight and high-performance object detector built upon You Only Look Once Version 5(YOLOv5).The proposed model integrates three key components:(1)a Multi-Scale Dilated Attention(MSDA)module to enhance semantic feature extraction across varying receptive fields;(2)Depthwise Separable Convolution(DSC)to reduce computational cost and improve model generalization;and(3)a Visual Global Attention Upsampling(VGAU)module that leverages high-level contextual information to refine low-level features for precise localization.Extensive experiments on three public road defect benchmarks demonstrate that MGD-YOLO outperforms state-of-the-art models in both detection accuracy and efficiency.Notably,our model achieves 87.9%accuracy in crack detection,88.3%overall precision on TD-RD dataset,while maintaining fast inference speed and a compact architecture.These results highlight the potential of MGD-YOLO for deployment in real-time,resource-constrained scenarios,paving the way for practical and scalable intelligent road maintenance systems. 展开更多
关键词 YOLO road damage detection object detection computer vision deep learning
在线阅读 下载PDF
Infrared road object detection algorithm based on spatial depth channel attention network and improved YOLOv8
5
作者 LI Song SHI Tao +1 位作者 JING Fangke CUI Jie 《Optoelectronics Letters》 2025年第8期491-498,共8页
Aiming at the problems of low detection accuracy and large model size of existing object detection algorithms applied to complex road scenes,an improved you only look once version 8(YOLOv8)object detection algorithm f... Aiming at the problems of low detection accuracy and large model size of existing object detection algorithms applied to complex road scenes,an improved you only look once version 8(YOLOv8)object detection algorithm for infrared images,F-YOLOv8,is proposed.First,a spatial-to-depth network replaces the traditional backbone network's strided convolution or pooling layer.At the same time,it combines with the channel attention mechanism so that the neural network focuses on the channels with large weight values to better extract low-resolution image feature information;then an improved feature pyramid network of lightweight bidirectional feature pyramid network(L-BiFPN)is proposed,which can efficiently fuse features of different scales.In addition,a loss function of insertion of union based on the minimum point distance(MPDIoU)is introduced for bounding box regression,which obtains faster convergence speed and more accurate regression results.Experimental results on the FLIR dataset show that the improved algorithm can accurately detect infrared road targets in real time with 3%and 2.2%enhancement in mean average precision at 50%IoU(mAP50)and mean average precision at 50%—95%IoU(mAP50-95),respectively,and 38.1%,37.3%and 16.9%reduction in the number of model parameters,the model weight,and floating-point operations per second(FLOPs),respectively.To further demonstrate the detection capability of the improved algorithm,it is tested on the public dataset PASCAL VOC,and the results show that F-YOLO has excellent generalized detection performance. 展开更多
关键词 feature pyramid network infrared road object detection infrared imagesf yolov backbone networks channel attention mechanism spatial depth channel attention network object detection improved YOLOv
原文传递
YOLO-MFD:Remote Sensing Image Object Detection with Multi-Scale Fusion Dynamic Head
6
作者 Zhongyuan Zhang Wenqiu Zhu 《Computers, Materials & Continua》 SCIE EI 2024年第5期2547-2563,共17页
Remote sensing imagery,due to its high altitude,presents inherent challenges characterized by multiple scales,limited target areas,and intricate backgrounds.These inherent traits often lead to increased miss and false... Remote sensing imagery,due to its high altitude,presents inherent challenges characterized by multiple scales,limited target areas,and intricate backgrounds.These inherent traits often lead to increased miss and false detection rates when applying object recognition algorithms tailored for remote sensing imagery.Additionally,these complexities contribute to inaccuracies in target localization and hinder precise target categorization.This paper addresses these challenges by proposing a solution:The YOLO-MFD model(YOLO-MFD:Remote Sensing Image Object Detection withMulti-scale Fusion Dynamic Head).Before presenting our method,we delve into the prevalent issues faced in remote sensing imagery analysis.Specifically,we emphasize the struggles of existing object recognition algorithms in comprehensively capturing critical image features amidst varying scales and complex backgrounds.To resolve these issues,we introduce a novel approach.First,we propose the implementation of a lightweight multi-scale module called CEF.This module significantly improves the model’s ability to comprehensively capture important image features by merging multi-scale feature information.It effectively addresses the issues of missed detection and mistaken alarms that are common in remote sensing imagery.Second,an additional layer of small target detection heads is added,and a residual link is established with the higher-level feature extraction module in the backbone section.This allows the model to incorporate shallower information,significantly improving the accuracy of target localization in remotely sensed images.Finally,a dynamic head attentionmechanism is introduced.This allows themodel to exhibit greater flexibility and accuracy in recognizing shapes and targets of different sizes.Consequently,the precision of object detection is significantly improved.The trial results show that the YOLO-MFD model shows improvements of 6.3%,3.5%,and 2.5%over the original YOLOv8 model in Precision,map@0.5 and map@0.5:0.95,separately.These results illustrate the clear advantages of the method. 展开更多
关键词 object detection YOLOv8 multi-scale attention mechanism dynamic detection head
在线阅读 下载PDF
MSC-YOLO:Improved YOLOv7 Based on Multi-Scale Spatial Context for Small Object Detection in UAV-View
7
作者 Xiangyan Tang Chengchun Ruan +2 位作者 Xiulai Li Binbin Li Cebin Fu 《Computers, Materials & Continua》 SCIE EI 2024年第4期983-1003,共21页
Accurately identifying small objects in high-resolution aerial images presents a complex and crucial task in thefield of small object detection on unmanned aerial vehicles(UAVs).This task is challenging due to variati... Accurately identifying small objects in high-resolution aerial images presents a complex and crucial task in thefield of small object detection on unmanned aerial vehicles(UAVs).This task is challenging due to variations inUAV flight altitude,differences in object scales,as well as factors like flight speed and motion blur.To enhancethe detection efficacy of small targets in drone aerial imagery,we propose an enhanced You Only Look Onceversion 7(YOLOv7)algorithm based on multi-scale spatial context.We build the MSC-YOLO model,whichincorporates an additional prediction head,denoted as P2,to improve adaptability for small objects.We replaceconventional downsampling with a Spatial-to-Depth Convolutional Combination(CSPDC)module to mitigatethe loss of intricate feature details related to small objects.Furthermore,we propose a Spatial Context Pyramidwith Multi-Scale Attention(SCPMA)module,which captures spatial and channel-dependent features of smalltargets acrossmultiple scales.This module enhances the perception of spatial contextual features and the utilizationof multiscale feature information.On the Visdrone2023 and UAVDT datasets,MSC-YOLO achieves remarkableresults,outperforming the baseline method YOLOv7 by 3.0%in terms ofmean average precision(mAP).The MSCYOLOalgorithm proposed in this paper has demonstrated satisfactory performance in detecting small targets inUAV aerial photography,providing strong support for practical applications. 展开更多
关键词 Small object detection YOLOv7 multi-scale attention spatial context
在线阅读 下载PDF
Hybrid receptive field network for small object detection on drone view 被引量:1
8
作者 Zhaodong CHEN Hongbing JI +2 位作者 Yongquan ZHANG Wenke LIU Zhigang ZHU 《Chinese Journal of Aeronautics》 2025年第2期322-338,共17页
Drone-based small object detection is of great significance in practical applications such as military actions, disaster rescue, transportation, etc. However, the severe scale differences in objects captured by drones... Drone-based small object detection is of great significance in practical applications such as military actions, disaster rescue, transportation, etc. However, the severe scale differences in objects captured by drones and lack of detail information for small-scale objects make drone-based small object detection a formidable challenge. To address these issues, we first develop a mathematical model to explore how changing receptive fields impacts the polynomial fitting results. Subsequently, based on the obtained conclusions, we propose a simple but effective Hybrid Receptive Field Network (HRFNet), whose modules include Hybrid Feature Augmentation (HFA), Hybrid Feature Pyramid (HFP) and Dual Scale Head (DSH). Specifically, HFA employs parallel dilated convolution kernels of different sizes to extend shallow features with different receptive fields, committed to improving the multi-scale adaptability of the network;HFP enhances the perception of small objects by capturing contextual information across layers, while DSH reconstructs the original prediction head utilizing a set of high-resolution features and ultrahigh-resolution features. In addition, in order to train HRFNet, the corresponding dual-scale loss function is designed. Finally, comprehensive evaluation results on public benchmarks such as VisDrone-DET and TinyPerson demonstrate the robustness of the proposed method. Most impressively, the proposed HRFNet achieves a mAP of 51.0 on VisDrone-DET with 29.3 M parameters, which outperforms the extant state-of-the-art detectors. HRFNet also performs excellently in complex scenarios captured by drones, achieving the best performance on the CS-Drone dataset we built. 展开更多
关键词 Drone remote sensing object detection on drone view Small object detector Hybrid receptive field Feature pyramid network Feature augmentation multi-scale object detection
原文传递
DDFNet:real-time salient object detection with dual-branch decoding fusion for steel plate surface defects
9
作者 Tao Wang Wang-zhe Du +5 位作者 Xu-wei Li Hua-xin Liu Yuan-ming Liu Xiao-miao Niu Ya-xing Liu Tao Wang 《Journal of Iron and Steel Research International》 2025年第8期2421-2433,共13页
A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decod... A novel dual-branch decoding fusion convolutional neural network model(DDFNet)specifically designed for real-time salient object detection(SOD)on steel surfaces is proposed.DDFNet is based on a standard encoder–decoder architecture.DDFNet integrates three key innovations:first,we introduce a novel,lightweight multi-scale progressive aggregation residual network that effectively suppresses background interference and refines defect details,enabling efficient salient feature extraction.Then,we propose an innovative dual-branch decoding fusion structure,comprising the refined defect representation branch and the enhanced defect representation branch,which enhance accuracy in defect region identification and feature representation.Additionally,to further improve the detection of small and complex defects,we incorporate a multi-scale attention fusion module.Experimental results on the public ESDIs-SOD dataset show that DDFNet,with only 3.69 million parameters,achieves detection performance comparable to current state-of-the-art models,demonstrating its potential for real-time industrial applications.Furthermore,our DDFNet-L variant consistently outperforms leading methods in detection performance.The code is available at https://github.com/13140W/DDFNet. 展开更多
关键词 Steel plate surface defect Real-time detection Salient object detection Dual-branch decoder multi-scale attention fusion multi-scale residual fusion
原文传递
SPD-YOLO:A Novel Lightweight YOLO Modelfor Road Information Detection
10
作者 Guoliang Li Xianxin Ke +1 位作者 Tao Xue Xiangyu Liao 《Journal of Beijing Institute of Technology》 2025年第5期482-495,共14页
Rapid and high-precision speed bump detection is critical for autonomous driving and road safety,yet it faces challenges from non-standard appearances and complex environments.To address this issue,this study proposes... Rapid and high-precision speed bump detection is critical for autonomous driving and road safety,yet it faces challenges from non-standard appearances and complex environments.To address this issue,this study proposes a you only look once(YOLO)algorithm for speed bump detection(SPD-YOLO),a lightweight model based on YOLO11s that integrates three core innova-tive modules to balance detection precision and computational efficiency:it replaces YOLO11s’original backbone with StarNet,which uses‘star operations’to map features into high-dimensional nonlinear spaces for enhanced feature representation while maintaining computational efficiency;its neck incorporates context feature calibration(CFC)and spatial feature calibration(SFC)to improve detection performance without significant computational overhead;and its detection head adopts a lightweight shared convolutional detection(LSCD)structure combined with GroupNorm,minimizing computational complexity while preserving multi-scale feature fusion efficacy.Experi-ments on a custom speed bump dataset show SPD-YOLO achieves a mean average precision(mAP)of 79.9%,surpassing YOLO11s by 1.3%and YOLO12s by 1.2%while reducing parameters by 26.3%and floating-point operations per second(FLOPs)by 29.5%,enabling real-time deploy-ment on resource-constrained platforms. 展开更多
关键词 LIGHTWEIGHT object detection road speed bump detection YOLO11 algorithm
在线阅读 下载PDF
MADF-YOLOv8:A Lightweight Model for Road Distress Detection Based on Adaptive Multiscale Feature Fusion
11
作者 Tao OuYang Haohui Yu +3 位作者 Guanlin Pan Yan Cui Qingling Chang Xiulong Fu 《Journal of Electronic Research and Application》 2025年第6期96-104,共9页
Efficient road distress detection is crucial for transportation safety.To address the challenge of balancing detection accuracy,efficiency,and multi-scale feature fusion in existing methods,this paper proposes a light... Efficient road distress detection is crucial for transportation safety.To address the challenge of balancing detection accuracy,efficiency,and multi-scale feature fusion in existing methods,this paper proposes a lightweight model named MADF-YOLOv8.The model enhances multi-scale feature extraction capability by introducing the Multi-Scale Ghost Residual Convolution(MSGRConv)and the Multiscale Adaptive Feature Processing Module(MAFP).Furthermore,it constructs a Multi-scale Dynamic sampling Bidirectional Feature Pyramid Network(MD-BiFPN)and incorporates the C2f-Faster module to optimize feature fusion efficiency.Experiments on the RDD2022 dataset demonstrate that the proposed model achieves a mean Average Precision at 0.5 Intersection over Union(mAP@0.5)of 88.6%with only 2.312 million parameters.Its overall performance surpasses various mainstream detectors,achieving an exceptional balance between accuracy and efficiency. 展开更多
关键词 road distress detection multi-scale feature fusion YOLOv8
在线阅读 下载PDF
Coupling the Power of YOLOv9 with Transformer for Small Object Detection in Remote-Sensing Images
12
作者 Mohammad Barr 《Computer Modeling in Engineering & Sciences》 2025年第4期593-616,共24页
Recent years have seen a surge in interest in object detection on remote sensing images for applications such as surveillance andmanagement.However,challenges like small object detection,scale variation,and the presen... Recent years have seen a surge in interest in object detection on remote sensing images for applications such as surveillance andmanagement.However,challenges like small object detection,scale variation,and the presence of closely packed objects in these images hinder accurate detection.Additionally,the motion blur effect further complicates the identification of such objects.To address these issues,we propose enhanced YOLOv9 with a transformer head(YOLOv9-TH).The model introduces an additional prediction head for detecting objects of varying sizes and swaps the original prediction heads for transformer heads to leverage self-attention mechanisms.We further improve YOLOv9-TH using several strategies,including data augmentation,multi-scale testing,multi-model integration,and the introduction of an additional classifier.The cross-stage partial(CSP)method and the ghost convolution hierarchical graph(GCHG)are combined to improve detection accuracy by better utilizing feature maps,widening the receptive field,and precisely extracting multi-scale objects.Additionally,we incorporate the E-SimAM attention mechanism to address low-resolution feature loss.Extensive experiments on the VisDrone2021 and DIOR datasets demonstrate the effectiveness of YOLOv9-TH,showing good improvement in mAP compared to the best existing methods.The YOLOv9-TH-e achieved 54.2% of mAP50 on the VisDrone2021 dataset and 92.3% of mAP on the DIOR dataset.The results confirmthemodel’s robustness and suitability for real-world applications,particularly for small object detection in remote sensing images. 展开更多
关键词 Remote sensing images YOLOv9-TH multi-scale object detection transformer heads VisDrone2021 dataset
在线阅读 下载PDF
Speed and Accuracy Tradeoff for LiDAR Data Based Road Boundary Detection 被引量:6
13
作者 Guojun Wang Jian Wu +1 位作者 Rui He Bin Tian 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第6期1210-1220,共11页
Road boundary detection is essential for autonomous vehicle localization and decision-making,especially under GPS signal loss and lane discontinuities.For road boundary detection in structural environments,obstacle oc... Road boundary detection is essential for autonomous vehicle localization and decision-making,especially under GPS signal loss and lane discontinuities.For road boundary detection in structural environments,obstacle occlusions and large road curvature are two significant challenges.However,an effective and fast solution for these problems has remained elusive.To solve these problems,a speed and accuracy tradeoff method for LiDAR-based road boundary detection in structured environments is proposed.The proposed method consists of three main stages:1)a multi-feature based method is applied to extract feature points;2)a road-segmentation-line-based method is proposed for classifying left and right feature points;3)an iterative Gaussian Process Regression(GPR)is employed for filtering out false points and extracting boundary points.To demonstrate the effectiveness of the proposed method,KITTI datasets is used for comprehensive experiments,and the performance of our approach is tested under different road conditions.Comprehensive experiments show the roadsegmentation-line-based method can classify left,and right feature points on structured curved roads,and the proposed iterative Gaussian Process Regression can extract road boundary points on varied road shapes and traffic conditions.Meanwhile,the proposed road boundary detection method can achieve real-time performance with an average of 70.5 ms per frame. 展开更多
关键词 3D-LiDAR autonomous vehicle object detection point cloud road boundary
在线阅读 下载PDF
Zero-DCE++Inspired Object Detection in Less Illuminated Environment Using Improved YOLOv5 被引量:3
14
作者 Ananthakrishnan Balasundaram Anshuman Mohanty +3 位作者 Ayesha Shaik Krishnadoss Pradeep Kedalu Poornachary Vijayakumar Muthu Subash Kavitha 《Computers, Materials & Continua》 SCIE EI 2023年第12期2751-2769,共19页
Automated object detection has received the most attention over the years.Use cases ranging from autonomous driving applications to military surveillance systems,require robust detection of objects in different illumi... Automated object detection has received the most attention over the years.Use cases ranging from autonomous driving applications to military surveillance systems,require robust detection of objects in different illumination conditions.State-of-the-art object detectors tend to fare well in object detection during daytime conditions.However,their performance is severely hampered in night light conditions due to poor illumination.To address this challenge,the manuscript proposes an improved YOLOv5-based object detection framework for effective detection in unevenly illuminated nighttime conditions.Firstly,the preprocessing strategies involve using the Zero-DCE++approach to enhance lowlight images.It is followed by optimizing the existing YOLOv5 architecture by integrating the Convolutional Block Attention Module(CBAM)in the backbone network to boost model learning capability and Depthwise Convolutional module(DWConv)in the neck network for efficient compression of network parameters.The Night Object Detection(NOD)and Exclusively Dark(ExDARK)dataset has been used for this work.The proposed framework detects classes like humans,bicycles,and cars.Experiments demonstrate that the proposed architecture achieved a higher Mean Average Precision(mAP)along with a reduction in model size and total parameters,respectively.The proposed model is lighter by 11.24%in terms of model size and 12.38%in terms of parameters when compared to baseline YOLOv5. 展开更多
关键词 object detection deep learning nighttime road scenes YOLOv5 DWConv Zero-DCE++ CBAM
在线阅读 下载PDF
Neighborhood fusion-based hierarchical parallel feature pyramid network for object detection 被引量:3
15
作者 Mo Lingfei Hu Shuming 《Journal of Southeast University(English Edition)》 EI CAS 2020年第3期252-263,共12页
In order to improve the detection accuracy of small objects,a neighborhood fusion-based hierarchical parallel feature pyramid network(NFPN)is proposed.Unlike the layer-by-layer structure adopted in the feature pyramid... In order to improve the detection accuracy of small objects,a neighborhood fusion-based hierarchical parallel feature pyramid network(NFPN)is proposed.Unlike the layer-by-layer structure adopted in the feature pyramid network(FPN)and deconvolutional single shot detector(DSSD),where the bottom layer of the feature pyramid network relies on the top layer,NFPN builds the feature pyramid network with no connections between the upper and lower layers.That is,it only fuses shallow features on similar scales.NFPN is highly portable and can be embedded in many models to further boost performance.Extensive experiments on PASCAL VOC 2007,2012,and COCO datasets demonstrate that the NFPN-based SSD without intricate tricks can exceed the DSSD model in terms of detection accuracy and inference speed,especially for small objects,e.g.,4%to 5%higher mAP(mean average precision)than SSD,and 2%to 3%higher mAP than DSSD.On VOC 2007 test set,the NFPN-based SSD with 300×300 input reaches 79.4%mAP at 34.6 frame/s,and the mAP can raise to 82.9%after using the multi-scale testing strategy. 展开更多
关键词 computer vision deep convolutional neural network object detection hierarchical parallel feature pyramid network multi-scale feature fusion
在线阅读 下载PDF
Bidirectional parallel multi-branch convolution feature pyramid network for target detection in aerial images of swarm UAVs 被引量:4
16
作者 Lei Fu Wen-bin Gu +3 位作者 Wei Li Liang Chen Yong-bao Ai Hua-lei Wang 《Defence Technology(防务技术)》 SCIE EI CAS CSCD 2021年第4期1531-1541,共11页
In this paper,based on a bidirectional parallel multi-branch feature pyramid network(BPMFPN),a novel one-stage object detector called BPMFPN Det is proposed for real-time detection of ground multi-scale targets by swa... In this paper,based on a bidirectional parallel multi-branch feature pyramid network(BPMFPN),a novel one-stage object detector called BPMFPN Det is proposed for real-time detection of ground multi-scale targets by swarm unmanned aerial vehicles(UAVs).First,the bidirectional parallel multi-branch convolution modules are used to construct the feature pyramid to enhance the feature expression abilities of different scale feature layers.Next,the feature pyramid is integrated into the single-stage object detection framework to ensure real-time performance.In order to validate the effectiveness of the proposed algorithm,experiments are conducted on four datasets.For the PASCAL VOC dataset,the proposed algorithm achieves the mean average precision(mAP)of 85.4 on the VOC 2007 test set.With regard to the detection in optical remote sensing(DIOR)dataset,the proposed algorithm achieves 73.9 mAP.For vehicle detection in aerial imagery(VEDAI)dataset,the detection accuracy of small land vehicle(slv)targets reaches 97.4 mAP.For unmanned aerial vehicle detection and tracking(UAVDT)dataset,the proposed BPMFPN Det achieves the mAP of 48.75.Compared with the previous state-of-the-art methods,the results obtained by the proposed algorithm are more competitive.The experimental results demonstrate that the proposed algorithm can effectively solve the problem of real-time detection of ground multi-scale targets in aerial images of swarm UAVs. 展开更多
关键词 Aerial images object detection Feature pyramid networks multi-scale feature fusion Swarm UAVs
在线阅读 下载PDF
Industrial Fusion Cascade Detection of Solder Joint
17
作者 Chunyuan Li Peng Zhang +2 位作者 Shuangming Wang Lie Liu Mingquan Shi 《Computers, Materials & Continua》 SCIE EI 2024年第10期1197-1214,共18页
With the remarkable advancements in machine vision research and its ever-expanding applications,scholars have increasingly focused on harnessing various vision methodologies within the industrial realm.Specifically,de... With the remarkable advancements in machine vision research and its ever-expanding applications,scholars have increasingly focused on harnessing various vision methodologies within the industrial realm.Specifically,detecting vehicle floor welding points poses unique challenges,including high operational costs and limited portability in practical settings.To address these challenges,this paper innovatively integrates template matching and the Faster RCNN algorithm,presenting an industrial fusion cascaded solder joint detection algorithm that seamlessly blends template matching with deep learning techniques.This algorithm meticulously weights and fuses the optimized features of both methodologies,enhancing the overall detection capabilities.Furthermore,it introduces an optimized multi-scale and multi-template matching approach,leveraging a diverse array of templates and image pyramid algorithms to bolster the accuracy and resilience of object detection.By integrating deep learning algorithms with this multi-scale and multi-template matching strategy,the cascaded target matching algorithm effectively accurately identifies solder joint types and positions.A comprehensive welding point dataset,labeled by experts specifically for vehicle detection,was constructed based on images from authentic industrial environments to validate the algorithm’s performance.Experiments demonstrate the algorithm’s compelling performance in industrial scenarios,outperforming the single-template matching algorithm by 21.3%,the multi-scale and multitemplate matching algorithm by 3.4%,the Faster RCNN algorithm by 19.7%,and the YOLOv9 algorithm by 17.3%in terms of solder joint detection accuracy.This optimized algorithm exhibits remarkable robustness and portability,ideally suited for detecting solder joints across diverse vehicle workpieces.Notably,this study’s dataset and feature fusion approach can be a valuable resource for other algorithms seeking to enhance their solder joint detection capabilities.This work thus not only presents a novel and effective solution for industrial solder joint detection but lays the groundwork for future advancements in this critical area. 展开更多
关键词 Cascade object detection deep learning feature fusion multi-scale and multi-template matching solder joint dataset
在线阅读 下载PDF
MFF‒YOLO:多尺度特征融合的轻量级道路缺陷检测算法
18
作者 侯涛 张田明 牛宏侠 《工程科学与技术》 北大核心 2026年第1期303-312,共10页
道路缺陷检测是道路养护的前提与基础,对道路安全保障非常重要。针对现有道路缺陷检测算法未能有效平衡检测精度与运算复杂度,从而难以应用于移动终端设备的问题,本文在YOLOv7-tiny的基础上提出了一种多尺度特征融合(multiscale feature... 道路缺陷检测是道路养护的前提与基础,对道路安全保障非常重要。针对现有道路缺陷检测算法未能有效平衡检测精度与运算复杂度,从而难以应用于移动终端设备的问题,本文在YOLOv7-tiny的基础上提出了一种多尺度特征融合(multiscale feature fusion,MFF)的轻量级道路缺陷检测算法MFF‒YOLO。首先,设计多尺度特征融合模块(MFFBlock)与下采样模块(DSB),并在此基础上构建高效的骨干多尺度特征提取网络(MFENet),以增强多尺度特征的提取能力。然后,颈部特征融合网络采用Slim-Neck设计范式,即利用GSConv与VoV‒GSCSPC模块实现颈部特征的聚合,降低计算复杂度,轻量化网络。同时,颈部特征融合网络采用FSF‒PAFPN结构以实现对多尺度特征的高效融合,提升算法对道路缺陷的定位与分类能力。最后,利用K-Means算法对道路缺陷数据集进行聚类,获取更符合道路缺陷目标形状特点的先验框,降低算法的训练难度,提升检测精度。在数据集RDD2022上的实验结果表明,相较于YOLOv7-tiny,MFF‒YOLO轻量化显著,参数量与运算量分别降低了约25.1%与约25.8%。此外,MFF‒YOLO的平均精确率均值达到了60.1%,较YOLOv7-tiny提升了2.3个百分点,为多种对比算法中的最高值。同时,MFF‒YOLO在检测效果方面表现出色,能够对道路缺陷进行精准定位并分类,且其检测帧率达到81帧/s,表现出较高的实时性。MFF‒YOLO实现了检测精度与计算复杂度的有效平衡,为在移动终端设备上实现道路缺陷检测提供参考。 展开更多
关键词 道路缺陷 目标检测 YOLOv7-tiny 轻量化
在线阅读 下载PDF
盲道识别与障碍物检测的多任务模型
19
作者 徐浩闻 李维乾 《计算机与现代化》 2026年第1期30-39,共10页
本文提出Amaterasu-YOLO多任务模型,旨在提升盲道区域分割与障碍物检测的精度与效率。该模型结合自适应串联模块(ECD)和多感受野空间注意力模块(MRSA),能够在复杂城市环境中实现高精度的盲道分割和障碍物检测。通过多任务学习的方式,Ama... 本文提出Amaterasu-YOLO多任务模型,旨在提升盲道区域分割与障碍物检测的精度与效率。该模型结合自适应串联模块(ECD)和多感受野空间注意力模块(MRSA),能够在复杂城市环境中实现高精度的盲道分割和障碍物检测。通过多任务学习的方式,Amaterasu-YOLO不仅优化了盲道分割和障碍物检测的联合任务,还显著降低了计算负担,提高了模型在资源受限的边缘设备上的应用效率。实验结果表明,Amaterasu-YOLO在盲道区域分割与障碍物检测任务上均取得了良好的性能,分别达到了90%的分割精度和85%的障碍物检测准确率。与传统单任务方法相比,模型展现出更强的鲁棒性和实用性,在智能城市建设和视障人士出行安全等领域具有广泛的应用潜力。 展开更多
关键词 YOLOv8 盲道分割 障碍物检测 多任务模型 注意力机制 目标检测
在线阅读 下载PDF
基于YOLOv5n-BGF的雾天道路目标检测算法
20
作者 郝宇翔 甄国涌 储成群 《现代电子技术》 北大核心 2026年第5期37-43,共7页
针对在雾天环境下的道路目标检测存在特征提取难度大、精度与效率难以平衡的问题,文中在YOLOv5n的基础上创新性地提出一种高效且轻量的YOLOv5n-BGF变体,该变体结合了双向特征金字塔网络(BiFPN)模型,利用双向连接的结构特点,更加有效地... 针对在雾天环境下的道路目标检测存在特征提取难度大、精度与效率难以平衡的问题,文中在YOLOv5n的基础上创新性地提出一种高效且轻量的YOLOv5n-BGF变体,该变体结合了双向特征金字塔网络(BiFPN)模型,利用双向连接的结构特点,更加有效地结合不同尺度的特征;其次,引入GELAN模块代替颈部网络中的C3结构,在减少计算量的同时增强了有效特征的提取;最后,考虑不同样本的边界框回归问题,采用Focaler-IoU来进一步提高检测性能。在本地平台针对非公开雾天道路目标检测数据集D-8800进行验证,实验结果表明,相较于基础模型YOLOv5n,改进后的YOLOv5n-BGF的mAP@0.5提升了5.3%,参数量减少了25%,GFLOPs仅为3.5,YOLOv5n-BGF凭借其卓越的性能,在雾天道路目标检测数据集D-8800上的表现优于其他目标检测模型,为雾天道路目标检测提供了高效的解决方案。 展开更多
关键词 雾天道路 目标检测 YOLOv5n变体 BiFPN GELAN 轻量化设计
在线阅读 下载PDF
上一页 1 2 7 下一页 到第
使用帮助 返回顶部