In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,par...In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,particularly in snowy environments,remains a challenge.Snow-covered roads introduce unpredictable surface conditions,occlusions,and reduced visibility,that require robust and adaptive path detection algorithms.This paper presents an enhanced road detection framework for snowy environments,leveraging Simple Framework forContrastive Learning of Visual Representations(SimCLR)for Self-Supervised pretraining,hyperparameter optimization,and uncertainty-aware object detection to improve the performance of YouOnly Look Once version 8(YOLOv8).Themodel is trained and evaluated on a custom-built dataset collected from snowy roads in Tromsø,Norway,which covers a range of snow textures,illumination conditions,and road geometries.The proposed framework achieves scores in terms of mAP@50 equal to 99%and mAP@50–95 equal to 97%,demonstrating the effectiveness of YOLOv8 for real-time road detection in extreme winter conditions.The findings contribute to the safe and reliable deployment of autonomous vehicles in Arctic environments,enabling robust decision-making in hazardous weather conditions.This research lays the groundwork for more resilient perceptionmodels in self-driving systems,paving the way for the future development of intelligent and adaptive transportation networks.展开更多
Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SG...Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SGW-YOLOv8 based on the improvement of YOLOv8.First,the Adaptive Fine-Grained Channel Attention(FCA)module is fused with the backbone layer of the YOLOv8 network to improve the feature extraction ability of the model.This paper uses the YOLOv8 network backbone layer to improve the feature extraction capability of the model.Second,the Efficient Multi-Scale Attention(C2f_EMA)module is replaced with the C2f module in the Neck layer of the network to improve the detection performance of the model for small underwater targets.Finally,the loss function is optimized to Weighted Intersection over Union(WIoU)to replace the original loss function,so that the model is better adapted to the target detection task in the complex ocean background.The improved algorithm has been experimented with on the Underwater Robot Picking Contest(URPC)dataset,and the results show that the improved algorithm achieves a detection accuracy of 84.5,which is 2.3%higher than that before the improvement,and at the same time,it can accurately detect the small-target marine organisms and adapts to the task of detecting marine organisms in various complex environments.展开更多
In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm...In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm based on YOLOv8 was proposed in this study.To begin with,the CoordAtt attention mechanism was employed to enhance the feature extraction capability of the backbone network,thereby reducing interference from backgrounds.Additionally,the BiFPN feature fusion network with an added small object detection layer was used to enhance the model's ability to perceive for small objects.Furthermore,a multi-level fusion module was designed and proposed to effectively integrate shallow and deep information.The use of an enhanced MPDIoU loss function further improved detection performance.The experimental results based on the publicly available VisDrone2019 dataset showed that the improved model outperformed the YOLOv8 baseline model,mAP@0.5 improved by 20%,and the improved method improved the detection accuracy of the model for small targets.展开更多
An improved CSYOLOv8 model based on YOLOv8 model is developed specifically for identifying defects in printed circuit board(PCB).Firstly,a composite backbone network is designed to carry out additional feature extract...An improved CSYOLOv8 model based on YOLOv8 model is developed specifically for identifying defects in printed circuit board(PCB).Firstly,a composite backbone network is designed to carry out additional feature extraction,which enriches the expression ability of features and enhances the detection accuracy of the model.Secondly,a YOLO-FPN(Feature pyramid network)structure is designed to supplant the original neck network,which enhances the feature fusion ability of the model and improves the detection accuracy of small target objects.Furthermore,to enhance the model’s capability to extract tubular features,dynamic snake convolution is implemented.Finally,MPDIoU loss function is employed to enhance both the convergence rate and the precision of the model.Experiments show that the mAP of the improved model on the PCB defect dataset reaches 96.6%,which is 4.5%higher than that of the YOLOv8 model,and the number of parameters is only 3256862,and the average detection speed is 51.8 frames per second,which meets the requirements of detection accuracy and efficiency.展开更多
UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,comp...UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,complex backgrounds,and variable lighting persist due to the unique perspective of UAV imagery.To address these issues,this paper introduces DAFPN-YOLO,an innovative model based on YOLOv8s(You Only Look Once version 8s).Themodel strikes a balance between detection accuracy and speed while reducing parameters,making itwell-suited for multi-object detection tasks from drone perspectives.A key feature of DAFPN-YOLO is the enhanced Drone-AFPN(Adaptive Feature Pyramid Network),which adaptively fuses multi-scale features to optimize feature extraction and enhance spatial and small-object information.To leverage Drone-AFPN’smulti-scale capabilities fully,a dedicated 160×160 small-object detection head was added,significantly boosting detection accuracy for small targets.In the backbone,the C2f_Dual(Cross Stage Partial with Cross-Stage Feature Fusion Dual)module and SPPELAN(Spatial Pyramid Pooling with Enhanced LocalAttentionNetwork)modulewere integrated.These components improve feature extraction and information aggregationwhile reducing parameters and computational complexity,enhancing inference efficiency.Additionally,Shape-IoU(Shape Intersection over Union)is used as the loss function for bounding box regression,enabling more precise shape-based object matching.Experimental results on the VisDrone 2019 dataset demonstrate the effectiveness ofDAFPN-YOLO.Compared to YOLOv8s,the proposedmodel achieves a 5.4 percentage point increase inmAP@0.5,a 3.8 percentage point improvement in mAP@0.5:0.95,and a 17.2%reduction in parameter count.These results highlight DAFPN-YOLO’s advantages in UAV-based object detection,offering valuable insights for applying deep learning to UAV-specific multi-object detection tasks.展开更多
为提高水域鱼类资源监测的自动化程度和实时分析能力,结合YOLOv8X(You only look once version 8-extra large)目标检测模型、ByteTrack(ByteTrack:a strong baseline for multi-object tracking)算法与双频识别声呐(Dual-frequency ide...为提高水域鱼类资源监测的自动化程度和实时分析能力,结合YOLOv8X(You only look once version 8-extra large)目标检测模型、ByteTrack(ByteTrack:a strong baseline for multi-object tracking)算法与双频识别声呐(Dual-frequency identification sonar,DIDSON)数据,开发了1种快速、准确的鱼类目标识别与计数方法。实验结果表明,YOLOv8X与ByteTrack联合方法与传统的Echoview软件识别精度接近(偏差率仅为1.36%),但处理时间显著减少(单条测线从约30 min减少至约3 min),表现出较强的实时处理能力和泛化性能。同时,通过重复实验验证了该方法的稳定性,确认其在不同场景中的可靠性。本研究方法与成果为水域鱼类资源的自动化监测提供了可靠的技术支持,可广泛地应用于大范围高频次的渔业资源监测与管理工作中。展开更多
针对井工矿复杂环境下安全帽检测面临的挑战,提出了一种改进的YOLOv8目标检测模型。井工矿环境的特殊性,如光线昏暗、粉尘弥漫、背景复杂及矿工姿态多样性,导致现有检测算法在目标遮挡、小目标识别及恶劣环境条件下的检测性能下降。为...针对井工矿复杂环境下安全帽检测面临的挑战,提出了一种改进的YOLOv8目标检测模型。井工矿环境的特殊性,如光线昏暗、粉尘弥漫、背景复杂及矿工姿态多样性,导致现有检测算法在目标遮挡、小目标识别及恶劣环境条件下的检测性能下降。为解决这些问题,从以下几个方面对YOLOv8模型进行了改进:首先,引入残差块+卷积块注意力模块(Residual Block+Convolutional Block Attention Module,ResBlock+CBAM),通过跳跃连接和注意力机制,显著提升了模型对小目标和遮挡目标的检测能力;其次,设计动态检测头,将注意力机制分解为尺度、空间和任务三个独立维度,增强了模型对目标多样性的适应能力;再次,提出用于边界框回归的损失函数——最小点距离的交并比(Minimum Point Distance based Intersection over Union,MPDIoU),综合考虑边界框的重叠面积、中心点距离以及宽度和高度偏差,优化了边界框回归的精度;最后,构建矿井安全帽检测数据集,包含4300张在复杂井下环境下拍摄的图像,用于模型的训练和验证。试验结果表明,改进后的模型在精确率、召回率、mAP50和mAP50-95等关键评价指标上均取得了显著提升。与当前主流的目标检测算法(如Faster R-CNN、SSD、YOLOXs和YOLOv8)相比,提出的模型在检测精度上大幅领先,同时保持了较低的计算复杂度(GFLOPs为16.9),使该模型更适合在实际矿井环境中实时运行。研究不仅提升了井工矿安全帽检测的准确性和鲁棒性,还为复杂环境下的目标检测任务提供了新的思路和方法,具有实际应用价值。展开更多
文摘In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,particularly in snowy environments,remains a challenge.Snow-covered roads introduce unpredictable surface conditions,occlusions,and reduced visibility,that require robust and adaptive path detection algorithms.This paper presents an enhanced road detection framework for snowy environments,leveraging Simple Framework forContrastive Learning of Visual Representations(SimCLR)for Self-Supervised pretraining,hyperparameter optimization,and uncertainty-aware object detection to improve the performance of YouOnly Look Once version 8(YOLOv8).Themodel is trained and evaluated on a custom-built dataset collected from snowy roads in Tromsø,Norway,which covers a range of snow textures,illumination conditions,and road geometries.The proposed framework achieves scores in terms of mAP@50 equal to 99%and mAP@50–95 equal to 97%,demonstrating the effectiveness of YOLOv8 for real-time road detection in extreme winter conditions.The findings contribute to the safe and reliable deployment of autonomous vehicles in Arctic environments,enabling robust decision-making in hazardous weather conditions.This research lays the groundwork for more resilient perceptionmodels in self-driving systems,paving the way for the future development of intelligent and adaptive transportation networks.
基金supported by 2023IT020 of the Industry-University-Research Innovation Fund for Chinese Universities-New Generation Information Technology Innovation ProgramPX-972024121 of the Education&Teaching Reform Program of Guangdong Ocean University。
文摘Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SGW-YOLOv8 based on the improvement of YOLOv8.First,the Adaptive Fine-Grained Channel Attention(FCA)module is fused with the backbone layer of the YOLOv8 network to improve the feature extraction ability of the model.This paper uses the YOLOv8 network backbone layer to improve the feature extraction capability of the model.Second,the Efficient Multi-Scale Attention(C2f_EMA)module is replaced with the C2f module in the Neck layer of the network to improve the detection performance of the model for small underwater targets.Finally,the loss function is optimized to Weighted Intersection over Union(WIoU)to replace the original loss function,so that the model is better adapted to the target detection task in the complex ocean background.The improved algorithm has been experimented with on the Underwater Robot Picking Contest(URPC)dataset,and the results show that the improved algorithm achieves a detection accuracy of 84.5,which is 2.3%higher than that before the improvement,and at the same time,it can accurately detect the small-target marine organisms and adapts to the task of detecting marine organisms in various complex environments.
文摘In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm based on YOLOv8 was proposed in this study.To begin with,the CoordAtt attention mechanism was employed to enhance the feature extraction capability of the backbone network,thereby reducing interference from backgrounds.Additionally,the BiFPN feature fusion network with an added small object detection layer was used to enhance the model's ability to perceive for small objects.Furthermore,a multi-level fusion module was designed and proposed to effectively integrate shallow and deep information.The use of an enhanced MPDIoU loss function further improved detection performance.The experimental results based on the publicly available VisDrone2019 dataset showed that the improved model outperformed the YOLOv8 baseline model,mAP@0.5 improved by 20%,and the improved method improved the detection accuracy of the model for small targets.
基金supported by Natural Science Foundation of Gansu Province(No.22JR5RA320)。
文摘An improved CSYOLOv8 model based on YOLOv8 model is developed specifically for identifying defects in printed circuit board(PCB).Firstly,a composite backbone network is designed to carry out additional feature extraction,which enriches the expression ability of features and enhances the detection accuracy of the model.Secondly,a YOLO-FPN(Feature pyramid network)structure is designed to supplant the original neck network,which enhances the feature fusion ability of the model and improves the detection accuracy of small target objects.Furthermore,to enhance the model’s capability to extract tubular features,dynamic snake convolution is implemented.Finally,MPDIoU loss function is employed to enhance both the convergence rate and the precision of the model.Experiments show that the mAP of the improved model on the PCB defect dataset reaches 96.6%,which is 4.5%higher than that of the YOLOv8 model,and the number of parameters is only 3256862,and the average detection speed is 51.8 frames per second,which meets the requirements of detection accuracy and efficiency.
基金supported by the National Natural Science Foundation of China(Grant Nos.62101275 and 62101274).
文摘UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,complex backgrounds,and variable lighting persist due to the unique perspective of UAV imagery.To address these issues,this paper introduces DAFPN-YOLO,an innovative model based on YOLOv8s(You Only Look Once version 8s).Themodel strikes a balance between detection accuracy and speed while reducing parameters,making itwell-suited for multi-object detection tasks from drone perspectives.A key feature of DAFPN-YOLO is the enhanced Drone-AFPN(Adaptive Feature Pyramid Network),which adaptively fuses multi-scale features to optimize feature extraction and enhance spatial and small-object information.To leverage Drone-AFPN’smulti-scale capabilities fully,a dedicated 160×160 small-object detection head was added,significantly boosting detection accuracy for small targets.In the backbone,the C2f_Dual(Cross Stage Partial with Cross-Stage Feature Fusion Dual)module and SPPELAN(Spatial Pyramid Pooling with Enhanced LocalAttentionNetwork)modulewere integrated.These components improve feature extraction and information aggregationwhile reducing parameters and computational complexity,enhancing inference efficiency.Additionally,Shape-IoU(Shape Intersection over Union)is used as the loss function for bounding box regression,enabling more precise shape-based object matching.Experimental results on the VisDrone 2019 dataset demonstrate the effectiveness ofDAFPN-YOLO.Compared to YOLOv8s,the proposedmodel achieves a 5.4 percentage point increase inmAP@0.5,a 3.8 percentage point improvement in mAP@0.5:0.95,and a 17.2%reduction in parameter count.These results highlight DAFPN-YOLO’s advantages in UAV-based object detection,offering valuable insights for applying deep learning to UAV-specific multi-object detection tasks.
文摘为提高水域鱼类资源监测的自动化程度和实时分析能力,结合YOLOv8X(You only look once version 8-extra large)目标检测模型、ByteTrack(ByteTrack:a strong baseline for multi-object tracking)算法与双频识别声呐(Dual-frequency identification sonar,DIDSON)数据,开发了1种快速、准确的鱼类目标识别与计数方法。实验结果表明,YOLOv8X与ByteTrack联合方法与传统的Echoview软件识别精度接近(偏差率仅为1.36%),但处理时间显著减少(单条测线从约30 min减少至约3 min),表现出较强的实时处理能力和泛化性能。同时,通过重复实验验证了该方法的稳定性,确认其在不同场景中的可靠性。本研究方法与成果为水域鱼类资源的自动化监测提供了可靠的技术支持,可广泛地应用于大范围高频次的渔业资源监测与管理工作中。
文摘针对井工矿复杂环境下安全帽检测面临的挑战,提出了一种改进的YOLOv8目标检测模型。井工矿环境的特殊性,如光线昏暗、粉尘弥漫、背景复杂及矿工姿态多样性,导致现有检测算法在目标遮挡、小目标识别及恶劣环境条件下的检测性能下降。为解决这些问题,从以下几个方面对YOLOv8模型进行了改进:首先,引入残差块+卷积块注意力模块(Residual Block+Convolutional Block Attention Module,ResBlock+CBAM),通过跳跃连接和注意力机制,显著提升了模型对小目标和遮挡目标的检测能力;其次,设计动态检测头,将注意力机制分解为尺度、空间和任务三个独立维度,增强了模型对目标多样性的适应能力;再次,提出用于边界框回归的损失函数——最小点距离的交并比(Minimum Point Distance based Intersection over Union,MPDIoU),综合考虑边界框的重叠面积、中心点距离以及宽度和高度偏差,优化了边界框回归的精度;最后,构建矿井安全帽检测数据集,包含4300张在复杂井下环境下拍摄的图像,用于模型的训练和验证。试验结果表明,改进后的模型在精确率、召回率、mAP50和mAP50-95等关键评价指标上均取得了显著提升。与当前主流的目标检测算法(如Faster R-CNN、SSD、YOLOXs和YOLOv8)相比,提出的模型在检测精度上大幅领先,同时保持了较低的计算复杂度(GFLOPs为16.9),使该模型更适合在实际矿井环境中实时运行。研究不仅提升了井工矿安全帽检测的准确性和鲁棒性,还为复杂环境下的目标检测任务提供了新的思路和方法,具有实际应用价值。