Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOL...Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOLOv8 model for traffic sign detection is proposed.Firstly,by adding Coordinate Attention(CA)to the Backbone,the model gains location information,improving detection accuracy.Secondly,we also introduce EIoU to the localization function to address the ambiguity in aspect ratio descriptions by calculating the width-height difference based on CIoU.Additionally,Focal Loss is incorporated to balance sample difficulty,enhancing regression accuracy.Finally,the model,YOLOv8-CE(YOLOv8-Coordinate Attention-EIoU),is tested on the Jetson Nano,achieving real-time street scene detection and outperforming the Raspberry Pi 4B.Experimental results show that YOLOv8-CE excels in various complex scenarios,improving mAP by 2.8%over the original YOLOv8.The model size and computational effort remain similar,with the Jetson Nano achieving an inference time of 96 ms,significantly faster than the Raspberry Pi 4B.展开更多
Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakt...Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakthroughs in this field,in the face of complex scenes,such as image blur and target occlusion,the traffic sign detection continues to exhibit limited accuracy,accompanied by false positives and missed detections.To address the above problems,a traffic sign detection algorithm,You Only Look Once-based Skip Dynamic Way(YOLO-SDW)based on You Only Look Once version 8 small(YOLOv8s),is proposed.Firstly,a Skip Connection Reconstruction(SCR)module is introduced to efficiently integrate fine-grained feature information and enhance the detection accuracy of the algorithm in complex scenes.Secondly,a C2f module based on Dynamic Snake Convolution(C2f-DySnake)is proposed to dynamically adjust the receptive field information,improve the algorithm’s feature extraction ability for blurred or occluded targets,and reduce the occurrence of false detections and missed detections.Finally,the Wise Powerful IoU v2(WPIoUv2)loss function is proposed to further improve the detection accuracy of the algorithm.Experimental results show that the average precision mAP@0.5 of YOLO-SDW on the TT100K dataset is 89.2%,and mAP@0.5:0.95 is 68.5%,which is 4%and 3.3%higher than the YOLOv8s baseline,respectively.YOLO-SDW ensures real-time performance while having higher accuracy.展开更多
In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds...In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds in current intelligent inspection algorithms,this paper proposes CG-YOLOv8,a lightweight and improved model based on YOLOv8n for PCB surface defect detection.The proposed method optimizes the network architecture and compresses parameters to reduce model complexity while maintaining high detection accuracy,thereby enhancing the capability of identifying diverse defects under complex conditions.Specifically,a cascaded multi-receptive field(CMRF)module is adopted to replace the SPPF module in the backbone to improve feature perception,and an inverted residual mobile block(IRMB)is integrated into the C2f module to further enhance performance.Additionally,conventional convolution layers are replaced with GSConv to reduce computational cost,and a lightweight Convolutional Block Attention Module based Convolution(CBAMConv)module is introduced after Grouped Spatial Convolution(GSConv)to preserve accuracy through attention mechanisms.The detection head is also optimized by removing medium and large-scale detection layers,thereby enhancing the model’s ability to detect small-scale defects and further reducing complexity.Experimental results show that,compared to the original YOLOv8n,the proposed CG-YOLOv8 reduces parameter count by 53.9%,improves mAP@0.5 by 2.2%,and increases precision and recall by 2.0%and 1.8%,respectively.These improvements demonstrate that CG-YOLOv8 offers an efficient and lightweight solution for PCB surface defect detection.展开更多
为精准识别与分类不同花期杭白菊,满足自动化采摘要求,该研究提出一种基于改进YOLOv8s的杭白菊检测模型-YOLOv8s-RDL。首先,该研究将颈部网络(neck)的C2f(faster implementation of CSP bottleneck with 2 convolutions)模块替换为RCS-O...为精准识别与分类不同花期杭白菊,满足自动化采摘要求,该研究提出一种基于改进YOLOv8s的杭白菊检测模型-YOLOv8s-RDL。首先,该研究将颈部网络(neck)的C2f(faster implementation of CSP bottleneck with 2 convolutions)模块替换为RCS-OSA(one-shot aggregation of reparameterized convolution based on channel shuffle)模块,以提升骨干网络(backbone)特征融合效率;其次,将检测头更换为DyHead(dynamic head),并融合DCNv3(deformable convolutional networks v3),借助多头自注意力机制增强目标检测头的表达能力;最后,采用LAMP(layer-adaptive magnitude-based pruning)通道剪枝算法减少参数量,降低模型复杂度。试验结果表明,YOLOv8s-RDL模型在菊米和胎菊的花期分类中平均精度分别达到96.3%和97.7%,相较于YOLOv8s模型,分别提升了3.8和1.5个百分点,同时权重文件大小较YOLOv8s减小了6 MB。该研究引入TIDE(toolkit for identifying detection and segmentation errors)评估指标,结果显示,YOLOv8s-RDL模型分类错误和背景检测错误相较YOLOv8s模型分别降低0.55和1.26。该研究为杭白菊分花期自动化采摘提供了理论依据和技术支撑。展开更多
在林业管理中,及时发现火灾并识别其规模对于安全防护和治理火灾至关重要。针对现有火灾检测算法存在的精度低、漏检误检和实时性不足等问题,提出一种无人机航拍图像下火灾实时检测算法——MDSYOLOv8。以YOLOv8为基线算法,将骨干网络第...在林业管理中,及时发现火灾并识别其规模对于安全防护和治理火灾至关重要。针对现有火灾检测算法存在的精度低、漏检误检和实时性不足等问题,提出一种无人机航拍图像下火灾实时检测算法——MDSYOLOv8。以YOLOv8为基线算法,将骨干网络第7层卷积模块和颈部网络卷积模块替换成动态蛇形卷积(DSConv),提高算法的特征提取性能,并强化算法对微小特征的学习能力;然后在颈部与检测头之间添加多维协作注意力机制(MCA),加强颈部特征融合,增强算法对小目标的检测能力,并抑制无关背景信息;最后使用SIoU损失函数替换原YOLOv8中的CIoU损失函数,加快算法的收敛速度和回归精度。实验结果表明,MDSYOLOv8在公开数据集KMU上对烟雾目标的检测精度mAP达到95.89%,相较于基线YOLOv8提高了3.33个百分点,具有卓越的检测性能。此外,本研究采集互联网上的无人机航拍火灾图像制作UFF(UAV field fire)数据集,主要对象为火焰和烟雾,包含森林和城市等火灾隐患可能发生场景。在自制数据集UFF上进行深度实验分析,MDSYOLOv8的检测精度达到93.98%,检测速度为54帧/s,并且能同时识别烟雾和火焰两种火灾场景中的主要目标,与主流目标检测方法相比,在检测精度和效率方面均展现出明显优势,更加契合航拍场景下的火灾检测应用。展开更多
基金supported by Heilongjiang Provincial Natural Science Foundation of China(LH2023E055)the National Key R&D Program of China(2021YFB2600502).
文摘Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOLOv8 model for traffic sign detection is proposed.Firstly,by adding Coordinate Attention(CA)to the Backbone,the model gains location information,improving detection accuracy.Secondly,we also introduce EIoU to the localization function to address the ambiguity in aspect ratio descriptions by calculating the width-height difference based on CIoU.Additionally,Focal Loss is incorporated to balance sample difficulty,enhancing regression accuracy.Finally,the model,YOLOv8-CE(YOLOv8-Coordinate Attention-EIoU),is tested on the Jetson Nano,achieving real-time street scene detection and outperforming the Raspberry Pi 4B.Experimental results show that YOLOv8-CE excels in various complex scenarios,improving mAP by 2.8%over the original YOLOv8.The model size and computational effort remain similar,with the Jetson Nano achieving an inference time of 96 ms,significantly faster than the Raspberry Pi 4B.
基金funded by Key research and development Program of Henan Province(No.251111211200)National Natural Science Foundation of China(Grant No.U2004163).
文摘Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakthroughs in this field,in the face of complex scenes,such as image blur and target occlusion,the traffic sign detection continues to exhibit limited accuracy,accompanied by false positives and missed detections.To address the above problems,a traffic sign detection algorithm,You Only Look Once-based Skip Dynamic Way(YOLO-SDW)based on You Only Look Once version 8 small(YOLOv8s),is proposed.Firstly,a Skip Connection Reconstruction(SCR)module is introduced to efficiently integrate fine-grained feature information and enhance the detection accuracy of the algorithm in complex scenes.Secondly,a C2f module based on Dynamic Snake Convolution(C2f-DySnake)is proposed to dynamically adjust the receptive field information,improve the algorithm’s feature extraction ability for blurred or occluded targets,and reduce the occurrence of false detections and missed detections.Finally,the Wise Powerful IoU v2(WPIoUv2)loss function is proposed to further improve the detection accuracy of the algorithm.Experimental results show that the average precision mAP@0.5 of YOLO-SDW on the TT100K dataset is 89.2%,and mAP@0.5:0.95 is 68.5%,which is 4%and 3.3%higher than the YOLOv8s baseline,respectively.YOLO-SDW ensures real-time performance while having higher accuracy.
基金funded by the Joint Funds of the National Natural Science Foundation of China(U2341223)the Beijing Municipal Natural Science Foundation(No.4232067).
文摘In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds in current intelligent inspection algorithms,this paper proposes CG-YOLOv8,a lightweight and improved model based on YOLOv8n for PCB surface defect detection.The proposed method optimizes the network architecture and compresses parameters to reduce model complexity while maintaining high detection accuracy,thereby enhancing the capability of identifying diverse defects under complex conditions.Specifically,a cascaded multi-receptive field(CMRF)module is adopted to replace the SPPF module in the backbone to improve feature perception,and an inverted residual mobile block(IRMB)is integrated into the C2f module to further enhance performance.Additionally,conventional convolution layers are replaced with GSConv to reduce computational cost,and a lightweight Convolutional Block Attention Module based Convolution(CBAMConv)module is introduced after Grouped Spatial Convolution(GSConv)to preserve accuracy through attention mechanisms.The detection head is also optimized by removing medium and large-scale detection layers,thereby enhancing the model’s ability to detect small-scale defects and further reducing complexity.Experimental results show that,compared to the original YOLOv8n,the proposed CG-YOLOv8 reduces parameter count by 53.9%,improves mAP@0.5 by 2.2%,and increases precision and recall by 2.0%and 1.8%,respectively.These improvements demonstrate that CG-YOLOv8 offers an efficient and lightweight solution for PCB surface defect detection.
文摘为精准识别与分类不同花期杭白菊,满足自动化采摘要求,该研究提出一种基于改进YOLOv8s的杭白菊检测模型-YOLOv8s-RDL。首先,该研究将颈部网络(neck)的C2f(faster implementation of CSP bottleneck with 2 convolutions)模块替换为RCS-OSA(one-shot aggregation of reparameterized convolution based on channel shuffle)模块,以提升骨干网络(backbone)特征融合效率;其次,将检测头更换为DyHead(dynamic head),并融合DCNv3(deformable convolutional networks v3),借助多头自注意力机制增强目标检测头的表达能力;最后,采用LAMP(layer-adaptive magnitude-based pruning)通道剪枝算法减少参数量,降低模型复杂度。试验结果表明,YOLOv8s-RDL模型在菊米和胎菊的花期分类中平均精度分别达到96.3%和97.7%,相较于YOLOv8s模型,分别提升了3.8和1.5个百分点,同时权重文件大小较YOLOv8s减小了6 MB。该研究引入TIDE(toolkit for identifying detection and segmentation errors)评估指标,结果显示,YOLOv8s-RDL模型分类错误和背景检测错误相较YOLOv8s模型分别降低0.55和1.26。该研究为杭白菊分花期自动化采摘提供了理论依据和技术支撑。
文摘在林业管理中,及时发现火灾并识别其规模对于安全防护和治理火灾至关重要。针对现有火灾检测算法存在的精度低、漏检误检和实时性不足等问题,提出一种无人机航拍图像下火灾实时检测算法——MDSYOLOv8。以YOLOv8为基线算法,将骨干网络第7层卷积模块和颈部网络卷积模块替换成动态蛇形卷积(DSConv),提高算法的特征提取性能,并强化算法对微小特征的学习能力;然后在颈部与检测头之间添加多维协作注意力机制(MCA),加强颈部特征融合,增强算法对小目标的检测能力,并抑制无关背景信息;最后使用SIoU损失函数替换原YOLOv8中的CIoU损失函数,加快算法的收敛速度和回归精度。实验结果表明,MDSYOLOv8在公开数据集KMU上对烟雾目标的检测精度mAP达到95.89%,相较于基线YOLOv8提高了3.33个百分点,具有卓越的检测性能。此外,本研究采集互联网上的无人机航拍火灾图像制作UFF(UAV field fire)数据集,主要对象为火焰和烟雾,包含森林和城市等火灾隐患可能发生场景。在自制数据集UFF上进行深度实验分析,MDSYOLOv8的检测精度达到93.98%,检测速度为54帧/s,并且能同时识别烟雾和火焰两种火灾场景中的主要目标,与主流目标检测方法相比,在检测精度和效率方面均展现出明显优势,更加契合航拍场景下的火灾检测应用。