In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,par...In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,particularly in snowy environments,remains a challenge.Snow-covered roads introduce unpredictable surface conditions,occlusions,and reduced visibility,that require robust and adaptive path detection algorithms.This paper presents an enhanced road detection framework for snowy environments,leveraging Simple Framework forContrastive Learning of Visual Representations(SimCLR)for Self-Supervised pretraining,hyperparameter optimization,and uncertainty-aware object detection to improve the performance of YouOnly Look Once version 8(YOLOv8).Themodel is trained and evaluated on a custom-built dataset collected from snowy roads in Tromsø,Norway,which covers a range of snow textures,illumination conditions,and road geometries.The proposed framework achieves scores in terms of mAP@50 equal to 99%and mAP@50–95 equal to 97%,demonstrating the effectiveness of YOLOv8 for real-time road detection in extreme winter conditions.The findings contribute to the safe and reliable deployment of autonomous vehicles in Arctic environments,enabling robust decision-making in hazardous weather conditions.This research lays the groundwork for more resilient perceptionmodels in self-driving systems,paving the way for the future development of intelligent and adaptive transportation networks.展开更多
UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,comp...UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,complex backgrounds,and variable lighting persist due to the unique perspective of UAV imagery.To address these issues,this paper introduces DAFPN-YOLO,an innovative model based on YOLOv8s(You Only Look Once version 8s).Themodel strikes a balance between detection accuracy and speed while reducing parameters,making itwell-suited for multi-object detection tasks from drone perspectives.A key feature of DAFPN-YOLO is the enhanced Drone-AFPN(Adaptive Feature Pyramid Network),which adaptively fuses multi-scale features to optimize feature extraction and enhance spatial and small-object information.To leverage Drone-AFPN’smulti-scale capabilities fully,a dedicated 160×160 small-object detection head was added,significantly boosting detection accuracy for small targets.In the backbone,the C2f_Dual(Cross Stage Partial with Cross-Stage Feature Fusion Dual)module and SPPELAN(Spatial Pyramid Pooling with Enhanced LocalAttentionNetwork)modulewere integrated.These components improve feature extraction and information aggregationwhile reducing parameters and computational complexity,enhancing inference efficiency.Additionally,Shape-IoU(Shape Intersection over Union)is used as the loss function for bounding box regression,enabling more precise shape-based object matching.Experimental results on the VisDrone 2019 dataset demonstrate the effectiveness ofDAFPN-YOLO.Compared to YOLOv8s,the proposedmodel achieves a 5.4 percentage point increase inmAP@0.5,a 3.8 percentage point improvement in mAP@0.5:0.95,and a 17.2%reduction in parameter count.These results highlight DAFPN-YOLO’s advantages in UAV-based object detection,offering valuable insights for applying deep learning to UAV-specific multi-object detection tasks.展开更多
Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SG...Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SGW-YOLOv8 based on the improvement of YOLOv8.First,the Adaptive Fine-Grained Channel Attention(FCA)module is fused with the backbone layer of the YOLOv8 network to improve the feature extraction ability of the model.This paper uses the YOLOv8 network backbone layer to improve the feature extraction capability of the model.Second,the Efficient Multi-Scale Attention(C2f_EMA)module is replaced with the C2f module in the Neck layer of the network to improve the detection performance of the model for small underwater targets.Finally,the loss function is optimized to Weighted Intersection over Union(WIoU)to replace the original loss function,so that the model is better adapted to the target detection task in the complex ocean background.The improved algorithm has been experimented with on the Underwater Robot Picking Contest(URPC)dataset,and the results show that the improved algorithm achieves a detection accuracy of 84.5,which is 2.3%higher than that before the improvement,and at the same time,it can accurately detect the small-target marine organisms and adapts to the task of detecting marine organisms in various complex environments.展开更多
To address critical challenges in nighttime ship detection—high small-target missed detection(over 20%),insufficient lightweighting,and limited generalization due to scarce,low-quality datasets—this study proposes a...To address critical challenges in nighttime ship detection—high small-target missed detection(over 20%),insufficient lightweighting,and limited generalization due to scarce,low-quality datasets—this study proposes a systematic solution.First,a high-quality Night-Ships dataset is constructed via CycleGAN-based day-night transfer,combined with a dual-threshold cleaning strategy(Laplacian variance sharpness filtering and brightness-color deviation screening).Second,a Cross-stage Lightweight Fusion-You Only Look Once version 8(CLF-YOLOv8)is proposed with key improvements:the Neck network is reconstructed by replacing Cross Stage Partial(CSP)structure with the Cross Stage Partial Multi-Scale Convolutional Block(CSP-MSCB)and integrating Bidirectional Feature Pyramid Network(BiFPN)for weighted multi-scale fusion to enhance small-target detection;a Lightweight Shared Convolutional and Separated Batch Normalization Detection-Head(LSCSBD-Head)with shared convolutions and layer-wise Batch Normalization(BN)reduces parameters to 1.8M(42% fewer than YOLOv8n);and the FocalMinimum Point Distance Intersection over Union(Focal-MPDIoU)loss combines Minimum Point Distance Intersection over Union(MPDIoU)geometric constraints and Focal weighting to optimize low-overlap targets.Experiments show CLFYOLOv8 achieves 97.6%mAP@0.5(0.7% higher than YOLOv8n)with 1.8 M parameters,outperforming mainstream models in small-target detection,overlapping target discrimination,and adaptability to complex lighting.展开更多
Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakt...Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakthroughs in this field,in the face of complex scenes,such as image blur and target occlusion,the traffic sign detection continues to exhibit limited accuracy,accompanied by false positives and missed detections.To address the above problems,a traffic sign detection algorithm,You Only Look Once-based Skip Dynamic Way(YOLO-SDW)based on You Only Look Once version 8 small(YOLOv8s),is proposed.Firstly,a Skip Connection Reconstruction(SCR)module is introduced to efficiently integrate fine-grained feature information and enhance the detection accuracy of the algorithm in complex scenes.Secondly,a C2f module based on Dynamic Snake Convolution(C2f-DySnake)is proposed to dynamically adjust the receptive field information,improve the algorithm’s feature extraction ability for blurred or occluded targets,and reduce the occurrence of false detections and missed detections.Finally,the Wise Powerful IoU v2(WPIoUv2)loss function is proposed to further improve the detection accuracy of the algorithm.Experimental results show that the average precision mAP@0.5 of YOLO-SDW on the TT100K dataset is 89.2%,and mAP@0.5:0.95 is 68.5%,which is 4%and 3.3%higher than the YOLOv8s baseline,respectively.YOLO-SDW ensures real-time performance while having higher accuracy.展开更多
In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds...In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds in current intelligent inspection algorithms,this paper proposes CG-YOLOv8,a lightweight and improved model based on YOLOv8n for PCB surface defect detection.The proposed method optimizes the network architecture and compresses parameters to reduce model complexity while maintaining high detection accuracy,thereby enhancing the capability of identifying diverse defects under complex conditions.Specifically,a cascaded multi-receptive field(CMRF)module is adopted to replace the SPPF module in the backbone to improve feature perception,and an inverted residual mobile block(IRMB)is integrated into the C2f module to further enhance performance.Additionally,conventional convolution layers are replaced with GSConv to reduce computational cost,and a lightweight Convolutional Block Attention Module based Convolution(CBAMConv)module is introduced after Grouped Spatial Convolution(GSConv)to preserve accuracy through attention mechanisms.The detection head is also optimized by removing medium and large-scale detection layers,thereby enhancing the model’s ability to detect small-scale defects and further reducing complexity.Experimental results show that,compared to the original YOLOv8n,the proposed CG-YOLOv8 reduces parameter count by 53.9%,improves mAP@0.5 by 2.2%,and increases precision and recall by 2.0%and 1.8%,respectively.These improvements demonstrate that CG-YOLOv8 offers an efficient and lightweight solution for PCB surface defect detection.展开更多
Object detection,a major challenge in computer vision and pattern recognition,plays a significant part in many applications,crossing artificial intelligence,face recognition,and autonomous driving.It involves focusing...Object detection,a major challenge in computer vision and pattern recognition,plays a significant part in many applications,crossing artificial intelligence,face recognition,and autonomous driving.It involves focusing on identifying the detection,localization,and categorization of targets in images.A particularly important emerging task is distinguishing real animals from toy replicas in real-time,mostly for smart camera systems in both urban and natural environments.However,that difficult task is affected by factors such as showing angle,occlusion,light intensity,variations,and texture differences.To tackle these challenges,this paper recommends Group Sparse YOLOv8(You Only Look Once version 8),an improved real-time object detection algorithm that improves YOLOv8 by integrating group sparsity regularization.This adjustment improves efficiency and accuracy while utilizing the computational costs and power consumption,including a frame selection approach.And a hybrid parallel processing method that merges pipelining with dataflow strategies to improve the performance.Established using a custom dataset of toy and real animal images along with well-known datasets,namely ImageNet,MSCOCO,and CIFAR-10/100.The combination of Group Sparsity with YOLOv8 shows high detection accuracy with lower latency.Here provides a real and resource-efficient solution for intelligent camera systems and improves real-time object detection and classification in environments,differentiating between real and toy animals.展开更多
With the gradual development of automatic driving technology,people’s attention is no longer limited to daily automatic driving target detection.In response to the problem that it is difficult to achieve fast and acc...With the gradual development of automatic driving technology,people’s attention is no longer limited to daily automatic driving target detection.In response to the problem that it is difficult to achieve fast and accurate detection of visual targets in complex scenes of automatic driving at night,a detection algorithm based on improved YOLOv8s was proposed.Firsly,By adding Triplet Attention module into the lower sampling layer of the original model,the model can effectively retain and enhance feature information related to target detection on the lower-resolution feature map.This enhancement improved the robustness of the target detection network and reduced instances of missed detections.Secondly,the Soft-NMS algorithm was introduced to address the challenges of dealing with dense targets,overlapping objects,and complex scenes.This algorithm effectively reduced false and missed positives,thereby improved overall detection performance when faced with highly overlapping detection results.Finally,the experimental results on the MPDIoU loss function dataset showed that compared with the original model,the improved method,in which mAP and accuracy are increased by 2.9%and 2.8%respectively,can achieve better detection accuracy and speed in night vehicle detection.It can effectively improve the problem of target detection in night scenes.展开更多
In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm...In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm based on YOLOv8 was proposed in this study.To begin with,the CoordAtt attention mechanism was employed to enhance the feature extraction capability of the backbone network,thereby reducing interference from backgrounds.Additionally,the BiFPN feature fusion network with an added small object detection layer was used to enhance the model's ability to perceive for small objects.Furthermore,a multi-level fusion module was designed and proposed to effectively integrate shallow and deep information.The use of an enhanced MPDIoU loss function further improved detection performance.The experimental results based on the publicly available VisDrone2019 dataset showed that the improved model outperformed the YOLOv8 baseline model,mAP@0.5 improved by 20%,and the improved method improved the detection accuracy of the model for small targets.展开更多
Pulmonary nodules represent an early manifestation of lung cancer.However,pulmonary nodules only constitute a small portion of the overall image,posing challenges for physicians in image interpretation and potentially...Pulmonary nodules represent an early manifestation of lung cancer.However,pulmonary nodules only constitute a small portion of the overall image,posing challenges for physicians in image interpretation and potentially leading to false positives or missed detections.To solve these problems,the YOLOv8 network is enhanced by adding deformable convolution and atrous spatial pyramid pooling(ASPP),along with the integration of a coordinate attention(CA)mechanism.This allows the network to focus on small targets while expanding the receptive field without losing resolution.At the same time,context information on the target is gathered and feature expression is enhanced by attention modules in different directions.It effectively improves the positioning accuracy and achieves good results on the LUNA16 dataset.Compared with other detection algorithms,it improves the accuracy of pulmonary nodule detection to a certain extent.展开更多
Aiming at the problem that the existing algorithms for vehicle detection in smart factories are difficult to detect partial occlusion of vehicles,vulnerable to background interference,lack of global vision,and excessi...Aiming at the problem that the existing algorithms for vehicle detection in smart factories are difficult to detect partial occlusion of vehicles,vulnerable to background interference,lack of global vision,and excessive suppression of real targets,which ultimately cause accuracy degradation.At the same time,to facilitate the subsequent positioning of vehicles in the factory,this paper proposes an improved YOLOv8 algorithm.Firstly,the RFCAConv module is combined to improve the original YOLOv8 backbone.Pay attention to the different features in the receptive field,and give priority to the spatial features of the receptive field to capture more vehicle feature information and solve the problem that the vehicle is partially occluded and difficult to detect.Secondly,the SFE module is added to the neck of v8,which improves the saliency of the target in the reasoning process and reduces the influence of background interference on vehicle detection.Finally,the head of the RT-DETR algorithm is used to replace the head in the original YOLOv8 algorithm,which avoids the excessive suppression of the real target while combining the context information.The experimental results show that compared with the original YOLOv8 algorithm,the detection accuracy of the improved YOLOv8 algorithm is improved by 4.6%on the self-made smart factory data set,and the detection speed also meets the real-time requirements of smart factory vehicle detection and subsequent vehicle positioning.展开更多
Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOL...Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOLOv8 model for traffic sign detection is proposed.Firstly,by adding Coordinate Attention(CA)to the Backbone,the model gains location information,improving detection accuracy.Secondly,we also introduce EIoU to the localization function to address the ambiguity in aspect ratio descriptions by calculating the width-height difference based on CIoU.Additionally,Focal Loss is incorporated to balance sample difficulty,enhancing regression accuracy.Finally,the model,YOLOv8-CE(YOLOv8-Coordinate Attention-EIoU),is tested on the Jetson Nano,achieving real-time street scene detection and outperforming the Raspberry Pi 4B.Experimental results show that YOLOv8-CE excels in various complex scenarios,improving mAP by 2.8%over the original YOLOv8.The model size and computational effort remain similar,with the Jetson Nano achieving an inference time of 96 ms,significantly faster than the Raspberry Pi 4B.展开更多
For the characteristics of small,dense distribution,high diversity of defects and high precision and fast detection in the process of PCB(Printed Circuit Board)defect detection,a defect detection algorithm based on YO...For the characteristics of small,dense distribution,high diversity of defects and high precision and fast detection in the process of PCB(Printed Circuit Board)defect detection,a defect detection algorithm based on YOLOv8n was proposed in this study.Firstly,the original C2f module of YOLOv8n was improved into a C2FFaster-EMA module to reduce the number of parameters and floating-point operations(FLOPs).Additionally,the WIoUv3 loss function was introduced to mitigate the negative impact of low-quality defect images on model training.Consequently,a reduction in model size and an enhancement in detection precision were achieved.Finally,the ablation and comparative experiments were conducted on an augmented Deep PCB dataset,and the generalization experiments were performed on the PCB Defect-Augmented dataset.The results indicated that the proposed model reduces the number of parameters by 23.3%and FLOPs by 20%,P by 0.7%,mAP@0.5 by 0.3%,and mAP@0.5:0.95 by 3.9%,respectively,compared to the original YOLOv8n model.Furthermore,the comparative experiments demonstrated that the proposed model achieves higher accuracy and mAP compared to YOLOv5n and YOLOv5s.It was concluded that the proposed method satisfies the requirements for both accuracy and speed in PCB defect detection.展开更多
In recent years,with the development of synthetic aperture radar(SAR)technology and the widespread application of deep learning,lightweight detection of SAR images has emerged as a research direction.The ultimate goal...In recent years,with the development of synthetic aperture radar(SAR)technology and the widespread application of deep learning,lightweight detection of SAR images has emerged as a research direction.The ultimate goal is to reduce computational and storage requirements while ensuring detection accuracy and reliability,making it an ideal choice for achieving rapid response and efficient processing.In this regard,a lightweight SAR ship target detection algorithm based on YOLOv8 was proposed in this study.Firstly,the C2f-Sc module was designed by fusing the C2f in the backbone network with the ScConv to reduce spatial redundancy and channel redundancy between features in convolutional neural networks.At the same time,the Ghost module was introduced into the neck network to effectively reduce model parameters and computational complexity.A relatively lightweight EMA attention mechanism was added to the neck network to promote the effective fusion of features at different levels.Experimental results showed that the Parameters and GFLOPs of the improved model are reduced by 8.5%and 7.0%when mAP@0.5 and mAP@0.5:0.95 are increased by 0.7%and 1.8%,respectively.It makes the model lightweight and improves the detection accuracy,which has certain application value.展开更多
Synthetic Aperture Radar(SAR)has become one of the most effective tools in ship detection.However,due to significant background interference,small targets,and challenges related to target scattering intensity in SAR i...Synthetic Aperture Radar(SAR)has become one of the most effective tools in ship detection.However,due to significant background interference,small targets,and challenges related to target scattering intensity in SAR images,current ship target detection faces serious issues of missed detections and false positives,and the network structures are overly complex.To address this issue,this paper proposes a lightweight model based on YOLOv8,named OD-YOLOv8.Firstly,we adopt a simplified neural network architecture,VanillaNet,to replace the backbone network,significantly reducing the number of parameters and computational complexity while ensuring accuracy.Secondly,we introduce a dynamic,multi-dimensional attention mechanism by designing the ODC2f module with ODConv to replace the original C2f module and using GSConv to replace two down-sampling convolutions to reduce the number of parameters.Then,to alleviate the issues of missed detections and false positives for small targets,we discard one of the original large target detection layers and add a detection layer specifically for small targets.Finally,based on a dynamic non-monotonic focusing mechanism,we employ the Wise-IoU(Intersection over Union)loss function to significantly improve detection accuracy.Experimental results on the HRSID dataset show that,compared to the original YOLOv8,OD-YOLOv8 improves mAP@0.5 and mAP@0.5–0.95 by 2.7%and 3.5%,respectively,while reducing the number of parameters and GFLOPs by 72.9%and 4.9%,respectively.Moreover,the model also performs exceptionally well on the SSDD dataset,with AP and AP50 increasing by 1.7%and 0.4%,respectively.OD-YOLOv8 achieves an excellent balance between model lightweightness and accuracy,making it highly valuable for end-to-end industrial deployment.展开更多
Addressing challenges in accurately detecting persimmon fruit quality in orchards—such as reliance on manual grading,low efficiency,severe foliage obstruction,and subtle differences between quality grades—this paper...Addressing challenges in accurately detecting persimmon fruit quality in orchards—such as reliance on manual grading,low efficiency,severe foliage obstruction,and subtle differences between quality grades—this paper proposes a lightweight persimmon detection model based on an improved YOLOv8s architecture.First,the Conv layer in the backbone network is replaced with an ADown module to reduce model complexity.Second,MSFAN is introduced in the Neck layer to fully extract texture features from persimmon images,highlighting differences between quality grades.Finally,the Wise-IoU loss function mitigates the impact of low-quality sample data on grading accuracy.The improved model accurately identifies and separates persimmons of varying quality,effectively addressing quality grading detection in complex backgrounds.This provides a viable technical approach for achieving persimmon quality grading detection.展开更多
This study systematically addresses the limitations of traditional pest detection methods and proposes an optimized version of the YOLOv8 object detection model.By integrating the GhostConv convolution module and the ...This study systematically addresses the limitations of traditional pest detection methods and proposes an optimized version of the YOLOv8 object detection model.By integrating the GhostConv convolution module and the C3Ghost module,the Polarized Self-Attention(PSA)mechanism is incorporated to enhance the model’s capacity for extracting pest features.Experimental results demonstrate that the improved YOLOv8+Ghost+PSA model achieves outstanding performance in critical metrics such as precision,recall,and mean Average Precision(mAP),with a computational cost of only 5.3 GFLOPs,making it highly suitable for deployment in resource-constrained agricultural environments.展开更多
In response to themissed and false detections that are easily caused by the large variety of and significant differences among aluminum surface defects,a detection algorithm based on an improved You Only Look Once(YOL...In response to themissed and false detections that are easily caused by the large variety of and significant differences among aluminum surface defects,a detection algorithm based on an improved You Only Look Once(YOLO)v8n network is proposed.First,a C2f_DWR_DRB module is constructed by introducing a dilation-wise residual(DWR)module and a dilated reparameterization block(DRB)to replace the C2f module at the high level of the backbone network,enriching the gradient flow information and increasing the effective receptive field(ERF).Second,an efficient local attention(ELA)mechanism is fused with the high-level screening-feature pyramid networks(HS-FPN)module,and an ELA_HSFPN is designed to replace the original feature fusion module,enhancing the ability of the network to cope with multiscale detection tasks.Moreover,a lightweight shared convolutional detection head(SCDH)is introduced to reduce the number of parameters and the computational complexity of the module while enhancing the performance and generalizability of the model.Finally,the soft intersection over union(SIoU)replaces the original loss function to improve the convergence speed and prediction accuracy of the model.Experimental results show that compared with that of the original YOLOv8n model,the mAP@0.5 of the improved algorithm is increased by 5.1%,the number of parameters and computational complexity are reduced by 33.3%and 32.1%,respectively,and the FPS is increased by 4.9%.Compared with other mainstream object detection algorithms,the improved algorithm still leads in terms of core indicators and has good generalizability for surface defects encountered in other industrial scenarios.展开更多
The service life of internal combustion engines is significantly influenced by surface defects in cylinder liners.To address the limitations of traditional detection methods,we propose an enhanced YOLOv8 model with Sw...The service life of internal combustion engines is significantly influenced by surface defects in cylinder liners.To address the limitations of traditional detection methods,we propose an enhanced YOLOv8 model with Swin Transformer as the backbone network.This approach leverages Swin Transformer's multi-head self-attention mechanism for improved feature extraction of defects spanning various scales.Integrated with the YOLOv8 detection head,our model achieves a mean average precision of 85.1%on our dataset,outperforming baseline methods by 1.4%.The model's effectiveness is further demonstrated on a steel-surface defect dataset,indicating its broad applicability in industrial surface defect detection.Our work highlights the potential of combining Swin Transformer and YOLOv8 for accurate and efficient defect detection.展开更多
In response to challenges posed by complex backgrounds,diverse target angles,and numerous small targets in remote sensing images,alongside the issue of high resource consumption hindering model deployment,we propose a...In response to challenges posed by complex backgrounds,diverse target angles,and numerous small targets in remote sensing images,alongside the issue of high resource consumption hindering model deployment,we propose an enhanced,lightweight you only look once version 8 small(YOLOv8s)detection algorithm.Regarding network improvements,we first replace tradi-tional horizontal boxes with rotated boxes for target detection,effectively addressing difficulties in feature extraction caused by varying target angles.Second,we design a module integrating convolu-tional neural networks(CNN)and Transformer components to replace specific C2f modules in the backbone network,thereby expanding the model’s receptive field and enhancing feature extraction in complex backgrounds.Finally,we introduce a feature calibration structure to mitigate potential feature mismatches during feature fusion.For model compression,we employ a lightweight channel pruning technique based on localized mean average precision(LMAP)to eliminate redundancies in the enhanced model.Although this approach results in some loss of detection accuracy,it effec-tively reduces the number of parameters,computational load,and model size.Additionally,we employ channel-level knowledge distillation to recover accuracy in the pruned model,further enhancing detection performance.Experimental results indicate that the enhanced algorithm achieves a 6.1%increase in mAP50 compared to YOLOv8s,while simultaneously reducing parame-ters,computational load,and model size by 57.7%,28.8%,and 52.3%,respectively.展开更多
文摘In recent years,advancements in autonomous vehicle technology have accelerated,promising safer and more efficient transportation systems.However,achieving fully autonomous driving in challenging weather conditions,particularly in snowy environments,remains a challenge.Snow-covered roads introduce unpredictable surface conditions,occlusions,and reduced visibility,that require robust and adaptive path detection algorithms.This paper presents an enhanced road detection framework for snowy environments,leveraging Simple Framework forContrastive Learning of Visual Representations(SimCLR)for Self-Supervised pretraining,hyperparameter optimization,and uncertainty-aware object detection to improve the performance of YouOnly Look Once version 8(YOLOv8).Themodel is trained and evaluated on a custom-built dataset collected from snowy roads in Tromsø,Norway,which covers a range of snow textures,illumination conditions,and road geometries.The proposed framework achieves scores in terms of mAP@50 equal to 99%and mAP@50–95 equal to 97%,demonstrating the effectiveness of YOLOv8 for real-time road detection in extreme winter conditions.The findings contribute to the safe and reliable deployment of autonomous vehicles in Arctic environments,enabling robust decision-making in hazardous weather conditions.This research lays the groundwork for more resilient perceptionmodels in self-driving systems,paving the way for the future development of intelligent and adaptive transportation networks.
基金supported by the National Natural Science Foundation of China(Grant Nos.62101275 and 62101274).
文摘UAV-based object detection is rapidly expanding in both civilian and military applications,including security surveillance,disaster assessment,and border patrol.However,challenges such as small objects,occlusions,complex backgrounds,and variable lighting persist due to the unique perspective of UAV imagery.To address these issues,this paper introduces DAFPN-YOLO,an innovative model based on YOLOv8s(You Only Look Once version 8s).Themodel strikes a balance between detection accuracy and speed while reducing parameters,making itwell-suited for multi-object detection tasks from drone perspectives.A key feature of DAFPN-YOLO is the enhanced Drone-AFPN(Adaptive Feature Pyramid Network),which adaptively fuses multi-scale features to optimize feature extraction and enhance spatial and small-object information.To leverage Drone-AFPN’smulti-scale capabilities fully,a dedicated 160×160 small-object detection head was added,significantly boosting detection accuracy for small targets.In the backbone,the C2f_Dual(Cross Stage Partial with Cross-Stage Feature Fusion Dual)module and SPPELAN(Spatial Pyramid Pooling with Enhanced LocalAttentionNetwork)modulewere integrated.These components improve feature extraction and information aggregationwhile reducing parameters and computational complexity,enhancing inference efficiency.Additionally,Shape-IoU(Shape Intersection over Union)is used as the loss function for bounding box regression,enabling more precise shape-based object matching.Experimental results on the VisDrone 2019 dataset demonstrate the effectiveness ofDAFPN-YOLO.Compared to YOLOv8s,the proposedmodel achieves a 5.4 percentage point increase inmAP@0.5,a 3.8 percentage point improvement in mAP@0.5:0.95,and a 17.2%reduction in parameter count.These results highlight DAFPN-YOLO’s advantages in UAV-based object detection,offering valuable insights for applying deep learning to UAV-specific multi-object detection tasks.
基金supported by 2023IT020 of the Industry-University-Research Innovation Fund for Chinese Universities-New Generation Information Technology Innovation ProgramPX-972024121 of the Education&Teaching Reform Program of Guangdong Ocean University。
文摘Aiming at the problems of insufficient feature extraction ability for small targets,complex image background,and low detection accuracy in marine life detection,this paper proposes a marine life detection algorithm SGW-YOLOv8 based on the improvement of YOLOv8.First,the Adaptive Fine-Grained Channel Attention(FCA)module is fused with the backbone layer of the YOLOv8 network to improve the feature extraction ability of the model.This paper uses the YOLOv8 network backbone layer to improve the feature extraction capability of the model.Second,the Efficient Multi-Scale Attention(C2f_EMA)module is replaced with the C2f module in the Neck layer of the network to improve the detection performance of the model for small underwater targets.Finally,the loss function is optimized to Weighted Intersection over Union(WIoU)to replace the original loss function,so that the model is better adapted to the target detection task in the complex ocean background.The improved algorithm has been experimented with on the Underwater Robot Picking Contest(URPC)dataset,and the results show that the improved algorithm achieves a detection accuracy of 84.5,which is 2.3%higher than that before the improvement,and at the same time,it can accurately detect the small-target marine organisms and adapts to the task of detecting marine organisms in various complex environments.
基金the Shandong Provincial Key Research and Development Program(Grant No.2024SFGC0201).
文摘To address critical challenges in nighttime ship detection—high small-target missed detection(over 20%),insufficient lightweighting,and limited generalization due to scarce,low-quality datasets—this study proposes a systematic solution.First,a high-quality Night-Ships dataset is constructed via CycleGAN-based day-night transfer,combined with a dual-threshold cleaning strategy(Laplacian variance sharpness filtering and brightness-color deviation screening).Second,a Cross-stage Lightweight Fusion-You Only Look Once version 8(CLF-YOLOv8)is proposed with key improvements:the Neck network is reconstructed by replacing Cross Stage Partial(CSP)structure with the Cross Stage Partial Multi-Scale Convolutional Block(CSP-MSCB)and integrating Bidirectional Feature Pyramid Network(BiFPN)for weighted multi-scale fusion to enhance small-target detection;a Lightweight Shared Convolutional and Separated Batch Normalization Detection-Head(LSCSBD-Head)with shared convolutions and layer-wise Batch Normalization(BN)reduces parameters to 1.8M(42% fewer than YOLOv8n);and the FocalMinimum Point Distance Intersection over Union(Focal-MPDIoU)loss combines Minimum Point Distance Intersection over Union(MPDIoU)geometric constraints and Focal weighting to optimize low-overlap targets.Experiments show CLFYOLOv8 achieves 97.6%mAP@0.5(0.7% higher than YOLOv8n)with 1.8 M parameters,outperforming mainstream models in small-target detection,overlapping target discrimination,and adaptability to complex lighting.
基金funded by Key research and development Program of Henan Province(No.251111211200)National Natural Science Foundation of China(Grant No.U2004163).
文摘Traffic sign detection is an important part of autonomous driving,and its recognition accuracy and speed are directly related to road traffic safety.Although convolutional neural networks(CNNs)have made certain breakthroughs in this field,in the face of complex scenes,such as image blur and target occlusion,the traffic sign detection continues to exhibit limited accuracy,accompanied by false positives and missed detections.To address the above problems,a traffic sign detection algorithm,You Only Look Once-based Skip Dynamic Way(YOLO-SDW)based on You Only Look Once version 8 small(YOLOv8s),is proposed.Firstly,a Skip Connection Reconstruction(SCR)module is introduced to efficiently integrate fine-grained feature information and enhance the detection accuracy of the algorithm in complex scenes.Secondly,a C2f module based on Dynamic Snake Convolution(C2f-DySnake)is proposed to dynamically adjust the receptive field information,improve the algorithm’s feature extraction ability for blurred or occluded targets,and reduce the occurrence of false detections and missed detections.Finally,the Wise Powerful IoU v2(WPIoUv2)loss function is proposed to further improve the detection accuracy of the algorithm.Experimental results show that the average precision mAP@0.5 of YOLO-SDW on the TT100K dataset is 89.2%,and mAP@0.5:0.95 is 68.5%,which is 4%and 3.3%higher than the YOLOv8s baseline,respectively.YOLO-SDW ensures real-time performance while having higher accuracy.
基金funded by the Joint Funds of the National Natural Science Foundation of China(U2341223)the Beijing Municipal Natural Science Foundation(No.4232067).
文摘In printed circuit board(PCB)manufacturing,surface defects can significantly affect product quality.To address the performance degradation,high false detection rates,and missed detections caused by complex backgrounds in current intelligent inspection algorithms,this paper proposes CG-YOLOv8,a lightweight and improved model based on YOLOv8n for PCB surface defect detection.The proposed method optimizes the network architecture and compresses parameters to reduce model complexity while maintaining high detection accuracy,thereby enhancing the capability of identifying diverse defects under complex conditions.Specifically,a cascaded multi-receptive field(CMRF)module is adopted to replace the SPPF module in the backbone to improve feature perception,and an inverted residual mobile block(IRMB)is integrated into the C2f module to further enhance performance.Additionally,conventional convolution layers are replaced with GSConv to reduce computational cost,and a lightweight Convolutional Block Attention Module based Convolution(CBAMConv)module is introduced after Grouped Spatial Convolution(GSConv)to preserve accuracy through attention mechanisms.The detection head is also optimized by removing medium and large-scale detection layers,thereby enhancing the model’s ability to detect small-scale defects and further reducing complexity.Experimental results show that,compared to the original YOLOv8n,the proposed CG-YOLOv8 reduces parameter count by 53.9%,improves mAP@0.5 by 2.2%,and increases precision and recall by 2.0%and 1.8%,respectively.These improvements demonstrate that CG-YOLOv8 offers an efficient and lightweight solution for PCB surface defect detection.
文摘Object detection,a major challenge in computer vision and pattern recognition,plays a significant part in many applications,crossing artificial intelligence,face recognition,and autonomous driving.It involves focusing on identifying the detection,localization,and categorization of targets in images.A particularly important emerging task is distinguishing real animals from toy replicas in real-time,mostly for smart camera systems in both urban and natural environments.However,that difficult task is affected by factors such as showing angle,occlusion,light intensity,variations,and texture differences.To tackle these challenges,this paper recommends Group Sparse YOLOv8(You Only Look Once version 8),an improved real-time object detection algorithm that improves YOLOv8 by integrating group sparsity regularization.This adjustment improves efficiency and accuracy while utilizing the computational costs and power consumption,including a frame selection approach.And a hybrid parallel processing method that merges pipelining with dataflow strategies to improve the performance.Established using a custom dataset of toy and real animal images along with well-known datasets,namely ImageNet,MSCOCO,and CIFAR-10/100.The combination of Group Sparsity with YOLOv8 shows high detection accuracy with lower latency.Here provides a real and resource-efficient solution for intelligent camera systems and improves real-time object detection and classification in environments,differentiating between real and toy animals.
文摘With the gradual development of automatic driving technology,people’s attention is no longer limited to daily automatic driving target detection.In response to the problem that it is difficult to achieve fast and accurate detection of visual targets in complex scenes of automatic driving at night,a detection algorithm based on improved YOLOv8s was proposed.Firsly,By adding Triplet Attention module into the lower sampling layer of the original model,the model can effectively retain and enhance feature information related to target detection on the lower-resolution feature map.This enhancement improved the robustness of the target detection network and reduced instances of missed detections.Secondly,the Soft-NMS algorithm was introduced to address the challenges of dealing with dense targets,overlapping objects,and complex scenes.This algorithm effectively reduced false and missed positives,thereby improved overall detection performance when faced with highly overlapping detection results.Finally,the experimental results on the MPDIoU loss function dataset showed that compared with the original model,the improved method,in which mAP and accuracy are increased by 2.9%and 2.8%respectively,can achieve better detection accuracy and speed in night vehicle detection.It can effectively improve the problem of target detection in night scenes.
文摘In response to the challenge of low detection accuracy and susceptibility to missed and false detections of small targets in unmanned aerial vehicles(UAVs)aerial images,an improved UAV image target detection algorithm based on YOLOv8 was proposed in this study.To begin with,the CoordAtt attention mechanism was employed to enhance the feature extraction capability of the backbone network,thereby reducing interference from backgrounds.Additionally,the BiFPN feature fusion network with an added small object detection layer was used to enhance the model's ability to perceive for small objects.Furthermore,a multi-level fusion module was designed and proposed to effectively integrate shallow and deep information.The use of an enhanced MPDIoU loss function further improved detection performance.The experimental results based on the publicly available VisDrone2019 dataset showed that the improved model outperformed the YOLOv8 baseline model,mAP@0.5 improved by 20%,and the improved method improved the detection accuracy of the model for small targets.
文摘Pulmonary nodules represent an early manifestation of lung cancer.However,pulmonary nodules only constitute a small portion of the overall image,posing challenges for physicians in image interpretation and potentially leading to false positives or missed detections.To solve these problems,the YOLOv8 network is enhanced by adding deformable convolution and atrous spatial pyramid pooling(ASPP),along with the integration of a coordinate attention(CA)mechanism.This allows the network to focus on small targets while expanding the receptive field without losing resolution.At the same time,context information on the target is gathered and feature expression is enhanced by attention modules in different directions.It effectively improves the positioning accuracy and achieves good results on the LUNA16 dataset.Compared with other detection algorithms,it improves the accuracy of pulmonary nodule detection to a certain extent.
基金funded by Changzhou Science and Technology Project(No.CZ20230025)Postgraduate Research&Practice Innovation Program of Jiangsu Province(No.XSJCX23_36).
文摘Aiming at the problem that the existing algorithms for vehicle detection in smart factories are difficult to detect partial occlusion of vehicles,vulnerable to background interference,lack of global vision,and excessive suppression of real targets,which ultimately cause accuracy degradation.At the same time,to facilitate the subsequent positioning of vehicles in the factory,this paper proposes an improved YOLOv8 algorithm.Firstly,the RFCAConv module is combined to improve the original YOLOv8 backbone.Pay attention to the different features in the receptive field,and give priority to the spatial features of the receptive field to capture more vehicle feature information and solve the problem that the vehicle is partially occluded and difficult to detect.Secondly,the SFE module is added to the neck of v8,which improves the saliency of the target in the reasoning process and reduces the influence of background interference on vehicle detection.Finally,the head of the RT-DETR algorithm is used to replace the head in the original YOLOv8 algorithm,which avoids the excessive suppression of the real target while combining the context information.The experimental results show that compared with the original YOLOv8 algorithm,the detection accuracy of the improved YOLOv8 algorithm is improved by 4.6%on the self-made smart factory data set,and the detection speed also meets the real-time requirements of smart factory vehicle detection and subsequent vehicle positioning.
基金supported by Heilongjiang Provincial Natural Science Foundation of China(LH2023E055)the National Key R&D Program of China(2021YFB2600502).
文摘Traffic sign detection in real scenarios is challenging due to their complexity and small size,often preventing existing deep learning models from achieving both high accuracy and real-time performance.An improved YOLOv8 model for traffic sign detection is proposed.Firstly,by adding Coordinate Attention(CA)to the Backbone,the model gains location information,improving detection accuracy.Secondly,we also introduce EIoU to the localization function to address the ambiguity in aspect ratio descriptions by calculating the width-height difference based on CIoU.Additionally,Focal Loss is incorporated to balance sample difficulty,enhancing regression accuracy.Finally,the model,YOLOv8-CE(YOLOv8-Coordinate Attention-EIoU),is tested on the Jetson Nano,achieving real-time street scene detection and outperforming the Raspberry Pi 4B.Experimental results show that YOLOv8-CE excels in various complex scenarios,improving mAP by 2.8%over the original YOLOv8.The model size and computational effort remain similar,with the Jetson Nano achieving an inference time of 96 ms,significantly faster than the Raspberry Pi 4B.
文摘For the characteristics of small,dense distribution,high diversity of defects and high precision and fast detection in the process of PCB(Printed Circuit Board)defect detection,a defect detection algorithm based on YOLOv8n was proposed in this study.Firstly,the original C2f module of YOLOv8n was improved into a C2FFaster-EMA module to reduce the number of parameters and floating-point operations(FLOPs).Additionally,the WIoUv3 loss function was introduced to mitigate the negative impact of low-quality defect images on model training.Consequently,a reduction in model size and an enhancement in detection precision were achieved.Finally,the ablation and comparative experiments were conducted on an augmented Deep PCB dataset,and the generalization experiments were performed on the PCB Defect-Augmented dataset.The results indicated that the proposed model reduces the number of parameters by 23.3%and FLOPs by 20%,P by 0.7%,mAP@0.5 by 0.3%,and mAP@0.5:0.95 by 3.9%,respectively,compared to the original YOLOv8n model.Furthermore,the comparative experiments demonstrated that the proposed model achieves higher accuracy and mAP compared to YOLOv5n and YOLOv5s.It was concluded that the proposed method satisfies the requirements for both accuracy and speed in PCB defect detection.
文摘In recent years,with the development of synthetic aperture radar(SAR)technology and the widespread application of deep learning,lightweight detection of SAR images has emerged as a research direction.The ultimate goal is to reduce computational and storage requirements while ensuring detection accuracy and reliability,making it an ideal choice for achieving rapid response and efficient processing.In this regard,a lightweight SAR ship target detection algorithm based on YOLOv8 was proposed in this study.Firstly,the C2f-Sc module was designed by fusing the C2f in the backbone network with the ScConv to reduce spatial redundancy and channel redundancy between features in convolutional neural networks.At the same time,the Ghost module was introduced into the neck network to effectively reduce model parameters and computational complexity.A relatively lightweight EMA attention mechanism was added to the neck network to promote the effective fusion of features at different levels.Experimental results showed that the Parameters and GFLOPs of the improved model are reduced by 8.5%and 7.0%when mAP@0.5 and mAP@0.5:0.95 are increased by 0.7%and 1.8%,respectively.It makes the model lightweight and improves the detection accuracy,which has certain application value.
基金supported by the Open Research Fund Program of State Key Laboratory of Maritime Technology and Safety in 2024funding from the National Natural Science Foundation of China(Grant No.52331012)the Natural Science Foundation of Shanghai(Grant No.21ZR1426500).
文摘Synthetic Aperture Radar(SAR)has become one of the most effective tools in ship detection.However,due to significant background interference,small targets,and challenges related to target scattering intensity in SAR images,current ship target detection faces serious issues of missed detections and false positives,and the network structures are overly complex.To address this issue,this paper proposes a lightweight model based on YOLOv8,named OD-YOLOv8.Firstly,we adopt a simplified neural network architecture,VanillaNet,to replace the backbone network,significantly reducing the number of parameters and computational complexity while ensuring accuracy.Secondly,we introduce a dynamic,multi-dimensional attention mechanism by designing the ODC2f module with ODConv to replace the original C2f module and using GSConv to replace two down-sampling convolutions to reduce the number of parameters.Then,to alleviate the issues of missed detections and false positives for small targets,we discard one of the original large target detection layers and add a detection layer specifically for small targets.Finally,based on a dynamic non-monotonic focusing mechanism,we employ the Wise-IoU(Intersection over Union)loss function to significantly improve detection accuracy.Experimental results on the HRSID dataset show that,compared to the original YOLOv8,OD-YOLOv8 improves mAP@0.5 and mAP@0.5–0.95 by 2.7%and 3.5%,respectively,while reducing the number of parameters and GFLOPs by 72.9%and 4.9%,respectively.Moreover,the model also performs exceptionally well on the SSDD dataset,with AP and AP50 increasing by 1.7%and 0.4%,respectively.OD-YOLOv8 achieves an excellent balance between model lightweightness and accuracy,making it highly valuable for end-to-end industrial deployment.
基金National Natural Science Foundation of China(61703363,62272284)Shanxi Provincial Basic Research Program(201801D121148)Yuncheng University Research and Innovation Team for Data Mining and Industrial Intelligence Applications(YCXYTD-202402)。
文摘Addressing challenges in accurately detecting persimmon fruit quality in orchards—such as reliance on manual grading,low efficiency,severe foliage obstruction,and subtle differences between quality grades—this paper proposes a lightweight persimmon detection model based on an improved YOLOv8s architecture.First,the Conv layer in the backbone network is replaced with an ADown module to reduce model complexity.Second,MSFAN is introduced in the Neck layer to fully extract texture features from persimmon images,highlighting differences between quality grades.Finally,the Wise-IoU loss function mitigates the impact of low-quality sample data on grading accuracy.The improved model accurately identifies and separates persimmons of varying quality,effectively addressing quality grading detection in complex backgrounds.This provides a viable technical approach for achieving persimmon quality grading detection.
文摘This study systematically addresses the limitations of traditional pest detection methods and proposes an optimized version of the YOLOv8 object detection model.By integrating the GhostConv convolution module and the C3Ghost module,the Polarized Self-Attention(PSA)mechanism is incorporated to enhance the model’s capacity for extracting pest features.Experimental results demonstrate that the improved YOLOv8+Ghost+PSA model achieves outstanding performance in critical metrics such as precision,recall,and mean Average Precision(mAP),with a computational cost of only 5.3 GFLOPs,making it highly suitable for deployment in resource-constrained agricultural environments.
基金supported by the Jiangsu Province Science and Technology Policy Guidance Program(Industry-University-Research Cooperation)/Forward-Looking Joint Research Project(BY2016005-05).
文摘In response to themissed and false detections that are easily caused by the large variety of and significant differences among aluminum surface defects,a detection algorithm based on an improved You Only Look Once(YOLO)v8n network is proposed.First,a C2f_DWR_DRB module is constructed by introducing a dilation-wise residual(DWR)module and a dilated reparameterization block(DRB)to replace the C2f module at the high level of the backbone network,enriching the gradient flow information and increasing the effective receptive field(ERF).Second,an efficient local attention(ELA)mechanism is fused with the high-level screening-feature pyramid networks(HS-FPN)module,and an ELA_HSFPN is designed to replace the original feature fusion module,enhancing the ability of the network to cope with multiscale detection tasks.Moreover,a lightweight shared convolutional detection head(SCDH)is introduced to reduce the number of parameters and the computational complexity of the module while enhancing the performance and generalizability of the model.Finally,the soft intersection over union(SIoU)replaces the original loss function to improve the convergence speed and prediction accuracy of the model.Experimental results show that compared with that of the original YOLOv8n model,the mAP@0.5 of the improved algorithm is increased by 5.1%,the number of parameters and computational complexity are reduced by 33.3%and 32.1%,respectively,and the FPS is increased by 4.9%.Compared with other mainstream object detection algorithms,the improved algorithm still leads in terms of core indicators and has good generalizability for surface defects encountered in other industrial scenarios.
基金supported by the Scientific and technological key project in Henan Province 22210224002the Natural Science Foundation of Henan Polytechnic University B2021-38.
文摘The service life of internal combustion engines is significantly influenced by surface defects in cylinder liners.To address the limitations of traditional detection methods,we propose an enhanced YOLOv8 model with Swin Transformer as the backbone network.This approach leverages Swin Transformer's multi-head self-attention mechanism for improved feature extraction of defects spanning various scales.Integrated with the YOLOv8 detection head,our model achieves a mean average precision of 85.1%on our dataset,outperforming baseline methods by 1.4%.The model's effectiveness is further demonstrated on a steel-surface defect dataset,indicating its broad applicability in industrial surface defect detection.Our work highlights the potential of combining Swin Transformer and YOLOv8 for accurate and efficient defect detection.
基金supported in part by the National Natural Foundation of China(Nos.52472334,U2368204)。
文摘In response to challenges posed by complex backgrounds,diverse target angles,and numerous small targets in remote sensing images,alongside the issue of high resource consumption hindering model deployment,we propose an enhanced,lightweight you only look once version 8 small(YOLOv8s)detection algorithm.Regarding network improvements,we first replace tradi-tional horizontal boxes with rotated boxes for target detection,effectively addressing difficulties in feature extraction caused by varying target angles.Second,we design a module integrating convolu-tional neural networks(CNN)and Transformer components to replace specific C2f modules in the backbone network,thereby expanding the model’s receptive field and enhancing feature extraction in complex backgrounds.Finally,we introduce a feature calibration structure to mitigate potential feature mismatches during feature fusion.For model compression,we employ a lightweight channel pruning technique based on localized mean average precision(LMAP)to eliminate redundancies in the enhanced model.Although this approach results in some loss of detection accuracy,it effec-tively reduces the number of parameters,computational load,and model size.Additionally,we employ channel-level knowledge distillation to recover accuracy in the pruned model,further enhancing detection performance.Experimental results indicate that the enhanced algorithm achieves a 6.1%increase in mAP50 compared to YOLOv8s,while simultaneously reducing parame-ters,computational load,and model size by 57.7%,28.8%,and 52.3%,respectively.