Reliable detection of traffic signs and lights(TSLs)at long range and under varying illumination is essen-tial for improving the perception and safety of autonomous driving systems(ADS).Traditional object detection mo...Reliable detection of traffic signs and lights(TSLs)at long range and under varying illumination is essen-tial for improving the perception and safety of autonomous driving systems(ADS).Traditional object detection models often exhibit significant performance degradation in real-world environments characterized by high dynamic range and complex lighting conditions.To overcome these limitations,this research presents FED-YOLOv10s,an improved and lightweight object detection framework based on You Only look Once v10(YOLOv10).The proposed model integrates a C2f-Faster block derived from FasterNet to reduce parameters and floating-point operations,an Efficient Multiscale Attention(EMA)mechanism to improve TSL-invariant feature extraction,and a deformable Convolution Networks v4(DCNv4)module to enhance multiscale spatial adaptability.Experimental findings demonstrate that the proposed architecture achieves an optimal balance between computational efficiency and detection accuracy,attaining an F1-score of 91.8%,and mAP@0.5 of 95.1%,while reducing parameters to 8.13 million.Comparative analyses across multiple traffic sign detection benchmarks demonstrate that FED-YOLOv10s outperforms state-of-the-art models in precision,recall,and mAP.These results highlight FED-YOLOv10s as a robust,efficient,and deployable solution for intelligent traffic perception in ADS.展开更多
In recent years,the country has spent significant workforce and material resources to prevent traffic accidents,particularly those caused by fatigued driving.The current studies mainly concentrate on driver physiologi...In recent years,the country has spent significant workforce and material resources to prevent traffic accidents,particularly those caused by fatigued driving.The current studies mainly concentrate on driver physiological signals,driving behavior,and vehicle information.However,most of the approaches are computationally intensive and inconvenient for real-time detection.Therefore,this paper designs a network that combines precision,speed and lightweight and proposes an algorithm for facial fatigue detection based on multi-feature fusion.Specifically,the face detection model takes YOLOv8(You Only Look Once version 8)as the basic framework,and replaces its backbone network with MobileNetv3.To focus on the significant regions in the image,CPCA(Channel Prior Convolution Attention)is adopted to enhance the network’s capacity for feature extraction.Meanwhile,the network training phase employs the Focal-EIOU(Focal and Efficient Intersection Over Union)loss function,which makes the network lightweight and increases the accuracy of target detection.Ultimately,the Dlib toolkit was employed to annotate 68 facial feature points.This study established an evaluation metric for facial fatigue and developed a novel fatigue detection algorithm to assess the driver’s condition.A series of comparative experiments were carried out on the self-built dataset.The suggested method’s mAP(mean Average Precision)values for object detection and fatigue detection are 96.71%and 95.75%,respectively,as well as the detection speed is 47 FPS(Frames Per Second).This method can balance the contradiction between computational complexity and model accuracy.Furthermore,it can be transplanted to NVIDIA Jetson Orin NX and quickly detect the driver’s state while maintaining a high degree of accuracy.It contributes to the development of automobile safety systems and reduces the occurrence of traffic accidents.展开更多
基金funded by the Deanship of Scientific Research(DSR)at King Abdulaziz University,Jeddah,Saudi Arabia under Grant No.IPP:172-830-2025.
文摘Reliable detection of traffic signs and lights(TSLs)at long range and under varying illumination is essen-tial for improving the perception and safety of autonomous driving systems(ADS).Traditional object detection models often exhibit significant performance degradation in real-world environments characterized by high dynamic range and complex lighting conditions.To overcome these limitations,this research presents FED-YOLOv10s,an improved and lightweight object detection framework based on You Only look Once v10(YOLOv10).The proposed model integrates a C2f-Faster block derived from FasterNet to reduce parameters and floating-point operations,an Efficient Multiscale Attention(EMA)mechanism to improve TSL-invariant feature extraction,and a deformable Convolution Networks v4(DCNv4)module to enhance multiscale spatial adaptability.Experimental findings demonstrate that the proposed architecture achieves an optimal balance between computational efficiency and detection accuracy,attaining an F1-score of 91.8%,and mAP@0.5 of 95.1%,while reducing parameters to 8.13 million.Comparative analyses across multiple traffic sign detection benchmarks demonstrate that FED-YOLOv10s outperforms state-of-the-art models in precision,recall,and mAP.These results highlight FED-YOLOv10s as a robust,efficient,and deployable solution for intelligent traffic perception in ADS.
基金supported by the Science and Technology Bureau of Xi’an project(24KGDW0049)the Key Research and Development Programof Shaanxi(2023-YBGY-264)the Key Research and Development Program of Guangxi(GK-AB20159032).
文摘In recent years,the country has spent significant workforce and material resources to prevent traffic accidents,particularly those caused by fatigued driving.The current studies mainly concentrate on driver physiological signals,driving behavior,and vehicle information.However,most of the approaches are computationally intensive and inconvenient for real-time detection.Therefore,this paper designs a network that combines precision,speed and lightweight and proposes an algorithm for facial fatigue detection based on multi-feature fusion.Specifically,the face detection model takes YOLOv8(You Only Look Once version 8)as the basic framework,and replaces its backbone network with MobileNetv3.To focus on the significant regions in the image,CPCA(Channel Prior Convolution Attention)is adopted to enhance the network’s capacity for feature extraction.Meanwhile,the network training phase employs the Focal-EIOU(Focal and Efficient Intersection Over Union)loss function,which makes the network lightweight and increases the accuracy of target detection.Ultimately,the Dlib toolkit was employed to annotate 68 facial feature points.This study established an evaluation metric for facial fatigue and developed a novel fatigue detection algorithm to assess the driver’s condition.A series of comparative experiments were carried out on the self-built dataset.The suggested method’s mAP(mean Average Precision)values for object detection and fatigue detection are 96.71%and 95.75%,respectively,as well as the detection speed is 47 FPS(Frames Per Second).This method can balance the contradiction between computational complexity and model accuracy.Furthermore,it can be transplanted to NVIDIA Jetson Orin NX and quickly detect the driver’s state while maintaining a high degree of accuracy.It contributes to the development of automobile safety systems and reduces the occurrence of traffic accidents.