As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a no...As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a novel,unified deep learning framework for vehicle detection,tracking,counting,and classification in aerial imagery designed explicitly for modern smart city infrastructure demands.Our approach begins with adaptive histogram equalization to optimize aerial image clarity,followed by a cutting-edge scene parsing technique using Mask2Former,enabling robust segmentation even in visually congested settings.Vehicle detection leverages the latest YOLOv11 architecture,delivering superior accuracy in aerial contexts by addressing occlusion,scale variance,and fine-grained object differentiation.We incorporate the highly efficient ByteTrack algorithm for tracking,enabling seamless identity preservation across frames.Vehicle counting is achieved through an unsupervised DBSCAN-based method,ensuring adaptability to varying traffic densities.We further introduce a hybrid feature extraction module combining Convolutional Neural Networks(CNNs)with Zernike Moments,capturing both deep semantic and geometric signatures of vehicles.The final classification is powered by NASNet,a neural architecture search-optimized model,ensuring high accuracy across diverse vehicle types and orientations.Extensive evaluations of the VAID benchmark dataset demonstrate the system’s outstanding performance,achieving 96%detection,94%tracking,and 96.4%classification accuracy.On the UAVDT dataset,the system attains 95%detection,93%tracking,and 95%classification accuracy,confirming its robustness across diverse aerial traffic scenarios.These results establish new benchmarks in aerial traffic analysis and validate the framework’s scalability,making it a powerful and adaptable solution for next-generation intelligent transportation systems and urban surveillance.展开更多
In this study,an automated multimodal system for detecting,classifying,and dating fruit was developed using a two-stage YOLOv11 pipeline.In the first stage,the YOLOv11 detection model locates individual date fruits in...In this study,an automated multimodal system for detecting,classifying,and dating fruit was developed using a two-stage YOLOv11 pipeline.In the first stage,the YOLOv11 detection model locates individual date fruits in real time by drawing bounding boxes around them.These bounding boxes are subsequently passed to a YOLOv11 classification model,which analyzes cropped images and assigns class labels.An additional counting module automatically tallies the detected fruits,offering a near-instantaneous estimation of quantity.The experimental results suggest high precision and recall for detection,high classification accuracy(across 15 classes),and near-perfect counting in real time.This paper presents a multi-stage pipeline for date fruit detection,classification,and automated counting,employing YOLOv11-based models to achieve high accuracy while maintaining real-time throughput.The results demonstrated that the detection precision exceeded 90%,the classification accuracy approached 92%,and the counting module correlated closely with the manual tallies.These findings confirm the potential of reducing manual labour and enhancing operational efficiency in post-harvesting processes.Future studies will include dataset expansion,user-centric interfaces,and integration with harvesting robotics.展开更多
基金funded by the Open Access Initiative of the University of Bremen and the DFG via SuUB BremenThe authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Group Project under grant number(RGP2/367/46)+1 种基金This research is supported and funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R410)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a novel,unified deep learning framework for vehicle detection,tracking,counting,and classification in aerial imagery designed explicitly for modern smart city infrastructure demands.Our approach begins with adaptive histogram equalization to optimize aerial image clarity,followed by a cutting-edge scene parsing technique using Mask2Former,enabling robust segmentation even in visually congested settings.Vehicle detection leverages the latest YOLOv11 architecture,delivering superior accuracy in aerial contexts by addressing occlusion,scale variance,and fine-grained object differentiation.We incorporate the highly efficient ByteTrack algorithm for tracking,enabling seamless identity preservation across frames.Vehicle counting is achieved through an unsupervised DBSCAN-based method,ensuring adaptability to varying traffic densities.We further introduce a hybrid feature extraction module combining Convolutional Neural Networks(CNNs)with Zernike Moments,capturing both deep semantic and geometric signatures of vehicles.The final classification is powered by NASNet,a neural architecture search-optimized model,ensuring high accuracy across diverse vehicle types and orientations.Extensive evaluations of the VAID benchmark dataset demonstrate the system’s outstanding performance,achieving 96%detection,94%tracking,and 96.4%classification accuracy.On the UAVDT dataset,the system attains 95%detection,93%tracking,and 95%classification accuracy,confirming its robustness across diverse aerial traffic scenarios.These results establish new benchmarks in aerial traffic analysis and validate the framework’s scalability,making it a powerful and adaptable solution for next-generation intelligent transportation systems and urban surveillance.
基金supported by the Deanship of Scientific Research,Vice Presidency for Graduate Studies and Scientific Research,King Faisal University,Saudi Arabia,Grant No.KFU250098.
文摘In this study,an automated multimodal system for detecting,classifying,and dating fruit was developed using a two-stage YOLOv11 pipeline.In the first stage,the YOLOv11 detection model locates individual date fruits in real time by drawing bounding boxes around them.These bounding boxes are subsequently passed to a YOLOv11 classification model,which analyzes cropped images and assigns class labels.An additional counting module automatically tallies the detected fruits,offering a near-instantaneous estimation of quantity.The experimental results suggest high precision and recall for detection,high classification accuracy(across 15 classes),and near-perfect counting in real time.This paper presents a multi-stage pipeline for date fruit detection,classification,and automated counting,employing YOLOv11-based models to achieve high accuracy while maintaining real-time throughput.The results demonstrated that the detection precision exceeded 90%,the classification accuracy approached 92%,and the counting module correlated closely with the manual tallies.These findings confirm the potential of reducing manual labour and enhancing operational efficiency in post-harvesting processes.Future studies will include dataset expansion,user-centric interfaces,and integration with harvesting robotics.