期刊文献+
共找到6篇文章
< 1 >
每页显示 20 50 100
Aerial Images for Intelligent Vehicle Detection and Classification via YOLOv11 and Deep Learner
1
作者 Ghulam Mujtaba Wenbiao Liu +3 位作者 Mohammed Alshehri Yahya AlQahtani Nouf Abdullah Almujally Hui Liu 《Computers, Materials & Continua》 2026年第1期1703-1721,共19页
As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a no... As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a novel,unified deep learning framework for vehicle detection,tracking,counting,and classification in aerial imagery designed explicitly for modern smart city infrastructure demands.Our approach begins with adaptive histogram equalization to optimize aerial image clarity,followed by a cutting-edge scene parsing technique using Mask2Former,enabling robust segmentation even in visually congested settings.Vehicle detection leverages the latest YOLOv11 architecture,delivering superior accuracy in aerial contexts by addressing occlusion,scale variance,and fine-grained object differentiation.We incorporate the highly efficient ByteTrack algorithm for tracking,enabling seamless identity preservation across frames.Vehicle counting is achieved through an unsupervised DBSCAN-based method,ensuring adaptability to varying traffic densities.We further introduce a hybrid feature extraction module combining Convolutional Neural Networks(CNNs)with Zernike Moments,capturing both deep semantic and geometric signatures of vehicles.The final classification is powered by NASNet,a neural architecture search-optimized model,ensuring high accuracy across diverse vehicle types and orientations.Extensive evaluations of the VAID benchmark dataset demonstrate the system’s outstanding performance,achieving 96%detection,94%tracking,and 96.4%classification accuracy.On the UAVDT dataset,the system attains 95%detection,93%tracking,and 95%classification accuracy,confirming its robustness across diverse aerial traffic scenarios.These results establish new benchmarks in aerial traffic analysis and validate the framework’s scalability,making it a powerful and adaptable solution for next-generation intelligent transportation systems and urban surveillance. 展开更多
关键词 Traffic management YOLOv11 autonomous vehicles intelligent traffic systems NASNet zernike moments
在线阅读 下载PDF
Traffic Vision:UAV-Based Vehicle Detection and Traffic Pattern Analysis via Deep Learning Classifier
2
作者 Mohammed Alnusayri Ghulam Mujtaba +4 位作者 Nouf Abdullah Almujally Shuoa S.Aitarbi Asaad Algarni Ahmad Jalal Jeongmin Park 《Computers, Materials & Continua》 2026年第3期266-284,共19页
This paper presents a unified Unmanned Aerial Vehicle-based(UAV-based)traffic monitoring framework that integrates vehicle detection,tracking,counting,motion prediction,and classification in a modular and co-optimized... This paper presents a unified Unmanned Aerial Vehicle-based(UAV-based)traffic monitoring framework that integrates vehicle detection,tracking,counting,motion prediction,and classification in a modular and co-optimized pipeline.Unlike prior works that address these tasks in isolation,our approach combines You Only Look Once(YOLO)v10 detection,ByteTrack tracking,optical-flow density estimation,Long Short-Term Memory-based(LSTM-based)trajectory forecasting,and hybrid Speeded-Up Robust Feature(SURF)+Gray-Level Co-occurrence Matrix(GLCM)feature engineering with VGG16 classification.Upon the validation across datasets(UAVDT and UAVID)our framework achieved a detection accuracy of 94.2%,and 92.3%detection accuracy when conducting a real-time UAV field validation.Our comprehensive evaluations,including multi-metric analyses,ablation studies,and cross-dataset validations,confirm the framework’s accuracy,efficiency,and generalizability.These results highlight the novelty of integrating complementary methods into a single framework,offering a practical solution for accurate and efficient UAV-based traffic monitoring. 展开更多
关键词 Smart traffic system drone devices machine learner dynamic complex scenes VGG-16 classifier
在线阅读 下载PDF
Intelligent Human Interaction Recognition with Multi-Modal Feature Extraction and Bidirectional LSTM
3
作者 Muhammad Hamdan Azhar Yanfeng Wu +4 位作者 Nouf Abdullah Almujally Shuaa S.Alharbi Asaad Algarni Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2026年第4期1632-1649,共18页
Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationall... Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationally intensive,sensitive to video resolution changes and often fail in crowded scenes.We propose a novel hybrid system that is computationally efficient,robust to degraded video quality and able to filter out irrelevant individuals,making it suitable for real-life use.The system leverages multi-modal handcrafted features for interaction representation and a deep learning classifier for capturing complex dependencies.Using Mask R-CNN and YOLO11-Pose,we extract grayscale silhouettes and keypoint coordinates of interacting individuals,while filtering out irrelevant individuals using a proposed algorithm.From these,we extract silhouette-based features(local ternary pattern and histogram of optical flow)and keypoint-based features(distances,angles and velocities)that capture distinct spatial and temporal information.A Bidirectional Long Short-Term Memory network(BiLSTM)then classifies the interactions.Extensive experiments on the UT Interaction,SBU Kinect Interaction and the ISR-UOL 3D social activity datasets demonstrate that our system achieves competitive accuracy.They also validate the effectiveness of the chosen features and classifier,along with the proposed system’s computational efficiency and robustness to occlusion. 展开更多
关键词 Human interaction recognition keypoint coordinates grayscale silhouettes bidirectional long shortterm memory network
在线阅读 下载PDF
Drone-Based Public Surveillance Using 3D Point Clouds and Neuro-Fuzzy Classifier
4
作者 Yawar Abbas Aisha Ahmed Alarfaj +3 位作者 Ebtisam Abdullah Alabdulqader Asaad Algarni Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2025年第3期4759-4776,共18页
Human Activity Recognition(HAR)in drone-captured videos has become popular because of the interest in various fields such as video surveillance,sports analysis,and human-robot interaction.However,recognizing actions f... Human Activity Recognition(HAR)in drone-captured videos has become popular because of the interest in various fields such as video surveillance,sports analysis,and human-robot interaction.However,recognizing actions from such videos poses the following challenges:variations of human motion,the complexity of backdrops,motion blurs,occlusions,and restricted camera angles.This research presents a human activity recognition system to address these challenges by working with drones’red-green-blue(RGB)videos.The first step in the proposed system involves partitioning videos into frames and then using bilateral filtering to improve the quality of object foregrounds while reducing background interference before converting from RGB to grayscale images.The YOLO(You Only Look Once)algorithm detects and extracts humans from each frame,obtaining their skeletons for further processing.The joint angles,displacement and velocity,histogram of oriented gradients(HOG),3D points,and geodesic Distance are included.These features are optimized using Quadratic Discriminant Analysis(QDA)and utilized in a Neuro-Fuzzy Classifier(NFC)for activity classification.Real-world evaluations on the Drone-Action,Unmanned Aerial Vehicle(UAV)-Gesture,and Okutama-Action datasets substantiate the proposed system’s superiority in accuracy rates over existing methods.In particular,the system obtains recognition rates of 93%for drone action,97%for UAV gestures,and 81%for Okutama-action,demonstrating the system’s reliability and ability to learn human activity from drone videos. 展开更多
关键词 Activity recognition geodesic distance pattern recognition neuro fuzzy classifier
在线阅读 下载PDF
Semantic Segmentation and YOLO Detector over Aerial Vehicle Images
5
作者 Asifa Mehmood Qureshi Abdul Haleem Butt +5 位作者 Abdulwahab Alazeb Naif Al Mudawi Mohammad Alonazi Nouf Abdullah Almujally Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 SCIE EI 2024年第8期3315-3332,共18页
Intelligent vehicle tracking and detection are crucial tasks in the realm of highway management.However,vehicles come in a range of sizes,which is challenging to detect,affecting the traffic monitoring system’s overa... Intelligent vehicle tracking and detection are crucial tasks in the realm of highway management.However,vehicles come in a range of sizes,which is challenging to detect,affecting the traffic monitoring system’s overall accuracy.Deep learning is considered to be an efficient method for object detection in vision-based systems.In this paper,we proposed a vision-based vehicle detection and tracking system based on a You Look Only Once version 5(YOLOv5)detector combined with a segmentation technique.The model consists of six steps.In the first step,all the extracted traffic sequence images are subjected to pre-processing to remove noise and enhance the contrast level of the images.These pre-processed images are segmented by labelling each pixel to extract the uniform regions to aid the detection phase.A single-stage detector YOLOv5 is used to detect and locate vehicles in images.Each detection was exposed to Speeded Up Robust Feature(SURF)feature extraction to track multiple vehicles.Based on this,a unique number is assigned to each vehicle to easily locate them in the succeeding image frames by extracting them using the feature-matching technique.Further,we implemented a Kalman filter to track multiple vehicles.In the end,the vehicle path is estimated by using the centroid points of the rectangular bounding box predicted by the tracking algorithm.The experimental results and comparison reveal that our proposed vehicle detection and tracking system outperformed other state-of-the-art systems.The proposed implemented system provided 94.1%detection precision for Roundabout and 96.1%detection precision for Vehicle Aerial Imaging from Drone(VAID)datasets,respectively. 展开更多
关键词 Semantic segmentation YOLOv5 vehicle detection and tracking Kalman filter SURF
在线阅读 下载PDF
A Hybrid Deep Learning Pipeline for Wearable Sensors-Based Human Activity Recognition
6
作者 Asaad Algarni Iqra Aijaz Abro +3 位作者 Mohammed Alshehri Yahya AlQahtani Abdulmonem Alshahrani Hui Liu 《Computers, Materials & Continua》 2025年第9期5879-5896,共18页
Inertial Sensor-based Daily Activity Recognition(IS-DAR)requires adaptable,data-efficient methods for effective multi-sensor use.This study presents an advanced detection system using body-worn sensors to accurately r... Inertial Sensor-based Daily Activity Recognition(IS-DAR)requires adaptable,data-efficient methods for effective multi-sensor use.This study presents an advanced detection system using body-worn sensors to accurately recognize activities.A structured pipeline enhances IS-DAR by applying signal preprocessing,feature extraction and optimization,followed by classification.Before segmentation,a Chebyshev filter removes noise,and Blackman window-ing improves signal representation.Discriminative features-Gaussian Mixture Model(GMM)with Mel-Frequency Cepstral Coefficients(MFCC),spectral entropy,quaternion-based features,and Gammatone Cepstral Coefficients(GCC)-are fused to expand the feature space.Unlike existing approaches,the proposed IS-DAR system uniquely inte-grates diverse handcrafted features using a novel fusion strategy combined with Bayesian-based optimization,enabling a more accurate and generalized activity recognition.The key contribution lies in the joint optimization and fusion of features via Bayesian-based subset selection,resulting in a compact and highly discriminative feature representation.These features are then fed into a Convolutional Neural Network(CNN)to effectively detect spatial-temporal patterns in activity signals.Testing on two public datasets-IM-WSHA and ENABL3S-achieved accuracy levels of 93.0%and 92.0%,respectively.The integration of advanced feature extraction methods with fusion and optimization techniques significantly enhanced detection performance,surpassing traditional methods.The obtained results establish the effectiveness of the proposed IS-DAR system for deployment in real-world activity recognition applications. 展开更多
关键词 Wearable sensors deep learning pattern recognition feature extraction
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部