The effective and timely diagnosis and treatment of ocular diseases are key to the rapid recovery of patients.Today,the mass disease that needs attention in this context is cataracts.Although deep learning has signifi...The effective and timely diagnosis and treatment of ocular diseases are key to the rapid recovery of patients.Today,the mass disease that needs attention in this context is cataracts.Although deep learning has significantly advanced the analysis of ocular disease images,there is a need for a probabilistic model to generate the distributions of potential outcomes and thusmake decisions related to uncertainty quantification.Therefore,this study implements a Bayesian Convolutional Neural Networks(BCNN)model for predicting cataracts by assigning probability values to the predictions.It prepares convolutional neural network(CNN)and BCNN models.The proposed BCNN model is CNN-based in which reparameterization is in the first and last layers of the CNN model.This study then trains them on a dataset of cataract images filtered from the ocular disease fundus images fromKaggle.The deep CNN model has an accuracy of 95%,while the BCNN model has an accuracy of 93.75% along with information on uncertainty estimation of cataracts and normal eye conditions.When compared with other methods,the proposed work reveals that it can be a promising solution for cataract prediction with uncertainty estimation.展开更多
As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a no...As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a novel,unified deep learning framework for vehicle detection,tracking,counting,and classification in aerial imagery designed explicitly for modern smart city infrastructure demands.Our approach begins with adaptive histogram equalization to optimize aerial image clarity,followed by a cutting-edge scene parsing technique using Mask2Former,enabling robust segmentation even in visually congested settings.Vehicle detection leverages the latest YOLOv11 architecture,delivering superior accuracy in aerial contexts by addressing occlusion,scale variance,and fine-grained object differentiation.We incorporate the highly efficient ByteTrack algorithm for tracking,enabling seamless identity preservation across frames.Vehicle counting is achieved through an unsupervised DBSCAN-based method,ensuring adaptability to varying traffic densities.We further introduce a hybrid feature extraction module combining Convolutional Neural Networks(CNNs)with Zernike Moments,capturing both deep semantic and geometric signatures of vehicles.The final classification is powered by NASNet,a neural architecture search-optimized model,ensuring high accuracy across diverse vehicle types and orientations.Extensive evaluations of the VAID benchmark dataset demonstrate the system’s outstanding performance,achieving 96%detection,94%tracking,and 96.4%classification accuracy.On the UAVDT dataset,the system attains 95%detection,93%tracking,and 95%classification accuracy,confirming its robustness across diverse aerial traffic scenarios.These results establish new benchmarks in aerial traffic analysis and validate the framework’s scalability,making it a powerful and adaptable solution for next-generation intelligent transportation systems and urban surveillance.展开更多
Remote sensing plays a pivotal role in environmental monitoring,disaster relief,and urban planning,where accurate scene classification of aerial images is essential.However,conventional convolutional neural networks(C...Remote sensing plays a pivotal role in environmental monitoring,disaster relief,and urban planning,where accurate scene classification of aerial images is essential.However,conventional convolutional neural networks(CNNs)struggle with long-range dependencies and preserving high-resolution features,limiting their effectiveness in complex aerial image analysis.To address these challenges,we propose a Hybrid HRNet-Swin Transformer model that synergizes the strengths of HRNet-W48 for high-resolution segmentation and the Swin Transformer for global feature extraction.This hybrid architecture ensures robust multi-scale feature fusion,capturing fine-grained details and broader contextual relationships in aerial imagery.Our methodology begins with preprocessing steps,including normalization,histogram equalization,and noise reduction,to enhance input data quality.The HRNet-W48 backbone maintains high-resolution feature maps throughout the network,enabling precise segmentation,while the Swin Transformer leverages hierarchical self-attention to model long-range dependencies efficiently.By integrating these components,our model achieves superior performance in segmentation and classification tasks compared to traditional CNNs and standalone transformer models.We evaluate our approach on two benchmark datasets:UC Merced and WHU-RS19.Experimental results demonstrate that the proposed hybrid model outperforms existing methods,achieving state-of-the-art accuracy while maintaining computational efficiency.Specifically,it excels in preserving fine spatial details and contextual understanding,critical for applications like land-use classification and disaster assessment.展开更多
The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS ...The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS models face several challenges,including training instability,high computational costs,and system failures.To address these limitations,we propose a Hybrid Wasserstein GAN and Autoencoder Model(WGAN-AE)for intrusion detection.The proposed framework leverages the stability of WGAN and the feature extraction capabilities of the Autoencoder Model.The model was trained and evaluated using two recent benchmark datasets,5GNIDD and IDSIoT2024.When trained on the 5GNIDD dataset,the model achieved an average area under the precisionrecall curve is 99.8%using five-fold cross-validation and demonstrated a high detection accuracy of 97.35%when tested on independent test data.Additionally,the model is well-suited for deployment on resource-limited Internetof-Things(IoT)devices due to its ability to detect attacks within microseconds and its small memory footprint of 60.24 kB.Similarly,when trained on the IDSIoT2024 dataset,the model achieved an average PR-AUC of 94.09%and an attack detection accuracy of 97.35%on independent test data,with a memory requirement of 61.84 kB.Extensive simulation results demonstrate that the proposed hybrid model effectively addresses the shortcomings of traditional GAN-based IDS approaches in terms of detection accuracy,computational efficiency,and applicability to real-world IoT environments.展开更多
Intelligent Traffic Management(ITM)has progressively developed into a critical component of modern transportation networks,significantly enhancing traffic flow and reducing congestion in urban environments.This resear...Intelligent Traffic Management(ITM)has progressively developed into a critical component of modern transportation networks,significantly enhancing traffic flow and reducing congestion in urban environments.This research proposes an enhanced framework that leverages Deep Q-Learning(DQL),Game Theory(GT),and Stochastic Optimization(SO)to tackle the complex dynamics in transportation networks.The DQL component utilizes the distribution of traffic conditions for epsilon-greedy policy formulation and action and choice reward calculation,ensuring resilient decision-making.GT models the interaction between vehicles and intersections through probabilistic distributions of various features to enhance performance.Results demonstrate that the proposed framework is a scalable solution for dynamic optimization in transportation networks.展开更多
Edge computing(EC)combined with the Internet of Things(IoT)provides a scalable and efficient solution for smart homes.Therapid proliferation of IoT devices poses real-time data processing and security challenges.EC ha...Edge computing(EC)combined with the Internet of Things(IoT)provides a scalable and efficient solution for smart homes.Therapid proliferation of IoT devices poses real-time data processing and security challenges.EC has become a transformative paradigm for addressing these challenges,particularly in intrusion detection and anomaly mitigation.The widespread connectivity of IoT edge networks has exposed them to various security threats,necessitating robust strategies to detect malicious activities.This research presents a privacy-preserving federated anomaly detection framework combined with Bayesian game theory(BGT)and double deep Q-learning(DDQL).The proposed framework integrates BGT to model attacker and defender interactions for dynamic threat level adaptation and resource availability.It also models a strategic layout between attackers and defenders that takes into account uncertainty.DDQL is incorporated to optimize decision-making and aids in learning optimal defense policies at the edge,thereby ensuring policy and decision optimization.Federated learning(FL)enables decentralized and unshared anomaly detection for sensitive data between devices.Data collection has been performed from various sensors in a real-time EC-IoT network to identify irregularities that occurred due to different attacks.The results reveal that the proposed model achieves high detection accuracy of up to 98%while maintaining low resource consumption.This study demonstrates the synergy between game theory and FL to strengthen anomaly detection in EC-IoT networks.展开更多
Content-Based Image Retrieval(CBIR)and image mining are becoming more important study fields in computer vision due to their wide range of applications in healthcare,security,and various domains.The image retrieval sy...Content-Based Image Retrieval(CBIR)and image mining are becoming more important study fields in computer vision due to their wide range of applications in healthcare,security,and various domains.The image retrieval system mainly relies on the efficiency and accuracy of the classification models.This research addresses the challenge of enhancing the image retrieval system by developing a novel approach,EfficientNet-Convolutional Neural Network(EffNet-CNN).The key objective of this research is to evaluate the proposed EffNet-CNN model’s performance in image classification,image mining,and CBIR.The novelty of the proposed EffNet-CNN model includes the integration of different techniques and modifications.The model includes the Mahalanobis distance metric for feature matching,which enhances the similarity measurements.The model extends EfficientNet architecture by incorporating additional convolutional layers,batch normalization,dropout,and pooling layers for improved hierarchical feature extraction.A systematic hyperparameter optimization using SGD,performance evaluation with three datasets,and data normalization for improving feature representations.The EffNet-CNN is assessed utilizing precision,accuracy,F-measure,and recall metrics across MS-COCO,CIFAR-10 and 100 datasets.The model achieved accuracy values ranging from 90.60%to 95.90%for the MS-COCO dataset,96.8%to 98.3%for the CIFAR-10 dataset and 92.9%to 98.6%for the CIFAR-100 dataset.A validation of the EffNet-CNN model’s results with other models reveals the proposed model’s superior performance.The results highlight the potential of the EffNet-CNN model proposed for image classification and its usefulness in image mining and CBIR.展开更多
Inertial Sensor-based Daily Activity Recognition(IS-DAR)requires adaptable,data-efficient methods for effective multi-sensor use.This study presents an advanced detection system using body-worn sensors to accurately r...Inertial Sensor-based Daily Activity Recognition(IS-DAR)requires adaptable,data-efficient methods for effective multi-sensor use.This study presents an advanced detection system using body-worn sensors to accurately recognize activities.A structured pipeline enhances IS-DAR by applying signal preprocessing,feature extraction and optimization,followed by classification.Before segmentation,a Chebyshev filter removes noise,and Blackman window-ing improves signal representation.Discriminative features-Gaussian Mixture Model(GMM)with Mel-Frequency Cepstral Coefficients(MFCC),spectral entropy,quaternion-based features,and Gammatone Cepstral Coefficients(GCC)-are fused to expand the feature space.Unlike existing approaches,the proposed IS-DAR system uniquely inte-grates diverse handcrafted features using a novel fusion strategy combined with Bayesian-based optimization,enabling a more accurate and generalized activity recognition.The key contribution lies in the joint optimization and fusion of features via Bayesian-based subset selection,resulting in a compact and highly discriminative feature representation.These features are then fed into a Convolutional Neural Network(CNN)to effectively detect spatial-temporal patterns in activity signals.Testing on two public datasets-IM-WSHA and ENABL3S-achieved accuracy levels of 93.0%and 92.0%,respectively.The integration of advanced feature extraction methods with fusion and optimization techniques significantly enhanced detection performance,surpassing traditional methods.The obtained results establish the effectiveness of the proposed IS-DAR system for deployment in real-world activity recognition applications.展开更多
Fundoscopic diagnosis involves assessing the proper functioning of the eye’s nerves,blood vessels,retinal health,and the impact of diabetes on the optic nerves.Fundus disorders are a major global health concern,affec...Fundoscopic diagnosis involves assessing the proper functioning of the eye’s nerves,blood vessels,retinal health,and the impact of diabetes on the optic nerves.Fundus disorders are a major global health concern,affecting millions of people worldwide due to their widespread occurrence.Fundus photography generates machine-based eye images that assist in diagnosing and treating ocular diseases such as diabetic retinopathy.As a result,accurate fundus detection is essential for early diagnosis and effective treatment,helping to prevent severe complications and improve patient outcomes.To address this need,this article introduces a Derivative Model for Fundus Detection using Deep NeuralNetworks(DMFD-DNN)to enhance diagnostic precision.Thismethod selects key features for fundus detection using the least derivative,which identifies features correlating with stored fundus images.Feature filtering relies on the minimum derivative,determined by extracting both similar and varying textures.In this research,the DNN model was integrated with the derivative model.Fundus images were segmented,features were extracted,and the DNN was iteratively trained to identify fundus regions reliably.The goal was to improve the precision of fundoscopic diagnosis by training the DNN incrementally,taking into account the least possible derivative across iterations,and using outputs from previous cycles.The hidden layer of the neural network operates on the most significant derivative,which may reduce precision across iterations.These derivatives are treated as inaccurate,and the model is subsequently trained using selective features and their corresponding extractions.The proposed model outperforms previous techniques in detecting fundus regions,achieving 94.98%accuracy and 91.57%sensitivity,with a minimal error rate of 5.43%.It significantly reduces feature extraction time to 1.462 s and minimizes computational overhead,thereby improving operational efficiency and scalability.Ultimately,the proposed model enhances diagnostic precision and reduces errors,leading to more effective fundus dysfunction diagnosis and treatment.展开更多
基金Saudi Arabia for funding this work through Small Research Group Project under Grant Number RGP.1/316/45.
文摘The effective and timely diagnosis and treatment of ocular diseases are key to the rapid recovery of patients.Today,the mass disease that needs attention in this context is cataracts.Although deep learning has significantly advanced the analysis of ocular disease images,there is a need for a probabilistic model to generate the distributions of potential outcomes and thusmake decisions related to uncertainty quantification.Therefore,this study implements a Bayesian Convolutional Neural Networks(BCNN)model for predicting cataracts by assigning probability values to the predictions.It prepares convolutional neural network(CNN)and BCNN models.The proposed BCNN model is CNN-based in which reparameterization is in the first and last layers of the CNN model.This study then trains them on a dataset of cataract images filtered from the ocular disease fundus images fromKaggle.The deep CNN model has an accuracy of 95%,while the BCNN model has an accuracy of 93.75% along with information on uncertainty estimation of cataracts and normal eye conditions.When compared with other methods,the proposed work reveals that it can be a promising solution for cataract prediction with uncertainty estimation.
基金funded by the Open Access Initiative of the University of Bremen and the DFG via SuUB BremenThe authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Group Project under grant number(RGP2/367/46)+1 种基金This research is supported and funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R410)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘As urban landscapes evolve and vehicular volumes soar,traditional traffic monitoring systems struggle to scale,often failing under the complexities of dense,dynamic,and occluded environments.This paper introduces a novel,unified deep learning framework for vehicle detection,tracking,counting,and classification in aerial imagery designed explicitly for modern smart city infrastructure demands.Our approach begins with adaptive histogram equalization to optimize aerial image clarity,followed by a cutting-edge scene parsing technique using Mask2Former,enabling robust segmentation even in visually congested settings.Vehicle detection leverages the latest YOLOv11 architecture,delivering superior accuracy in aerial contexts by addressing occlusion,scale variance,and fine-grained object differentiation.We incorporate the highly efficient ByteTrack algorithm for tracking,enabling seamless identity preservation across frames.Vehicle counting is achieved through an unsupervised DBSCAN-based method,ensuring adaptability to varying traffic densities.We further introduce a hybrid feature extraction module combining Convolutional Neural Networks(CNNs)with Zernike Moments,capturing both deep semantic and geometric signatures of vehicles.The final classification is powered by NASNet,a neural architecture search-optimized model,ensuring high accuracy across diverse vehicle types and orientations.Extensive evaluations of the VAID benchmark dataset demonstrate the system’s outstanding performance,achieving 96%detection,94%tracking,and 96.4%classification accuracy.On the UAVDT dataset,the system attains 95%detection,93%tracking,and 95%classification accuracy,confirming its robustness across diverse aerial traffic scenarios.These results establish new benchmarks in aerial traffic analysis and validate the framework’s scalability,making it a powerful and adaptable solution for next-generation intelligent transportation systems and urban surveillance.
基金supported by the ITP(Institute of Information&Communications Technology Planning&Evaluation)-ICAN(ICT Challenge and Advanced Network of HRD)(ITP-2025-RS-2022-00156326,33)grant funded by the Korea government(Ministry of Science and ICT)the Deanship of Research and Graduate Studies at King Khalid University for funding this work through the Large Group Project under grant number(RGP2/568/45)the Deanship of Scientific Research at Northern Border University,Arar,Saudi Arabia,for funding this research work through the Project Number"NBU-FFR-2025-231-03".
文摘Remote sensing plays a pivotal role in environmental monitoring,disaster relief,and urban planning,where accurate scene classification of aerial images is essential.However,conventional convolutional neural networks(CNNs)struggle with long-range dependencies and preserving high-resolution features,limiting their effectiveness in complex aerial image analysis.To address these challenges,we propose a Hybrid HRNet-Swin Transformer model that synergizes the strengths of HRNet-W48 for high-resolution segmentation and the Swin Transformer for global feature extraction.This hybrid architecture ensures robust multi-scale feature fusion,capturing fine-grained details and broader contextual relationships in aerial imagery.Our methodology begins with preprocessing steps,including normalization,histogram equalization,and noise reduction,to enhance input data quality.The HRNet-W48 backbone maintains high-resolution feature maps throughout the network,enabling precise segmentation,while the Swin Transformer leverages hierarchical self-attention to model long-range dependencies efficiently.By integrating these components,our model achieves superior performance in segmentation and classification tasks compared to traditional CNNs and standalone transformer models.We evaluate our approach on two benchmark datasets:UC Merced and WHU-RS19.Experimental results demonstrate that the proposed hybrid model outperforms existing methods,achieving state-of-the-art accuracy while maintaining computational efficiency.Specifically,it excels in preserving fine spatial details and contextual understanding,critical for applications like land-use classification and disaster assessment.
基金the Deanship of Research and Graduate Studies at King Khalid University for funding this work through Large Group Project under grant number(RGP.2/245/46)funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R760)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The research team thanks the Deanship of Graduate Studies and Scientific Research at Najran University for supporting the research project through the Nama’a program,with the project code NU/GP/SERC/13/352-1。
文摘The emergence of Generative Adversarial Network(GAN)techniques has garnered significant attention from the research community for the development of Intrusion Detection Systems(IDS).However,conventional GAN-based IDS models face several challenges,including training instability,high computational costs,and system failures.To address these limitations,we propose a Hybrid Wasserstein GAN and Autoencoder Model(WGAN-AE)for intrusion detection.The proposed framework leverages the stability of WGAN and the feature extraction capabilities of the Autoencoder Model.The model was trained and evaluated using two recent benchmark datasets,5GNIDD and IDSIoT2024.When trained on the 5GNIDD dataset,the model achieved an average area under the precisionrecall curve is 99.8%using five-fold cross-validation and demonstrated a high detection accuracy of 97.35%when tested on independent test data.Additionally,the model is well-suited for deployment on resource-limited Internetof-Things(IoT)devices due to its ability to detect attacks within microseconds and its small memory footprint of 60.24 kB.Similarly,when trained on the IDSIoT2024 dataset,the model achieved an average PR-AUC of 94.09%and an attack detection accuracy of 97.35%on independent test data,with a memory requirement of 61.84 kB.Extensive simulation results demonstrate that the proposed hybrid model effectively addresses the shortcomings of traditional GAN-based IDS approaches in terms of detection accuracy,computational efficiency,and applicability to real-world IoT environments.
基金the Deanship of Scientific Research at King Khalid University for funding this research through the Large Group Research Project under grant number RGP2/324/46.
文摘Intelligent Traffic Management(ITM)has progressively developed into a critical component of modern transportation networks,significantly enhancing traffic flow and reducing congestion in urban environments.This research proposes an enhanced framework that leverages Deep Q-Learning(DQL),Game Theory(GT),and Stochastic Optimization(SO)to tackle the complex dynamics in transportation networks.The DQL component utilizes the distribution of traffic conditions for epsilon-greedy policy formulation and action and choice reward calculation,ensuring resilient decision-making.GT models the interaction between vehicles and intersections through probabilistic distributions of various features to enhance performance.Results demonstrate that the proposed framework is a scalable solution for dynamic optimization in transportation networks.
基金The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding this work through the Large Group Project under grant number(RGP2/337/46)The research team thanks the Deanship of Graduate Studies and Scientific Research at Najran University for supporting the research project through the Nama’a program,with the project code NU/GP/SERC/13/352-4.
文摘Edge computing(EC)combined with the Internet of Things(IoT)provides a scalable and efficient solution for smart homes.Therapid proliferation of IoT devices poses real-time data processing and security challenges.EC has become a transformative paradigm for addressing these challenges,particularly in intrusion detection and anomaly mitigation.The widespread connectivity of IoT edge networks has exposed them to various security threats,necessitating robust strategies to detect malicious activities.This research presents a privacy-preserving federated anomaly detection framework combined with Bayesian game theory(BGT)and double deep Q-learning(DDQL).The proposed framework integrates BGT to model attacker and defender interactions for dynamic threat level adaptation and resource availability.It also models a strategic layout between attackers and defenders that takes into account uncertainty.DDQL is incorporated to optimize decision-making and aids in learning optimal defense policies at the edge,thereby ensuring policy and decision optimization.Federated learning(FL)enables decentralized and unshared anomaly detection for sensitive data between devices.Data collection has been performed from various sensors in a real-time EC-IoT network to identify irregularities that occurred due to different attacks.The results reveal that the proposed model achieves high detection accuracy of up to 98%while maintaining low resource consumption.This study demonstrates the synergy between game theory and FL to strengthen anomaly detection in EC-IoT networks.
基金The authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University,Kingdom of Saudi Arabia,for funding this work through the Small Research Group Project under Grant Number RGP.1/316/45.
文摘Content-Based Image Retrieval(CBIR)and image mining are becoming more important study fields in computer vision due to their wide range of applications in healthcare,security,and various domains.The image retrieval system mainly relies on the efficiency and accuracy of the classification models.This research addresses the challenge of enhancing the image retrieval system by developing a novel approach,EfficientNet-Convolutional Neural Network(EffNet-CNN).The key objective of this research is to evaluate the proposed EffNet-CNN model’s performance in image classification,image mining,and CBIR.The novelty of the proposed EffNet-CNN model includes the integration of different techniques and modifications.The model includes the Mahalanobis distance metric for feature matching,which enhances the similarity measurements.The model extends EfficientNet architecture by incorporating additional convolutional layers,batch normalization,dropout,and pooling layers for improved hierarchical feature extraction.A systematic hyperparameter optimization using SGD,performance evaluation with three datasets,and data normalization for improving feature representations.The EffNet-CNN is assessed utilizing precision,accuracy,F-measure,and recall metrics across MS-COCO,CIFAR-10 and 100 datasets.The model achieved accuracy values ranging from 90.60%to 95.90%for the MS-COCO dataset,96.8%to 98.3%for the CIFAR-10 dataset and 92.9%to 98.6%for the CIFAR-100 dataset.A validation of the EffNet-CNN model’s results with other models reveals the proposed model’s superior performance.The results highlight the potential of the EffNet-CNN model proposed for image classification and its usefulness in image mining and CBIR.
基金funded by the Open Access Initiative of the University of Bremen and the DFG via SuUB BremenThe authors extend their appreciation to the Deanship of Research and Graduate Studies at King Khalid University for funding thiswork through Large Group Project under grant number(RGP.2/568/45)The authors extend their appreciation to the Deanship of Scientific Research at Northern Border University,Arar,KSA for funding this research work through the Project Number“NBU-FFR-2025-231-04”.
文摘Inertial Sensor-based Daily Activity Recognition(IS-DAR)requires adaptable,data-efficient methods for effective multi-sensor use.This study presents an advanced detection system using body-worn sensors to accurately recognize activities.A structured pipeline enhances IS-DAR by applying signal preprocessing,feature extraction and optimization,followed by classification.Before segmentation,a Chebyshev filter removes noise,and Blackman window-ing improves signal representation.Discriminative features-Gaussian Mixture Model(GMM)with Mel-Frequency Cepstral Coefficients(MFCC),spectral entropy,quaternion-based features,and Gammatone Cepstral Coefficients(GCC)-are fused to expand the feature space.Unlike existing approaches,the proposed IS-DAR system uniquely inte-grates diverse handcrafted features using a novel fusion strategy combined with Bayesian-based optimization,enabling a more accurate and generalized activity recognition.The key contribution lies in the joint optimization and fusion of features via Bayesian-based subset selection,resulting in a compact and highly discriminative feature representation.These features are then fed into a Convolutional Neural Network(CNN)to effectively detect spatial-temporal patterns in activity signals.Testing on two public datasets-IM-WSHA and ENABL3S-achieved accuracy levels of 93.0%and 92.0%,respectively.The integration of advanced feature extraction methods with fusion and optimization techniques significantly enhanced detection performance,surpassing traditional methods.The obtained results establish the effectiveness of the proposed IS-DAR system for deployment in real-world activity recognition applications.
基金supported by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2021R1F1A1055408)supported by the Researchers Supporting Project Number(MHIRSP2024005)Almaarefa University,Riyadh,Saudi Arabia.
文摘Fundoscopic diagnosis involves assessing the proper functioning of the eye’s nerves,blood vessels,retinal health,and the impact of diabetes on the optic nerves.Fundus disorders are a major global health concern,affecting millions of people worldwide due to their widespread occurrence.Fundus photography generates machine-based eye images that assist in diagnosing and treating ocular diseases such as diabetic retinopathy.As a result,accurate fundus detection is essential for early diagnosis and effective treatment,helping to prevent severe complications and improve patient outcomes.To address this need,this article introduces a Derivative Model for Fundus Detection using Deep NeuralNetworks(DMFD-DNN)to enhance diagnostic precision.Thismethod selects key features for fundus detection using the least derivative,which identifies features correlating with stored fundus images.Feature filtering relies on the minimum derivative,determined by extracting both similar and varying textures.In this research,the DNN model was integrated with the derivative model.Fundus images were segmented,features were extracted,and the DNN was iteratively trained to identify fundus regions reliably.The goal was to improve the precision of fundoscopic diagnosis by training the DNN incrementally,taking into account the least possible derivative across iterations,and using outputs from previous cycles.The hidden layer of the neural network operates on the most significant derivative,which may reduce precision across iterations.These derivatives are treated as inaccurate,and the model is subsequently trained using selective features and their corresponding extractions.The proposed model outperforms previous techniques in detecting fundus regions,achieving 94.98%accuracy and 91.57%sensitivity,with a minimal error rate of 5.43%.It significantly reduces feature extraction time to 1.462 s and minimizes computational overhead,thereby improving operational efficiency and scalability.Ultimately,the proposed model enhances diagnostic precision and reduces errors,leading to more effective fundus dysfunction diagnosis and treatment.