Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal...Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.展开更多
Human Activity Recognition(HAR)represents a rapidly advancing research domain,propelled by continuous developments in sensor technologies and the Internet of Things(IoT).Deep learning has become the dominant paradigm ...Human Activity Recognition(HAR)represents a rapidly advancing research domain,propelled by continuous developments in sensor technologies and the Internet of Things(IoT).Deep learning has become the dominant paradigm in sensor-based HAR systems,offering significant advantages over traditional machine learning methods by eliminating manual feature extraction,enhancing recognition accuracy for complex activities,and enabling the exploitation of unlabeled data through generative models.This paper provides a comprehensive review of recent advancements and emerging trends in deep learning models developed for sensor-based human activity recognition(HAR)systems.We begin with an overview of fundamental HAR concepts in sensor-driven contexts,followed by a systematic categorization and summary of existing research.Our survey encompasses a wide range of deep learning approaches,including Multi-Layer Perceptrons(MLP),Convolutional Neural Networks(CNN),Recurrent Neural Networks(RNN),Long Short-Term Memory networks(LSTM),Gated Recurrent Units(GRU),Transformers,Deep Belief Networks(DBN),and hybrid architectures.A comparative evaluation of these models is provided,highlighting their performance,architectural complexity,and contributions to the field.Beyond Centralized deep learning models,we examine the role of Federated Learning(FL)in HAR,highlighting current applications and research directions.Finally,we discuss the growing importance of Explainable Artificial Intelligence(XAI)in sensor-based HAR,reviewing recent studies that integrate interpretability methods to enhance transparency and trustworthiness in deep learning-based HAR systems.展开更多
As mobile edge computing continues to develop,the demand for resource-intensive applications is steadily increasing,placing a significant strain on edge nodes.These nodes are normally subject to various constraints,fo...As mobile edge computing continues to develop,the demand for resource-intensive applications is steadily increasing,placing a significant strain on edge nodes.These nodes are normally subject to various constraints,for instance,limited processing capability,a few energy sources,and erratic availability being some of the common ones.Correspondingly,these problems require an effective task allocation algorithmto optimize the resources through continued high system performance and dependability in dynamic environments.This paper proposes an improved Particle Swarm Optimization technique,known as IPSO,for multi-objective optimization in edge computing to overcome these issues.To this end,the IPSO algorithm tries to make a trade-off between two important objectives,which are energy consumption minimization and task execution time reduction.Because of global optimal position mutation and dynamic adjustment to inertia weight,the proposed optimization algorithm can effectively distribute tasks among edge nodes.As a result,it reduces the execution time of tasks and energy consumption.In comparative assessments carried out by IPSO with benchmark methods such as Energy-aware Double-fitness Particle Swarm Optimization(EADPSO)and ICBA,IPSO provides better results than these algorithms.For the maximum task size,when compared with the benchmark methods,IPSO reduces the execution time by 17.1%and energy consumption by 31.58%.These results allow the conclusion that IPSO is an efficient and scalable technique for task allocation at the edge environment.It provides peak efficiency while handling scarce resources and variable workloads.展开更多
This review article provides a comprehensive analysis of the latest advancements and persistent challenges in Software-Defined Wide Area Networks(SD-WANs),with a particular emphasis on the multi-objective Controller P...This review article provides a comprehensive analysis of the latest advancements and persistent challenges in Software-Defined Wide Area Networks(SD-WANs),with a particular emphasis on the multi-objective Controller Placement Problem(CPP).As SD-WAN technology continues to gain prominence for its capacity to offer flexible and efficient network management,the task of 36optimally placing controllers—responsible for orchestrating and managing network traffic—remains a critical yet complex challenge.This review delves into recent innovations in multi-objective controller placement strategies,including clustering techniques,heuristic-based approaches,and the integration of machine learning and deep learning models.Each methodology is critically evaluated in terms of its ability to minimize network latency,enhance fault tolerance,and improve overall network performance.Furthermore,this paper discusses the inherent limitations and challenges associated with these techniques,providing a critical evaluation of their current utility and outlining potential avenues for future research.By offering a thorough overview of state-of-the-art approaches to multi-objective controller placement in SD-WANs,this review aims to inform ongoing advancements and highlight emerging research opportunities in this evolving field.展开更多
Blockchain interoperability enables seamless communication and asset transfer across isolated permissioned blockchain systems,but it introduces significant security and privacy vulnerabilities.This review aims to syst...Blockchain interoperability enables seamless communication and asset transfer across isolated permissioned blockchain systems,but it introduces significant security and privacy vulnerabilities.This review aims to systematically assess the security and privacy landscape of interoperability protocols for permissioned blockchains,identifying key properties,attack vectors,and countermeasures.Using PRISMA 2020 guidelines,we analysed 56 peerreviewed studies published between 2020 and 2025,retrieved from Scopus,ScienceDirect,Web of Science,and IEEE Xplore.The review focused on interoperability protocols for permissioned blockchains with security and privacy analyses,including only English-language journal articles and conference proceedings.Risk of bias in the included studies was assessed using the MMAT.Methods for presenting and synthesizing results included descriptive analysis,bibliometric analysis,and content analysis,with findings organized into tables,charts,and comparative summaries.The review classifies interoperability protocols into relay,sidechain,notary scheme,HTLC,and hybrid types and identifies 18 security and privacy properties along with 31 known attack types.Relay-based protocols showed the broadest security coverage,while HTLC and notary schemes demonstrated significant security gaps.Notably,93% of studies examined fewer than four properties or attack types,indicating a fragmented research landscape.The review identifies underexplored areas such as ACID properties,decentralization,and cross-chain attack resilience.It further highlights effective countermeasures,including cryptographic techniques,trusted execution environments,zero-knowledge proofs,and decentralized identity schemes.The findings suggest that despite growing adoption,current interoperability protocols lack comprehensive security evaluations.More holistic research is needed to ensure the resilience,trustworthiness,and scalability of cross-chain operations in permissioned blockchain ecosystems.展开更多
Parkinson’s disease(PD)is a progressive neurodegenerative disorder characterized by tremors,rigidity,and decreased movement.PD poses risks to individuals’lives and independence.Early detection of PD is essential bec...Parkinson’s disease(PD)is a progressive neurodegenerative disorder characterized by tremors,rigidity,and decreased movement.PD poses risks to individuals’lives and independence.Early detection of PD is essential because it allows timely intervention,which can slow disease progression and improve outcomes.Manual diagnosis of PD is problematic because it is difficult to capture the subtle patterns and changes that help diagnose PD.In addition,the subjectivity and lack of doctors compared to the number of patients constitute an obstacle to early diagnosis.Artificial intelligence(AI)techniques,especially deep and automated learning models,provide promising solutions to address deficiencies in manual diagnosis.This study develops robust systems for PD diagnosis by analyzing handwritten helical and wave graphical images.Handwritten graphic images of the PD dataset are enhanced using two overlapping filters,the average filter and the Laplacian filter,to improve image quality and highlight essential features.The enhanced images are segmented to isolate regions of interest(ROIs)from the rest of the image using a gradient vector flow(GVF)algorithm,which ensures that features are extracted from only relevant regions.The segmented ROIs are fed into convolutional neural network(CNN)models,namely DenseNet169,MobileNet,and VGG16,to extract fine and deep feature maps that capture complex patterns and representations relevant to PD diagnosis.Fine and deep feature maps extracted from individual CNN models are combined into fused feature vectors for DenseNet169-MobileNet,MobileNet-VGG16,DenseNet169-VGG16,and DenseNet169-MobileNet-VGG16 models.This fusion technique aims to combine complementary and robust features from several models,which improves the extracted features.Two feature selection algorithms are considered to remove redundancy and weak correlations within the combined feature set:Ant Colony Optimization(ACO)and Maximum Entropy Score-based Selection(MESbS).These algorithms identify and retain the most strongly correlated features while eliminating redundant and weakly correlated features,thus optimizing the features to improve system performance.The fused and enhanced feature vectors are fed into two powerful classifiers,XGBoost and random forest(RF),for accurate classification and differentiation between individuals with PD and healthy controls.The proposed hybrid systems show superior performance,where the RF classifier used the combined features from the DenseNet169-MobileNet-VGG16 models with the ACO feature selection method,achieving outstanding results:area under the curve(AUC)of 99%,sensitivity of 99.6%,99.3%accuracy,99.35%accuracy,and 99.65%specificity.展开更多
Lung cancer is the most dangerous and death-causing disease indicated by the presence of pulmonary nodules in the lung.It is mostly caused by the instinctive growth of cells in the lung.Lung nodule detection has a sig...Lung cancer is the most dangerous and death-causing disease indicated by the presence of pulmonary nodules in the lung.It is mostly caused by the instinctive growth of cells in the lung.Lung nodule detection has a significant role in detecting and screening lung cancer in Computed tomography(CT)scan images.Early detection plays an important role in the survival rate and treatment of lung cancer patients.Moreover,pulmonary nodule classification techniques based on the convolutional neural network can be used for the accurate and efficient detection of lung cancer.This work proposed an automatic nodule detection method in CT images based on modified AlexNet architecture and Support vector machine(SVM)algorithm namely LungNet-SVM.The proposed model consists of seven convolutional layers,three pooling layers,and two fully connected layers used to extract features.Support vector machine classifier is applied for the binary classification of nodules into benign andmalignant.The experimental analysis is performed by using the publicly available benchmark dataset Lung nodule analysis 2016(LUNA16).The proposed model has achieved 97.64%of accuracy,96.37%of sensitivity,and 99.08%of specificity.A comparative analysis has been carried out between the proposed LungNet-SVM model and existing stateof-the-art approaches for the classification of lung cancer.The experimental results indicate that the proposed LungNet-SVM model achieved remarkable performance on a LUNA16 dataset in terms of accuracy.展开更多
Universal Soil Loss Equation (USLE) is the most comprehensive technique available to predict the long term average annual rate of erosion on a field slope. USLE was governed by five factors include soil erodibility fa...Universal Soil Loss Equation (USLE) is the most comprehensive technique available to predict the long term average annual rate of erosion on a field slope. USLE was governed by five factors include soil erodibility factor (K), rainfall and runoff erodibility index (R), crop/vegetation and management factor (C), support practice factor (P) and slope length-gradient factor (LS). In the past, K, R and LS factors are extensively studied. But the impacts of factors C and P to outfall Total Suspended Solid (TSS) and % reduction of TSS are not fully studied yet. Therefore, this study employs Buffer Zone Calculator as a tool to determine the sediment removal efficiency for different C and P factors. The selected study areas are Santubong River, Kuching, Sarawak. Results show that the outfall TSS is increasing with the increase of C values. The most effective and efficient land use for reducing TSS among 17 land uses investigated is found to be forest with undergrowth, followed by mixed dipt. forest, forest with no undergrowth, cultivated grass, logging 30, logging 10^6, wet rice, new shifting agriculture, oil palm, rubber, cocoa, coffee, tea and lastly settlement/cleared land. Besides, results also indicate that the % reduction of TSS is increasing with the decrease of P factor. The most effective support practice to reduce the outfall TSS is found to be terracing, followed by contour-strip cropping, contouring and lastly not implementing any soil conservation practice.展开更多
The quick spread of the CoronavirusDisease(COVID-19)infection around the world considered a real danger for global health.The biological structure and symptoms of COVID-19 are similar to other viral chest maladies,whi...The quick spread of the CoronavirusDisease(COVID-19)infection around the world considered a real danger for global health.The biological structure and symptoms of COVID-19 are similar to other viral chest maladies,which makes it challenging and a big issue to improve approaches for efficient identification of COVID-19 disease.In this study,an automatic prediction of COVID-19 identification is proposed to automatically discriminate between healthy and COVID-19 infected subjects in X-ray images using two successful moderns are traditional machine learning methods(e.g.,artificial neural network(ANN),support vector machine(SVM),linear kernel and radial basis function(RBF),k-nearest neighbor(k-NN),Decision Tree(DT),andCN2 rule inducer techniques)and deep learningmodels(e.g.,MobileNets V2,ResNet50,GoogleNet,DarkNet andXception).A largeX-ray dataset has been created and developed,namely the COVID-19 vs.Normal(400 healthy cases,and 400 COVID cases).To the best of our knowledge,it is currently the largest publicly accessible COVID-19 dataset with the largest number of X-ray images of confirmed COVID-19 infection cases.Based on the results obtained from the experiments,it can be concluded that all the models performed well,deep learning models had achieved the optimum accuracy of 98.8%in ResNet50 model.In comparison,in traditional machine learning techniques, the SVM demonstrated the best result for an accuracy of 95% and RBFaccuracy 94% for the prediction of coronavirus disease 2019.展开更多
implementation of wireless technologies based on the vehicular ad hoc sensor network (VASNET) may provide support for the search and rescue (SAR) team to operate effectively in natural disaster events, such as lan...implementation of wireless technologies based on the vehicular ad hoc sensor network (VASNET) may provide support for the search and rescue (SAR) team to operate effectively in natural disaster events, such as landslide, earthquake, flooding, and tsunami. The operations of SAR team are very challenging in such events due to the possible damages of the existing telecommunication infrastructures. The existing deployment of the cellular communications infrastructure may be partially or completely destroyed after the occurrence of these natural disasters. Thus, the current VASNET infrastructure must be able to support the infrastructure-less network by integrating other green wireless technologies that can benefit the SAR team, which can indirectly save more human lives and reduce the number of casualties. Therefore, the integration of green Internet of things (loT) and VASNET is proposed to form a heterogeneous framework for data dissemination in SAR operations. In addition, this paper also discusses the existing lot framework in disaster scenarios with future research direction for IoT using on any aspect, especially related to the natural disaster scenarios.展开更多
Crop diseases have a significant impact on plant growth and can lead to reduced yields.Traditional methods of disease detection rely on the expertise of plant protection experts,which can be subjective and dependent o...Crop diseases have a significant impact on plant growth and can lead to reduced yields.Traditional methods of disease detection rely on the expertise of plant protection experts,which can be subjective and dependent on individual experience and knowledge.To address this,the use of digital image recognition technology and deep learning algorithms has emerged as a promising approach for automating plant disease identification.In this paper,we propose a novel approach that utilizes a convolutional neural network(CNN)model in conjunction with Inception v3 to identify plant leaf diseases.The research focuses on developing a mobile application that leverages this mechanism to identify diseases in plants and provide recommendations for overcoming specific diseases.The models were trained using a dataset consisting of 80,848 images representing 21 different plant leaves categorized into 60 distinct classes.Through rigorous training and evaluation,the proposed system achieved an impressive accuracy rate of 99%.This mobile application serves as a convenient and valuable advisory tool,providing early detection and guidance in real agricultural environments.The significance of this research lies in its potential to revolutionize plant disease detection and management practices.By automating the identification process through deep learning algorithms,the proposed system eliminates the subjective nature of expert-based diagnosis and reduces dependence on individual expertise.The integration of mobile technology further enhances accessibility and enables farmers and agricultural practitioners to swiftly and accurately identify diseases in their crops.展开更多
This paper proposes a novel framework to detect cyber-attacks using Machine Learning coupled with User Behavior Analytics.The framework models the user behavior as sequences of events representing the user activities ...This paper proposes a novel framework to detect cyber-attacks using Machine Learning coupled with User Behavior Analytics.The framework models the user behavior as sequences of events representing the user activities at such a network.The represented sequences are thenfitted into a recurrent neural network model to extract features that draw distinctive behavior for individual users.Thus,the model can recognize frequencies of regular behavior to profile the user manner in the network.The subsequent procedure is that the recurrent neural network would detect abnormal behavior by classifying unknown behavior to either regu-lar or irregular behavior.The importance of the proposed framework is due to the increase of cyber-attacks especially when the attack is triggered from such sources inside the network.Typically detecting inside attacks are much more challenging in that the security protocols can barely recognize attacks from trustful resources at the network,including users.Therefore,the user behavior can be extracted and ultimately learned to recognize insightful patterns in which the regular patterns reflect a normal network workflow.In contrast,the irregular patterns can trigger an alert for a potential cyber-attack.The framework has been fully described where the evaluation metrics have also been introduced.The experimental results show that the approach performed better compared to other approaches and AUC 0.97 was achieved using RNN-LSTM 1.The paper has been concluded with pro-viding the potential directions for future improvements.展开更多
Samarahan has transformed from a small village into education hub for the past 2 decades. Rapid development and population growth had led to speedy growth in water demand. The situation is getting worse as the pipes a...Samarahan has transformed from a small village into education hub for the past 2 decades. Rapid development and population growth had led to speedy growth in water demand. The situation is getting worse as the pipes are deteriorating due to pipe aging. Therefore, there is a need to study the adequacy of water supply and relationships among roughness coefficient (C) values in Hazen Williams’ Equation with head loss and water pressure due to pipe aging at Uni-Central, a residential area located at Samarahan Sarawak. Investigations were carried out with Ductile Iron, Abestos Cement and Cast Iron pipes at age categories of 0 - 10 years, 10 - 30 years, 30 - 50 years, 50 - 70 years and >70 years. Six critical nodes named as A, B, C, D, E and F were identified to study the water pressure and head loss. Model was developed with InfoWorks Water Supply (WS) Pro software. The impact of pipe aging and materials to water pressure and head loss was not significant at Nodes A, B, C and F. However, max water pressure at Nodes D and F were only reaching 6.30 m and 7.30 m, respectively for all investigations. Therefore, some improvement works are required. Results also show that Asbestos Cement pipe has the least impact on the head loss and water pressure, followed by Ductile Iron pipe and lastly Cast Iron pipe. Simulation results also revealed that older pipes have higher roughness coefficients, indicated with lower “C” values, thus increase the head loss and reduce the water pressure. In contrast, as “C” values increased, head loss will be reduced and water pressure will be increased.展开更多
The paper presents the simulation results of the comparison of three Queuing Mechanisms, First in First out (FIFO), Priority Queuing (PQ), and Weighted Fair Queuing (WFQ). Depending on their effects on the network’s ...The paper presents the simulation results of the comparison of three Queuing Mechanisms, First in First out (FIFO), Priority Queuing (PQ), and Weighted Fair Queuing (WFQ). Depending on their effects on the network’s Routers, the load of any algorithm of them over Router’s CPUs and memory usage, the delay occurred between routers when any algorithm has been used and the network application throughput. This comparison explains that, PQ doesn’t need high specification hardware (memory and CPU) but when used it is not fair, because it serves one application and ignore the other application and FIFO mechanism has smaller queuing delay, otherwise PQ has bigger delay.展开更多
Background The clinical significance of complete revascularization for ST segment elevation myocardial infarction (STEMI) pa- tients during admission is still debatable. Methods A total of 1406 STEMI patients from t...Background The clinical significance of complete revascularization for ST segment elevation myocardial infarction (STEMI) pa- tients during admission is still debatable. Methods A total of 1406 STEMI patients from the Korean Myocardial Infarction Registry with multivessel diseases without cardiogenic shock who underwent primary percutaneous coronary intervention (PPCI) were analyzed. We used propensity score matching (PSM) to control differences of baseline characteristics between culprit only intervention (CP) and multivessel percutaneous coronary interventions (MP), and between double vessel disease (DVD) and triple vessel disease (TVD). The major adverse cardiac event (MACE) was analyzed for one year after discharge. Results TVD patients showed higher incidence of MACE (14.2% vs. 8.6%, P = 0.01), any cause of revascularization (10.6% vs. 5.9%, P - 0.01), and repeated PCI (9.5% vs. 5.7%, P = 0.02), as compared to DVD patients during one year after discharge. MP reduced MACE effectively (7.3% vs. 13.8%, P = 0.03), as compared to CP for one year, but all cause of death (1.6% vs. 3.2%, P= 0.38), Ml (0.4% vs. 0.8%, P = 1.00), and any cause ofrevascularization (5.3% vs. 9.7%, P = 0.09) were comparable in the two treatment groups. Conclusions STEMI patients with TVD showed higher rate of MACE, as compared to DVD MP performed during PPCI or ad hoc during admission for STEMI patients without cardiogenic shock showed lower rate of MACE in this large scaled database. Therefore, MP could be considered as an effective treatment option for STEMI patients without cardiogenic shock.展开更多
Deep learning(DL)has emerged as a powerful subset of machine learning(ML)and artificial intelligence(AI),outperforming traditional ML methods,especially in handling unstructured and large datasets.Its impact spans acr...Deep learning(DL)has emerged as a powerful subset of machine learning(ML)and artificial intelligence(AI),outperforming traditional ML methods,especially in handling unstructured and large datasets.Its impact spans across various domains,including speech recognition,healthcare,autonomous vehicles,cybersecurity,predictive analytics,and more.However,the complexity and dynamic nature of real-world problems present challenges in designing effective deep learning models.Consequently,several deep learning models have been developed to address different problems and applications.In this article,we conduct a comprehensive survey of various deep learning models,including Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),Temporal Convolutional Networks(TCN),Transformer,Kolmogorov-Arnold Networks(KAN),Generative Models,Deep Reinforcement Learning(DRL),and Deep Transfer Learning.We examine the structure,applications,benefits,and limitations of each model.Furthermore,we perform an analysis using three publicly available datasets:IMDB,ARAS,and Fruit-360.We compared the performance of six renowned deep learning models:CNN,RNN,Long Short-Term Memory(LSTM),Bidirectional LSTM,Gated Recurrent Unit(GRU),and Bidirectional GRU alongside two newer models,TCN and Transformer,using the IMDB and ARAS datasets.Additionally,we evaluated the performance of eight CNN-based models,including VGG(Visual Geometry Group),Inception,ResNet(Residual Network),InceptionResNet,Xception(Extreme Inception),MobileNet,DenseNet(Dense Convolutional Network),and NASNet(Neural Architecture Search Network),for image classification tasks using the Fruit-360 dataset.展开更多
Due to natural disaster and global warning, one can expect unexpected fire, which causes panic among people and extent to death. To reduce the impact of fire, the authors propose a new method for predicting and rating...Due to natural disaster and global warning, one can expect unexpected fire, which causes panic among people and extent to death. To reduce the impact of fire, the authors propose a new method for predicting and rating fire in video through deep-learning models in this work such that rescue team can save lives of people. The proposed method explores a hybrid deep convolutional neural network, which involves motion detection and maximally stable extremal region for detecting and rating fire in video. Further, the authors propose to use a channel-wise attention mechanism of the deep neural network for detecting rating of fire level. Experimental results on a large dataset show the proposed method outperforms the existing methods for detecting and rating fire in video.展开更多
Organizations are adopting the Bring Your Own Device(BYOD)concept to enhance productivity and reduce expenses.However,this trend introduces security challenges,such as unauthorized access.Traditional access control sy...Organizations are adopting the Bring Your Own Device(BYOD)concept to enhance productivity and reduce expenses.However,this trend introduces security challenges,such as unauthorized access.Traditional access control systems,such as Attribute-Based Access Control(ABAC)and Role-Based Access Control(RBAC),are limited in their ability to enforce access decisions due to the variability and dynamism of attributes related to users and resources.This paper proposes a method for enforcing access decisions that is adaptable and dynamic,based on multilayer hybrid deep learning techniques,particularly the Tabular Deep Neural Network Tabular DNN method.This technique transforms all input attributes in an access request into a binary classification(allow or deny)using multiple layers,ensuring accurate and efficient access decision-making.The proposed solution was evaluated using the Kaggle Amazon access control policy dataset and demonstrated its effectiveness by achieving a 94%accuracy rate.Additionally,the proposed solution enhances the implementation of access decisions based on a variety of resource and user attributes while ensuring privacy through indirect communication with the Policy Administration Point(PAP).This solution significantly improves the flexibility of access control systems,making themmore dynamic and adaptable to the evolving needs ofmodern organizations.Furthermore,it offers a scalable approach to manage the complexities associated with the BYOD environment,providing a robust framework for secure and efficient access management.展开更多
Forecasting river flow is crucial for optimal planning,management,and sustainability using freshwater resources.Many machine learning(ML)approaches have been enhanced to improve streamflow prediction.Hybrid techniques...Forecasting river flow is crucial for optimal planning,management,and sustainability using freshwater resources.Many machine learning(ML)approaches have been enhanced to improve streamflow prediction.Hybrid techniques have been viewed as a viable method for enhancing the accuracy of univariate streamflow estimation when compared to standalone approaches.Current researchers have also emphasised using hybrid models to improve forecast accuracy.Accordingly,this paper conducts an updated literature review of applications of hybrid models in estimating streamflow over the last five years,summarising data preprocessing,univariate machine learning modelling strategy,advantages and disadvantages of standalone ML techniques,hybrid models,and performance metrics.This study focuses on two types of hybrid models:parameter optimisation-based hybrid models(OBH)and hybridisation of parameter optimisation-based and preprocessing-based hybridmodels(HOPH).Overall,this research supports the idea thatmeta-heuristic approaches precisely improveML techniques.It’s also one of the first efforts to comprehensively examine the efficiency of various meta-heuristic approaches(classified into four primary classes)hybridised with ML techniques.This study revealed that previous research applied swarm,evolutionary,physics,and hybrid metaheuristics with 77%,61%,12%,and 12%,respectively.Finally,there is still room for improving OBH and HOPH models by examining different data pre-processing techniques and metaheuristic algorithms.展开更多
文摘Honeycombing Lung(HCL)is a chronic lung condition marked by advanced fibrosis,resulting in enlarged air spaces with thick fibrotic walls,which are visible on Computed Tomography(CT)scans.Differentiating between normal lung tissue,honeycombing lungs,and Ground Glass Opacity(GGO)in CT images is often challenging for radiologists and may lead to misinterpretations.Although earlier studies have proposed models to detect and classify HCL,many faced limitations such as high computational demands,lower accuracy,and difficulty distinguishing between HCL and GGO.CT images are highly effective for lung classification due to their high resolution,3D visualization,and sensitivity to tissue density variations.This study introduces Honeycombing Lungs Network(HCL Net),a novel classification algorithm inspired by ResNet50V2 and enhanced to overcome the shortcomings of previous approaches.HCL Net incorporates additional residual blocks,refined preprocessing techniques,and selective parameter tuning to improve classification performance.The dataset,sourced from the University Malaya Medical Centre(UMMC)and verified by expert radiologists,consists of CT images of normal,honeycombing,and GGO lungs.Experimental evaluations across five assessments demonstrated that HCL Net achieved an outstanding classification accuracy of approximately 99.97%.It also recorded strong performance in other metrics,achieving 93%precision,100%sensitivity,89%specificity,and an AUC-ROC score of 97%.Comparative analysis with baseline feature engineering methods confirmed the superior efficacy of HCL Net.The model significantly reduces misclassification,particularly between honeycombing and GGO lungs,enhancing diagnostic precision and reliability in lung image analysis.
文摘Human Activity Recognition(HAR)represents a rapidly advancing research domain,propelled by continuous developments in sensor technologies and the Internet of Things(IoT).Deep learning has become the dominant paradigm in sensor-based HAR systems,offering significant advantages over traditional machine learning methods by eliminating manual feature extraction,enhancing recognition accuracy for complex activities,and enabling the exploitation of unlabeled data through generative models.This paper provides a comprehensive review of recent advancements and emerging trends in deep learning models developed for sensor-based human activity recognition(HAR)systems.We begin with an overview of fundamental HAR concepts in sensor-driven contexts,followed by a systematic categorization and summary of existing research.Our survey encompasses a wide range of deep learning approaches,including Multi-Layer Perceptrons(MLP),Convolutional Neural Networks(CNN),Recurrent Neural Networks(RNN),Long Short-Term Memory networks(LSTM),Gated Recurrent Units(GRU),Transformers,Deep Belief Networks(DBN),and hybrid architectures.A comparative evaluation of these models is provided,highlighting their performance,architectural complexity,and contributions to the field.Beyond Centralized deep learning models,we examine the role of Federated Learning(FL)in HAR,highlighting current applications and research directions.Finally,we discuss the growing importance of Explainable Artificial Intelligence(XAI)in sensor-based HAR,reviewing recent studies that integrate interpretability methods to enhance transparency and trustworthiness in deep learning-based HAR systems.
基金supported by the University Putra Malaysia and the Ministry of Higher Education Malaysia under grantNumber:(FRGS/1/2023/ICT11/UPM/02/3).
文摘As mobile edge computing continues to develop,the demand for resource-intensive applications is steadily increasing,placing a significant strain on edge nodes.These nodes are normally subject to various constraints,for instance,limited processing capability,a few energy sources,and erratic availability being some of the common ones.Correspondingly,these problems require an effective task allocation algorithmto optimize the resources through continued high system performance and dependability in dynamic environments.This paper proposes an improved Particle Swarm Optimization technique,known as IPSO,for multi-objective optimization in edge computing to overcome these issues.To this end,the IPSO algorithm tries to make a trade-off between two important objectives,which are energy consumption minimization and task execution time reduction.Because of global optimal position mutation and dynamic adjustment to inertia weight,the proposed optimization algorithm can effectively distribute tasks among edge nodes.As a result,it reduces the execution time of tasks and energy consumption.In comparative assessments carried out by IPSO with benchmark methods such as Energy-aware Double-fitness Particle Swarm Optimization(EADPSO)and ICBA,IPSO provides better results than these algorithms.For the maximum task size,when compared with the benchmark methods,IPSO reduces the execution time by 17.1%and energy consumption by 31.58%.These results allow the conclusion that IPSO is an efficient and scalable technique for task allocation at the edge environment.It provides peak efficiency while handling scarce resources and variable workloads.
文摘This review article provides a comprehensive analysis of the latest advancements and persistent challenges in Software-Defined Wide Area Networks(SD-WANs),with a particular emphasis on the multi-objective Controller Placement Problem(CPP).As SD-WAN technology continues to gain prominence for its capacity to offer flexible and efficient network management,the task of 36optimally placing controllers—responsible for orchestrating and managing network traffic—remains a critical yet complex challenge.This review delves into recent innovations in multi-objective controller placement strategies,including clustering techniques,heuristic-based approaches,and the integration of machine learning and deep learning models.Each methodology is critically evaluated in terms of its ability to minimize network latency,enhance fault tolerance,and improve overall network performance.Furthermore,this paper discusses the inherent limitations and challenges associated with these techniques,providing a critical evaluation of their current utility and outlining potential avenues for future research.By offering a thorough overview of state-of-the-art approaches to multi-objective controller placement in SD-WANs,this review aims to inform ongoing advancements and highlight emerging research opportunities in this evolving field.
基金supported by the International Scientific and Technological Cooperation Project of Huangpu and Development Districts in Guangzhou(2023GH17)the National Science and Technology Council in Taiwan under grant number NSTC-113-2224-E-027-001,Private Funding(PV009-2023)the KW IPPP(Research Maintenance Fee)Individual/Centre/Group(RMF1506-2021)at Universiti Malaya,Malaysia.
文摘Blockchain interoperability enables seamless communication and asset transfer across isolated permissioned blockchain systems,but it introduces significant security and privacy vulnerabilities.This review aims to systematically assess the security and privacy landscape of interoperability protocols for permissioned blockchains,identifying key properties,attack vectors,and countermeasures.Using PRISMA 2020 guidelines,we analysed 56 peerreviewed studies published between 2020 and 2025,retrieved from Scopus,ScienceDirect,Web of Science,and IEEE Xplore.The review focused on interoperability protocols for permissioned blockchains with security and privacy analyses,including only English-language journal articles and conference proceedings.Risk of bias in the included studies was assessed using the MMAT.Methods for presenting and synthesizing results included descriptive analysis,bibliometric analysis,and content analysis,with findings organized into tables,charts,and comparative summaries.The review classifies interoperability protocols into relay,sidechain,notary scheme,HTLC,and hybrid types and identifies 18 security and privacy properties along with 31 known attack types.Relay-based protocols showed the broadest security coverage,while HTLC and notary schemes demonstrated significant security gaps.Notably,93% of studies examined fewer than four properties or attack types,indicating a fragmented research landscape.The review identifies underexplored areas such as ACID properties,decentralization,and cross-chain attack resilience.It further highlights effective countermeasures,including cryptographic techniques,trusted execution environments,zero-knowledge proofs,and decentralized identity schemes.The findings suggest that despite growing adoption,current interoperability protocols lack comprehensive security evaluations.More holistic research is needed to ensure the resilience,trustworthiness,and scalability of cross-chain operations in permissioned blockchain ecosystems.
文摘Parkinson’s disease(PD)is a progressive neurodegenerative disorder characterized by tremors,rigidity,and decreased movement.PD poses risks to individuals’lives and independence.Early detection of PD is essential because it allows timely intervention,which can slow disease progression and improve outcomes.Manual diagnosis of PD is problematic because it is difficult to capture the subtle patterns and changes that help diagnose PD.In addition,the subjectivity and lack of doctors compared to the number of patients constitute an obstacle to early diagnosis.Artificial intelligence(AI)techniques,especially deep and automated learning models,provide promising solutions to address deficiencies in manual diagnosis.This study develops robust systems for PD diagnosis by analyzing handwritten helical and wave graphical images.Handwritten graphic images of the PD dataset are enhanced using two overlapping filters,the average filter and the Laplacian filter,to improve image quality and highlight essential features.The enhanced images are segmented to isolate regions of interest(ROIs)from the rest of the image using a gradient vector flow(GVF)algorithm,which ensures that features are extracted from only relevant regions.The segmented ROIs are fed into convolutional neural network(CNN)models,namely DenseNet169,MobileNet,and VGG16,to extract fine and deep feature maps that capture complex patterns and representations relevant to PD diagnosis.Fine and deep feature maps extracted from individual CNN models are combined into fused feature vectors for DenseNet169-MobileNet,MobileNet-VGG16,DenseNet169-VGG16,and DenseNet169-MobileNet-VGG16 models.This fusion technique aims to combine complementary and robust features from several models,which improves the extracted features.Two feature selection algorithms are considered to remove redundancy and weak correlations within the combined feature set:Ant Colony Optimization(ACO)and Maximum Entropy Score-based Selection(MESbS).These algorithms identify and retain the most strongly correlated features while eliminating redundant and weakly correlated features,thus optimizing the features to improve system performance.The fused and enhanced feature vectors are fed into two powerful classifiers,XGBoost and random forest(RF),for accurate classification and differentiation between individuals with PD and healthy controls.The proposed hybrid systems show superior performance,where the RF classifier used the combined features from the DenseNet169-MobileNet-VGG16 models with the ACO feature selection method,achieving outstanding results:area under the curve(AUC)of 99%,sensitivity of 99.6%,99.3%accuracy,99.35%accuracy,and 99.65%specificity.
文摘Lung cancer is the most dangerous and death-causing disease indicated by the presence of pulmonary nodules in the lung.It is mostly caused by the instinctive growth of cells in the lung.Lung nodule detection has a significant role in detecting and screening lung cancer in Computed tomography(CT)scan images.Early detection plays an important role in the survival rate and treatment of lung cancer patients.Moreover,pulmonary nodule classification techniques based on the convolutional neural network can be used for the accurate and efficient detection of lung cancer.This work proposed an automatic nodule detection method in CT images based on modified AlexNet architecture and Support vector machine(SVM)algorithm namely LungNet-SVM.The proposed model consists of seven convolutional layers,three pooling layers,and two fully connected layers used to extract features.Support vector machine classifier is applied for the binary classification of nodules into benign andmalignant.The experimental analysis is performed by using the publicly available benchmark dataset Lung nodule analysis 2016(LUNA16).The proposed model has achieved 97.64%of accuracy,96.37%of sensitivity,and 99.08%of specificity.A comparative analysis has been carried out between the proposed LungNet-SVM model and existing stateof-the-art approaches for the classification of lung cancer.The experimental results indicate that the proposed LungNet-SVM model achieved remarkable performance on a LUNA16 dataset in terms of accuracy.
文摘Universal Soil Loss Equation (USLE) is the most comprehensive technique available to predict the long term average annual rate of erosion on a field slope. USLE was governed by five factors include soil erodibility factor (K), rainfall and runoff erodibility index (R), crop/vegetation and management factor (C), support practice factor (P) and slope length-gradient factor (LS). In the past, K, R and LS factors are extensively studied. But the impacts of factors C and P to outfall Total Suspended Solid (TSS) and % reduction of TSS are not fully studied yet. Therefore, this study employs Buffer Zone Calculator as a tool to determine the sediment removal efficiency for different C and P factors. The selected study areas are Santubong River, Kuching, Sarawak. Results show that the outfall TSS is increasing with the increase of C values. The most effective and efficient land use for reducing TSS among 17 land uses investigated is found to be forest with undergrowth, followed by mixed dipt. forest, forest with no undergrowth, cultivated grass, logging 30, logging 10^6, wet rice, new shifting agriculture, oil palm, rubber, cocoa, coffee, tea and lastly settlement/cleared land. Besides, results also indicate that the % reduction of TSS is increasing with the decrease of P factor. The most effective support practice to reduce the outfall TSS is found to be terracing, followed by contour-strip cropping, contouring and lastly not implementing any soil conservation practice.
文摘The quick spread of the CoronavirusDisease(COVID-19)infection around the world considered a real danger for global health.The biological structure and symptoms of COVID-19 are similar to other viral chest maladies,which makes it challenging and a big issue to improve approaches for efficient identification of COVID-19 disease.In this study,an automatic prediction of COVID-19 identification is proposed to automatically discriminate between healthy and COVID-19 infected subjects in X-ray images using two successful moderns are traditional machine learning methods(e.g.,artificial neural network(ANN),support vector machine(SVM),linear kernel and radial basis function(RBF),k-nearest neighbor(k-NN),Decision Tree(DT),andCN2 rule inducer techniques)and deep learningmodels(e.g.,MobileNets V2,ResNet50,GoogleNet,DarkNet andXception).A largeX-ray dataset has been created and developed,namely the COVID-19 vs.Normal(400 healthy cases,and 400 COVID cases).To the best of our knowledge,it is currently the largest publicly accessible COVID-19 dataset with the largest number of X-ray images of confirmed COVID-19 infection cases.Based on the results obtained from the experiments,it can be concluded that all the models performed well,deep learning models had achieved the optimum accuracy of 98.8%in ResNet50 model.In comparison,in traditional machine learning techniques, the SVM demonstrated the best result for an accuracy of 95% and RBFaccuracy 94% for the prediction of coronavirus disease 2019.
文摘implementation of wireless technologies based on the vehicular ad hoc sensor network (VASNET) may provide support for the search and rescue (SAR) team to operate effectively in natural disaster events, such as landslide, earthquake, flooding, and tsunami. The operations of SAR team are very challenging in such events due to the possible damages of the existing telecommunication infrastructures. The existing deployment of the cellular communications infrastructure may be partially or completely destroyed after the occurrence of these natural disasters. Thus, the current VASNET infrastructure must be able to support the infrastructure-less network by integrating other green wireless technologies that can benefit the SAR team, which can indirectly save more human lives and reduce the number of casualties. Therefore, the integration of green Internet of things (loT) and VASNET is proposed to form a heterogeneous framework for data dissemination in SAR operations. In addition, this paper also discusses the existing lot framework in disaster scenarios with future research direction for IoT using on any aspect, especially related to the natural disaster scenarios.
基金supported by the Hainan Provincial Natural Science Foundation of China(No.123QN182)Hainan University Research Fund(Project Nos.KYQD(ZR)-22064,KYQD(ZR)-22063,and KYQD(ZR)-22065).
文摘Crop diseases have a significant impact on plant growth and can lead to reduced yields.Traditional methods of disease detection rely on the expertise of plant protection experts,which can be subjective and dependent on individual experience and knowledge.To address this,the use of digital image recognition technology and deep learning algorithms has emerged as a promising approach for automating plant disease identification.In this paper,we propose a novel approach that utilizes a convolutional neural network(CNN)model in conjunction with Inception v3 to identify plant leaf diseases.The research focuses on developing a mobile application that leverages this mechanism to identify diseases in plants and provide recommendations for overcoming specific diseases.The models were trained using a dataset consisting of 80,848 images representing 21 different plant leaves categorized into 60 distinct classes.Through rigorous training and evaluation,the proposed system achieved an impressive accuracy rate of 99%.This mobile application serves as a convenient and valuable advisory tool,providing early detection and guidance in real agricultural environments.The significance of this research lies in its potential to revolutionize plant disease detection and management practices.By automating the identification process through deep learning algorithms,the proposed system eliminates the subjective nature of expert-based diagnosis and reduces dependence on individual expertise.The integration of mobile technology further enhances accessibility and enables farmers and agricultural practitioners to swiftly and accurately identify diseases in their crops.
基金supported by the fund received from Al Baha University,8/1440.
文摘This paper proposes a novel framework to detect cyber-attacks using Machine Learning coupled with User Behavior Analytics.The framework models the user behavior as sequences of events representing the user activities at such a network.The represented sequences are thenfitted into a recurrent neural network model to extract features that draw distinctive behavior for individual users.Thus,the model can recognize frequencies of regular behavior to profile the user manner in the network.The subsequent procedure is that the recurrent neural network would detect abnormal behavior by classifying unknown behavior to either regu-lar or irregular behavior.The importance of the proposed framework is due to the increase of cyber-attacks especially when the attack is triggered from such sources inside the network.Typically detecting inside attacks are much more challenging in that the security protocols can barely recognize attacks from trustful resources at the network,including users.Therefore,the user behavior can be extracted and ultimately learned to recognize insightful patterns in which the regular patterns reflect a normal network workflow.In contrast,the irregular patterns can trigger an alert for a potential cyber-attack.The framework has been fully described where the evaluation metrics have also been introduced.The experimental results show that the approach performed better compared to other approaches and AUC 0.97 was achieved using RNN-LSTM 1.The paper has been concluded with pro-viding the potential directions for future improvements.
文摘Samarahan has transformed from a small village into education hub for the past 2 decades. Rapid development and population growth had led to speedy growth in water demand. The situation is getting worse as the pipes are deteriorating due to pipe aging. Therefore, there is a need to study the adequacy of water supply and relationships among roughness coefficient (C) values in Hazen Williams’ Equation with head loss and water pressure due to pipe aging at Uni-Central, a residential area located at Samarahan Sarawak. Investigations were carried out with Ductile Iron, Abestos Cement and Cast Iron pipes at age categories of 0 - 10 years, 10 - 30 years, 30 - 50 years, 50 - 70 years and >70 years. Six critical nodes named as A, B, C, D, E and F were identified to study the water pressure and head loss. Model was developed with InfoWorks Water Supply (WS) Pro software. The impact of pipe aging and materials to water pressure and head loss was not significant at Nodes A, B, C and F. However, max water pressure at Nodes D and F were only reaching 6.30 m and 7.30 m, respectively for all investigations. Therefore, some improvement works are required. Results also show that Asbestos Cement pipe has the least impact on the head loss and water pressure, followed by Ductile Iron pipe and lastly Cast Iron pipe. Simulation results also revealed that older pipes have higher roughness coefficients, indicated with lower “C” values, thus increase the head loss and reduce the water pressure. In contrast, as “C” values increased, head loss will be reduced and water pressure will be increased.
文摘The paper presents the simulation results of the comparison of three Queuing Mechanisms, First in First out (FIFO), Priority Queuing (PQ), and Weighted Fair Queuing (WFQ). Depending on their effects on the network’s Routers, the load of any algorithm of them over Router’s CPUs and memory usage, the delay occurred between routers when any algorithm has been used and the network application throughput. This comparison explains that, PQ doesn’t need high specification hardware (memory and CPU) but when used it is not fair, because it serves one application and ignore the other application and FIFO mechanism has smaller queuing delay, otherwise PQ has bigger delay.
文摘Background The clinical significance of complete revascularization for ST segment elevation myocardial infarction (STEMI) pa- tients during admission is still debatable. Methods A total of 1406 STEMI patients from the Korean Myocardial Infarction Registry with multivessel diseases without cardiogenic shock who underwent primary percutaneous coronary intervention (PPCI) were analyzed. We used propensity score matching (PSM) to control differences of baseline characteristics between culprit only intervention (CP) and multivessel percutaneous coronary interventions (MP), and between double vessel disease (DVD) and triple vessel disease (TVD). The major adverse cardiac event (MACE) was analyzed for one year after discharge. Results TVD patients showed higher incidence of MACE (14.2% vs. 8.6%, P = 0.01), any cause of revascularization (10.6% vs. 5.9%, P - 0.01), and repeated PCI (9.5% vs. 5.7%, P = 0.02), as compared to DVD patients during one year after discharge. MP reduced MACE effectively (7.3% vs. 13.8%, P = 0.03), as compared to CP for one year, but all cause of death (1.6% vs. 3.2%, P= 0.38), Ml (0.4% vs. 0.8%, P = 1.00), and any cause ofrevascularization (5.3% vs. 9.7%, P = 0.09) were comparable in the two treatment groups. Conclusions STEMI patients with TVD showed higher rate of MACE, as compared to DVD MP performed during PPCI or ad hoc during admission for STEMI patients without cardiogenic shock showed lower rate of MACE in this large scaled database. Therefore, MP could be considered as an effective treatment option for STEMI patients without cardiogenic shock.
文摘Deep learning(DL)has emerged as a powerful subset of machine learning(ML)and artificial intelligence(AI),outperforming traditional ML methods,especially in handling unstructured and large datasets.Its impact spans across various domains,including speech recognition,healthcare,autonomous vehicles,cybersecurity,predictive analytics,and more.However,the complexity and dynamic nature of real-world problems present challenges in designing effective deep learning models.Consequently,several deep learning models have been developed to address different problems and applications.In this article,we conduct a comprehensive survey of various deep learning models,including Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),Temporal Convolutional Networks(TCN),Transformer,Kolmogorov-Arnold Networks(KAN),Generative Models,Deep Reinforcement Learning(DRL),and Deep Transfer Learning.We examine the structure,applications,benefits,and limitations of each model.Furthermore,we perform an analysis using three publicly available datasets:IMDB,ARAS,and Fruit-360.We compared the performance of six renowned deep learning models:CNN,RNN,Long Short-Term Memory(LSTM),Bidirectional LSTM,Gated Recurrent Unit(GRU),and Bidirectional GRU alongside two newer models,TCN and Transformer,using the IMDB and ARAS datasets.Additionally,we evaluated the performance of eight CNN-based models,including VGG(Visual Geometry Group),Inception,ResNet(Residual Network),InceptionResNet,Xception(Extreme Inception),MobileNet,DenseNet(Dense Convolutional Network),and NASNet(Neural Architecture Search Network),for image classification tasks using the Fruit-360 dataset.
基金supported by National Key R&D Program of China under Grant no.2018YFC0407901the Natural Science Foundation of China under Grant Grant no.61702160,Grant 61672273 and Grant no.61832008+3 种基金the Science Foundation of Jiangsu under Grant BK20170892the Science Foundation for Distinguished Young Scholars of Jiangsu under Grant BK20160021Scientific Foundation of State Grid Corporation of China(Research on Ice-wind Disaster Feature Recognition and Prediction by Few-shot Machine Learning in Transmission Lines)the open Project of the National Key Lab for Novel Software Technology in NJU under Grant K-FKT2017B05.
文摘Due to natural disaster and global warning, one can expect unexpected fire, which causes panic among people and extent to death. To reduce the impact of fire, the authors propose a new method for predicting and rating fire in video through deep-learning models in this work such that rescue team can save lives of people. The proposed method explores a hybrid deep convolutional neural network, which involves motion detection and maximally stable extremal region for detecting and rating fire in video. Further, the authors propose to use a channel-wise attention mechanism of the deep neural network for detecting rating of fire level. Experimental results on a large dataset show the proposed method outperforms the existing methods for detecting and rating fire in video.
基金partly supported by the University of Malaya Impact Oriented Interdisci-plinary Research Grant under Grant IIRG008(A,B,C)-19IISS.
文摘Organizations are adopting the Bring Your Own Device(BYOD)concept to enhance productivity and reduce expenses.However,this trend introduces security challenges,such as unauthorized access.Traditional access control systems,such as Attribute-Based Access Control(ABAC)and Role-Based Access Control(RBAC),are limited in their ability to enforce access decisions due to the variability and dynamism of attributes related to users and resources.This paper proposes a method for enforcing access decisions that is adaptable and dynamic,based on multilayer hybrid deep learning techniques,particularly the Tabular Deep Neural Network Tabular DNN method.This technique transforms all input attributes in an access request into a binary classification(allow or deny)using multiple layers,ensuring accurate and efficient access decision-making.The proposed solution was evaluated using the Kaggle Amazon access control policy dataset and demonstrated its effectiveness by achieving a 94%accuracy rate.Additionally,the proposed solution enhances the implementation of access decisions based on a variety of resource and user attributes while ensuring privacy through indirect communication with the Policy Administration Point(PAP).This solution significantly improves the flexibility of access control systems,making themmore dynamic and adaptable to the evolving needs ofmodern organizations.Furthermore,it offers a scalable approach to manage the complexities associated with the BYOD environment,providing a robust framework for secure and efficient access management.
基金This paper’s logical organisation and content quality have been enhanced,so the authors thank anonymous reviewers and journal editors for assistance.
文摘Forecasting river flow is crucial for optimal planning,management,and sustainability using freshwater resources.Many machine learning(ML)approaches have been enhanced to improve streamflow prediction.Hybrid techniques have been viewed as a viable method for enhancing the accuracy of univariate streamflow estimation when compared to standalone approaches.Current researchers have also emphasised using hybrid models to improve forecast accuracy.Accordingly,this paper conducts an updated literature review of applications of hybrid models in estimating streamflow over the last five years,summarising data preprocessing,univariate machine learning modelling strategy,advantages and disadvantages of standalone ML techniques,hybrid models,and performance metrics.This study focuses on two types of hybrid models:parameter optimisation-based hybrid models(OBH)and hybridisation of parameter optimisation-based and preprocessing-based hybridmodels(HOPH).Overall,this research supports the idea thatmeta-heuristic approaches precisely improveML techniques.It’s also one of the first efforts to comprehensively examine the efficiency of various meta-heuristic approaches(classified into four primary classes)hybridised with ML techniques.This study revealed that previous research applied swarm,evolutionary,physics,and hybrid metaheuristics with 77%,61%,12%,and 12%,respectively.Finally,there is still room for improving OBH and HOPH models by examining different data pre-processing techniques and metaheuristic algorithms.