期刊文献+
共找到1,305篇文章
< 1 2 66 >
每页显示 20 50 100
An Effective and Secure Quality Assurance System for a Computer Science Program 被引量:1
1
作者 Mohammad Alkhatib 《Computer Systems Science & Engineering》 SCIE EI 2022年第6期975-995,共21页
Improving the quality assurance (QA) processes and acquiring accreditation are top priorities for academic programs. The learning outcomes (LOs)assessment and continuous quality improvement represent core components o... Improving the quality assurance (QA) processes and acquiring accreditation are top priorities for academic programs. The learning outcomes (LOs)assessment and continuous quality improvement represent core components ofthe quality assurance system (QAS). Current assessment methods suffer deficiencies related to accuracy and reliability, and they lack well-organized processes forcontinuous improvement planning. Moreover, the absence of automation, andintegration in QA processes forms a major obstacle towards developing efficientquality system. There is a pressing need to adopt security protocols that providerequired security services to safeguard the valuable information processed byQAS as well. This research proposes an effective methodology for LOs assessment and continuous improvement processes. The proposed approach ensuresmore accurate and reliable LOs assessment results and provides systematic wayfor utilizing those results in the continuous quality improvement. This systematicand well-specified QA processes were then utilized to model and implement automated and secure QAS that efficiently performs quality-related processes. Theproposed system adopts two security protocols that provide confidentiality, integrity, and authentication for quality data and reports. The security protocols avoidthe source repudiation, which is important in the quality reporting system. This isachieved through implementing powerful cryptographic algorithms. The QASenables efficient data collection and processing required for analysis and interpretation. It also prepares for the development of datasets that can be used in futureartificial intelligence (AI) researches to support decision making and improve thequality of academic programs. The proposed approach is implemented in a successful real case study for a computer science program. The current study servesscientific programs struggling to achieve academic accreditation, and gives rise tofully automating and integrating the QA processes and adopting modern AI andsecurity technologies to develop effective QAS. 展开更多
关键词 Quality assurance information security cryptographic algorithms education programs
在线阅读 下载PDF
Introduction to the Special Issue on Artificial Intelligence Emerging Trends and Sustainable Applications in Image Processing and Computer Vision
2
作者 Ahmad Taher Azar 《Computer Modeling in Engineering & Sciences》 2025年第7期29-36,共8页
The rapid development of artificial intelligence(AI),machine learning(ML),and deep learning(DL)in recent years has transformed many sectors.A fundamental shift has occurred in approaches to solving complex problems an... The rapid development of artificial intelligence(AI),machine learning(ML),and deep learning(DL)in recent years has transformed many sectors.A fundamental shift has occurred in approaches to solving complex problems and making decisions in many different fields.These advanced technologies have enabled significant breakthroughs in sectors including entertainment,finance,transportation,and healthcare.AI systems,which can analyze vast volumes of data,have significantly driven efficiency and innovation.With remarkable accuracy,patterns can be identified and predictions generated,improving decision-making processes and facilitating the development of more intelligent solutions.The increasing adoption of these technologies by organizations has expanded the potential for AI to change processes and improve results. 展开更多
关键词 machine learning deep learning dl analyze vast volumes datahave artificial intelligence ai machine learning ml advanced technologies solving complex problems efficiency innovationwith artificial intelligence
在线阅读 下载PDF
Efficient Arabic Essay Scoring with Hybrid Models: Feature Selection, Data Optimization, and Performance Trade-Offs
3
作者 Mohamed Ezz Meshrif Alruily +4 位作者 Ayman Mohamed Mostafa Alaa SAlaerjan Bader Aldughayfiq Hisham Allahem Abdulaziz Shehab 《Computers, Materials & Continua》 2026年第1期2274-2301,共28页
Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic... Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage. 展开更多
关键词 Automated essay scoring text-based features vector-based features embedding-based features feature selection optimal data efficiency
在线阅读 下载PDF
Mitigating Attribute Inference in Split Learning via Channel Pruning and Adversarial Training
4
作者 Afnan Alhindi Saad Al-Ahmadi Mohamed Maher Ben Ismail 《Computers, Materials & Continua》 2026年第3期1465-1489,共25页
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn... Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%. 展开更多
关键词 Split learning privacy-preserving split learning distributed collaborative machine learning channel pruning adversarial learning resource-constrained devices
在线阅读 下载PDF
TransCarbonNet:Multi-Day Grid Carbon Intensity Forecasting Using Hybrid Self-Attention and Bi-LSTM Temporal Fusion for Sustainable Energy Management
5
作者 Amel Ksibi Hatoon Albadah +1 位作者 Ghadah Aldehim Manel Ayadi 《Computer Modeling in Engineering & Sciences》 2026年第1期812-847,共36页
Sustainable energy systems will entail a change in the carbon intensity projections,which should be carried out in a proper manner to facilitate the smooth running of the grid and reduce greenhouse emissions.The prese... Sustainable energy systems will entail a change in the carbon intensity projections,which should be carried out in a proper manner to facilitate the smooth running of the grid and reduce greenhouse emissions.The present article outlines the TransCarbonNet,a novel hybrid deep learning framework with self-attention characteristics added to the bidirectional Long Short-Term Memory(Bi-LSTM)network to forecast the carbon intensity of the grid several days.The proposed temporal fusion model not only learns the local temporal interactions but also the long-term patterns of the carbon emission data;hence,it is able to give suitable forecasts over a period of seven days.TransCarbonNet takes advantage of a multi-head self-attention element to identify significant temporal connections,which means the Bi-LSTM element calculates sequential dependencies in both directions.Massive tests on two actual data sets indicate much improved results in comparison with the existing results,with mean relative errors of 15.3 percent and 12.7 percent,respectively.The framework has given explicable weights of attention that reveal critical periods that influence carbon intensity alterations,and informed decisions on the management of carbon sustainability.The effectiveness of the proposed solution has been validated in numerous cases of operations,and TransCarbonNet is established to be an effective tool when it comes to carbon-friendly optimization of the grid. 展开更多
关键词 Carbon intensity forecasting self-attention mechanism bidirectional LSTM temporal fusion sustainable energy management smart grid optimization deep learning
在线阅读 下载PDF
AI-Powered Anomaly Detection and Cybersecurity in Healthcare IoT with Fog-Edge
6
作者 Fatima Al-Quayed 《Computer Modeling in Engineering & Sciences》 2026年第1期1339-1372,共34页
The rapid proliferation of Internet of Things(IoT)devices in critical healthcare infrastructure has introduced significant security and privacy challenges that demand innovative,distributed architectural solutions.Thi... The rapid proliferation of Internet of Things(IoT)devices in critical healthcare infrastructure has introduced significant security and privacy challenges that demand innovative,distributed architectural solutions.This paper proposes FE-ACS(Fog-Edge Adaptive Cybersecurity System),a novel hierarchical security framework that intelligently distributes AI-powered anomaly detection algorithms across edge,fog,and cloud layers to optimize security efficacy,latency,and privacy.Our comprehensive evaluation demonstrates that FE-ACS achieves superior detection performance with an AUC-ROC of 0.985 and an F1-score of 0.923,while maintaining significantly lower end-to-end latency(18.7 ms)compared to cloud-centric(152.3 ms)and fog-only(34.5 ms)architectures.The system exhibits exceptional scalability,supporting up to 38,000 devices with logarithmic performance degradation—a 67×improvement over conventional cloud-based approaches.By incorporating differential privacy mechanisms with balanced privacy-utility tradeoffs(ε=1.0–1.5),FE-ACS maintains 90%–93%detection accuracy while ensuring strong privacy guarantees for sensitive healthcare data.Computational efficiency analysis reveals that our architecture achieves a detection rate of 12,400 events per second with only 12.3 mJ energy consumption per inference.In healthcare risk assessment,FE-ACS demonstrates robust operational viability with low patient safety risk(14.7%)and high system reliability(94.0%).The proposed framework represents a significant advancement in distributed security architectures,offering a scalable,privacy-preserving,and real-time solution for protecting healthcare IoT ecosystems against evolving cyber threats. 展开更多
关键词 AI-powered anomaly detection healthcare IoT fog computing CYBERSECURITY intrusion detection
在线阅读 下载PDF
Lightweight Multi-Agent Edge Framework for Cybersecurity and Resource Optimization in Mobile Sensor Networks
7
作者 Fatima Al-Quayed 《Computers, Materials & Continua》 2026年第1期919-934,共16页
Due to the growth of smart cities,many real-time systems have been developed to support smart cities using Internet of Things(IoT)and emerging technologies.They are formulated to collect the data for environment monit... Due to the growth of smart cities,many real-time systems have been developed to support smart cities using Internet of Things(IoT)and emerging technologies.They are formulated to collect the data for environment monitoring and automate the communication process.In recent decades,researchers have made many efforts to propose autonomous systems for manipulating network data and providing on-time responses in critical operations.However,the widespread use of IoT devices in resource-constrained applications and mobile sensor networks introduces significant research challenges for cybersecurity.These systems are vulnerable to a variety of cyberattacks,including unauthorized access,denial-of-service attacks,and data leakage,which compromise the network’s security.Additionally,uneven load balancing between mobile IoT devices,which frequently experience link interferences,compromises the trustworthiness of the system.This paper introduces a Multi-Agent secured framework using lightweight edge computing to enhance cybersecurity for sensor networks,aiming to leverage artificial intelligence for adaptive routing and multi-metric trust evaluation to achieve data privacy and mitigate potential threats.Moreover,it enhances the efficiency of distributed sensors for energy consumption through intelligent data analytics techniques,resulting in highly consistent and low-latency network communication.Using simulations,the proposed framework reveals its significant performance compared to state-of-the-art approaches for energy consumption by 43%,latency by 46%,network throughput by 51%,packet loss rate by 40%,and denial of service attacks by 42%. 展开更多
关键词 Artificial intelligence CYBERSECURITY edge computing Internet of Things threat detection
在线阅读 下载PDF
A Comprehensive Survey on Blockchain-Enabled Techniques and Federated Learning for Secure 5G/6G Networks:Challenges,Opportunities,and Future Directions
8
作者 Muhammad Asim Abdelhamied A.Ateya +4 位作者 Mudasir Ahmad Wani Gauhar Ali Mohammed ElAffendi Ahmed A.Abd El-Latif Reshma Siyal 《Computers, Materials & Continua》 2026年第3期117-161,共45页
The growing developments in 5G and 6G wireless communications have revolutionized communications technologies,providing faster speeds with reduced latency and improved connectivity to users.However,it raises significa... The growing developments in 5G and 6G wireless communications have revolutionized communications technologies,providing faster speeds with reduced latency and improved connectivity to users.However,it raises significant security challenges,including impersonation threats,data manipulation,distributed denial of service(DDoS)attacks,and privacy breaches.Traditional security measures are inadequate due to the decentralized and dynamic nature of next-generation networks.This survey provides a comprehensive review of how Federated Learning(FL),Blockchain,and Digital Twin(DT)technologies can collectively enhance the security of 5G and 6G systems.Blockchain offers decentralized,immutable,and transparent mechanisms for securing network transactions,while FL enables privacy-preserving collaborative learning without sharing raw data.Digital Twins create virtual replicas of network components,enabling real-time monitoring,anomaly detection,and predictive threat analysis.The survey examines major security issues in emerging wireless architectures and analyzes recent advancements that integrate FL,Blockchain,and DT to mitigate these threats.Additionally,it presents practical use cases,synthesizes key lessons learned,and identifies ongoing research challenges.Finally,the survey outlines future research directions to support the development of scalable,intelligent,and robust security frameworks for next-generation wireless networks. 展开更多
关键词 5G/6G blockchain federated learning edge computing security
在线阅读 下载PDF
FAIR-DQL:Fairness-Aware Deep Q-Learning for Enhanced Resource Allocation and RIS Optimization in High-Altitude Platform Networks
9
作者 Muhammad Ejaz Muhammad Asim +1 位作者 Mudasir Ahmad Wani Kashish Ara Shakil 《Computers, Materials & Continua》 2026年第3期758-779,共22页
The integration of High-Altitude Platform Stations(HAPS)with Reconfigurable Intelligent Surfaces(RIS)represents a critical advancement for next-generation wireless networks,offering unprecedented opportunities for ubi... The integration of High-Altitude Platform Stations(HAPS)with Reconfigurable Intelligent Surfaces(RIS)represents a critical advancement for next-generation wireless networks,offering unprecedented opportunities for ubiquitous connectivity.However,existing research reveals significant gaps in dynamic resource allocation,joint optimization,and equitable service provisioning under varying channel conditions,limiting practical deployment of these technologies.This paper addresses these challenges by proposing a novel Fairness-Aware Deep Q-Learning(FAIRDQL)framework for joint resource management and phase configuration in HAPS-RIS systems.Our methodology employs a comprehensive three-tier algorithmic architecture integrating adaptive power control,priority-based user scheduling,and dynamic learning mechanisms.The FAIR-DQL approach utilizes advanced reinforcement learning with experience replay and fairness-aware reward functions to balance competing objectives while adapting to dynamic environments.Key findings demonstrate substantial improvements:9.15 dB SINR gain,12.5 bps/Hz capacity,78%power efficiency,and 0.82 fairness index.The framework achieves rapid 40-episode convergence with consistent delay performance.These contributions establish new benchmarks for fairness-aware resource allocation in aerial communications,enabling practical HAPS-RIS deployments in rural connectivity,emergency communications,and urban networks. 展开更多
关键词 Wireless communication high-altitude platform station reconfigurable intelligent surfaces deep Q-learning
在线阅读 下载PDF
HMA-DER:A Hierarchical Attention and Expert Routing Framework for Accurate Gastrointestinal Disease Diagnosis
10
作者 Sara Tehsin Inzamam Mashood Nasir +4 位作者 Wiem Abdelbaki Fadwa Alrowais Khalid A.Alattas Sultan Almutairi Radwa Marzouk 《Computers, Materials & Continua》 2026年第4期701-736,共36页
Objective:Deep learning is employed increasingly in Gastroenterology(GI)endoscopy computer-aided diagnostics for polyp segmentation and multi-class disease detection.In the real world,implementation requires high accu... Objective:Deep learning is employed increasingly in Gastroenterology(GI)endoscopy computer-aided diagnostics for polyp segmentation and multi-class disease detection.In the real world,implementation requires high accuracy,therapeutically relevant explanations,strong calibration,domain generalization,and efficiency.Current Convolutional Neural Network(CNN)and transformer models compromise border precision and global context,generate attention maps that fail to align with expert reasoning,deteriorate during cross-center changes,and exhibit inadequate calibration,hence diminishing clinical trust.Methods:HMA-DER is a hierarchical multi-attention architecture that uses dilation-enhanced residual blocks and an explainability-aware Cognitive Alignment Score(CAS)regularizer to directly align attribution maps with reasoning signals from experts.The framework has additions that make it more resilient and a way to test for accuracy,macro-averaged F1 score,Area Under the Receiver Operating Characteristic Curve(AUROC),calibration(Expected Calibration Error(ECE),Brier Score),explainability(CAS,insertion/deletion AUC),cross-dataset transfer,and throughput.Results:HMA-DER gets Dice Similarity Coefficient scores of 89.5%and 86.0%on Kvasir-SEG and CVC-ClinicDB,beating the strongest baseline by+1.9 and+1.7 points.It gets 86.4%and 85.3%macro-F1 and 94.0%and 93.4%AUROC on HyperKvasir and GastroVision,which is better than the baseline by+1.4/+1.6macro-F1 and+1.2/+1.1AUROC.Ablation study shows that hierarchical attention gives the highest(+3.0),followed by CAS regularization(+2–3),dilatation(+1.5–2.0),and residual connections(+2–3).Cross-dataset validation demonstrates competitive zero-shot transfer(e.g.,KS→CVC Dice 82.7%),whereas multi-dataset training diminishes the domain gap,yielding an 88.1%primary-metric average.HMA-DER’s mixed-precision inference can handle 155 pictures per second,which helps with calibration.Conclusion:HMA-DER strikes a compromise between accuracy,explainability,robustness,and efficiency for the use of reliable GI computer-aided diagnosis in real-world clinical settings. 展开更多
关键词 Gastrointestinal image analysis polyp segmentation multi-attention deep learning explainable AI cognitive alignment score cross-dataset generalization
在线阅读 下载PDF
A Ransomware Detection Approach Based on LLM Embedding and Ensemble Learning
11
作者 Abdallah Ghourabi Hassen Chouaib 《Computers, Materials & Continua》 2026年第4期2327-2342,共16页
In recent years,ransomware attacks have become one of the most common and destructive types of cyberattacks.Their impact is significant on the operations,finances and reputation of affected companies.Despite the effor... In recent years,ransomware attacks have become one of the most common and destructive types of cyberattacks.Their impact is significant on the operations,finances and reputation of affected companies.Despite the efforts of researchers and security experts to protect information systems from these attacks,the threat persists and the proposed solutions are not able to significantly stop the spread of ransomware attacks.The latest remarkable achievements of large language models(LLMs)in NLP tasks have caught the attention of cybersecurity researchers to integrate thesemodels into security threat detection.Thesemodels offer high embedding capabilities,able to extract rich semantic representations and paving theway formore accurate and adaptive solutions.In this context,we propose a new approach for ransomware detection based on an ensemblemethod that leverages three distinctLLMembeddingmodels.This ensemble strategy takes advantage of the variety of embedding methods and the strengths of each model.In the proposed solution,each embedding model is associated with an independently trainedMLP classifier.The predictions obtained are then merged using a weighted voting technique,assigning each model an influence proportional to its performance.This approach makes it possible to exploit the complementarity of representations,improve detection accuracy and robustness,and offer a more reliable solution in the face of the growing diversity and complexity of modern ransomware. 展开更多
关键词 Ransomware detection ensemble learning LLM embedding
在线阅读 下载PDF
Traffic Vision:UAV-Based Vehicle Detection and Traffic Pattern Analysis via Deep Learning Classifier
12
作者 Mohammed Alnusayri Ghulam Mujtaba +4 位作者 Nouf Abdullah Almujally Shuoa S.Aitarbi Asaad Algarni Ahmad Jalal Jeongmin Park 《Computers, Materials & Continua》 2026年第3期266-284,共19页
This paper presents a unified Unmanned Aerial Vehicle-based(UAV-based)traffic monitoring framework that integrates vehicle detection,tracking,counting,motion prediction,and classification in a modular and co-optimized... This paper presents a unified Unmanned Aerial Vehicle-based(UAV-based)traffic monitoring framework that integrates vehicle detection,tracking,counting,motion prediction,and classification in a modular and co-optimized pipeline.Unlike prior works that address these tasks in isolation,our approach combines You Only Look Once(YOLO)v10 detection,ByteTrack tracking,optical-flow density estimation,Long Short-Term Memory-based(LSTM-based)trajectory forecasting,and hybrid Speeded-Up Robust Feature(SURF)+Gray-Level Co-occurrence Matrix(GLCM)feature engineering with VGG16 classification.Upon the validation across datasets(UAVDT and UAVID)our framework achieved a detection accuracy of 94.2%,and 92.3%detection accuracy when conducting a real-time UAV field validation.Our comprehensive evaluations,including multi-metric analyses,ablation studies,and cross-dataset validations,confirm the framework’s accuracy,efficiency,and generalizability.These results highlight the novelty of integrating complementary methods into a single framework,offering a practical solution for accurate and efficient UAV-based traffic monitoring. 展开更多
关键词 Smart traffic system drone devices machine learner dynamic complex scenes VGG-16 classifier
在线阅读 下载PDF
A Transformer-Based Deep Learning Framework with Semantic Encoding and Syntax-Aware LSTM for Fake Electronic News Detection
13
作者 Hamza Murad Khan Shakila Basheer +3 位作者 Mohammad Tabrez Quasim Raja`a Al-Naimi Vijaykumar Varadarajan Anwar Khan 《Computers, Materials & Continua》 2026年第1期1024-1048,共25页
With the increasing growth of online news,fake electronic news detection has become one of the most important paradigms of modern research.Traditional electronic news detection techniques are generally based on contex... With the increasing growth of online news,fake electronic news detection has become one of the most important paradigms of modern research.Traditional electronic news detection techniques are generally based on contextual understanding,sequential dependencies,and/or data imbalance.This makes distinction between genuine and fabricated news a challenging task.To address this problem,we propose a novel hybrid architecture,T5-SA-LSTM,which synergistically integrates the T5 Transformer for semantically rich contextual embedding with the Self-Attentionenhanced(SA)Long Short-Term Memory(LSTM).The LSTM is trained using the Adam optimizer,which provides faster and more stable convergence compared to the Stochastic Gradient Descend(SGD)and Root Mean Square Propagation(RMSProp).The WELFake and FakeNewsPrediction datasets are used,which consist of labeled news articles having fake and real news samples.Tokenization and Synthetic Minority Over-sampling Technique(SMOTE)methods are used for data preprocessing to ensure linguistic normalization and class imbalance.The incorporation of the Self-Attention(SA)mechanism enables the model to highlight critical words and phrases,thereby enhancing predictive accuracy.The proposed model is evaluated using accuracy,precision,recall(sensitivity),and F1-score as performance metrics.The model achieved 99%accuracy on the WELFake dataset and 96.5%accuracy on the FakeNewsPrediction dataset.It outperformed the competitive schemes such as T5-SA-LSTM(RMSProp),T5-SA-LSTM(SGD)and some other models. 展开更多
关键词 Fake news detection tokenization SMOTE text-to-text transfer transformer(T5) long short-term memory(LSTM) self-attention mechanism(SA) T5-SA-LSTM WELFake dataset FakeNewsPrediction dataset
在线阅读 下载PDF
Advancing Android Ransomware Detection with Hybrid AutoML and Ensemble Learning Approaches
14
作者 Kirubavathi Ganapathiyappan Chahana Ravikumar +3 位作者 Raghul Alagunachimuthu Ranganayaki Ayman Altameem Ateeq Ur Rehman Ahmad Almogren 《Computers, Materials & Continua》 2026年第4期737-766,共30页
Android smartphones have become an integral part of our daily lives,becoming targets for ransomware attacks.Such attacks encrypt user information and ask for payment to recover it.Conventional detection mechanisms,suc... Android smartphones have become an integral part of our daily lives,becoming targets for ransomware attacks.Such attacks encrypt user information and ask for payment to recover it.Conventional detection mechanisms,such as signature-based and heuristic techniques,often fail to detect new and polymorphic ransomware samples.To address this challenge,we employed various ensemble classifiers,such as Random Forest,Gradient Boosting,Bagging,and AutoML models.We aimed to showcase how AutoML can automate processes such as model selection,feature engineering,and hyperparameter optimization,to minimize manual effort while ensuring or enhancing performance compared to traditional approaches.We used this framework to test it with a publicly available dataset from the Kaggle repository,which contains features for Android ransomware network traffic.The dataset comprises 392,024 flow records,divided into eleven groups.There are ten classes for various ransomware types,including SVpeng,PornDroid,Koler,WannaLocker,and Lockerpin.There is also a class for regular traffic.We applied a three-step procedure to select themost relevant features:filter,wrapper,and embeddedmethods.The Bagging classifier was highly accurate,correctly getting 99.84%of the time.The FLAML AutoML framework was evenmore accurate,correctly getting 99.85%of the time.This is indicative of howwellAutoML performs in improving things with minimal human assistance.Our findings indicate that AutoML is an efficient,scalable,and flexible method to discover Android ransomware,and it will facilitate the development of next-generation intrusion detection systems. 展开更多
关键词 Automated machine learning(AutoML) ensemble learning intrusion detection system(IDS) ransomware traffic analysis android ransomware detection
在线阅读 下载PDF
Scalable and Resilient AI Framework for Malware Detection in Software-Defined Internet of Things
15
作者 Maha Abdelhaq Ahmad Sami Al-Shamayleh +2 位作者 Adnan Akhunzada Nikola Ivkovi´c Toobah Hasan 《Computers, Materials & Continua》 2026年第4期1307-1321,共15页
The rapid expansion of the Internet of Things(IoT)and Edge Artificial Intelligence(AI)has redefined automation and connectivity acrossmodern networks.However,the heterogeneity and limited resources of IoT devices expo... The rapid expansion of the Internet of Things(IoT)and Edge Artificial Intelligence(AI)has redefined automation and connectivity acrossmodern networks.However,the heterogeneity and limited resources of IoT devices expose them to increasingly sophisticated and persistentmalware attacks.These adaptive and stealthy threats can evade conventional detection,establish remote control,propagate across devices,exfiltrate sensitive data,and compromise network integrity.This study presents a Software-Defined Internet of Things(SD-IoT)control-plane-based,AI-driven framework that integrates Gated Recurrent Units(GRU)and Long Short-TermMemory(LSTM)networks for efficient detection of evolving multi-vector,malware-driven botnet attacks.The proposed CUDA-enabled hybrid deep learning(DL)framework performs centralized real-time detection without adding computational overhead to IoT nodes.A feature selection strategy combining variable clustering,attribute evaluation,one-R attribute evaluation,correlation analysis,and principal component analysis(PCA)enhances detection accuracy and reduces complexity.The framework is rigorously evaluated using the N_BaIoT dataset under k-fold cross-validation.Experimental results achieve 99.96%detection accuracy,a false positive rate(FPR)of 0.0035%,and a detection latency of 0.18 ms,confirming its high efficiency and scalability.The findings demonstrate the framework’s potential as a robust and intelligent security solution for next-generation IoT ecosystems. 展开更多
关键词 AI-driven malware analysis advanced persistent malware(APM) AI-poweredmalware detection deep learning(DL) malware-driven botnets software-defined internet of things(SD-IoT)
在线阅读 下载PDF
Explainable Ensemble Learning Framework for Early Detection of Autism Spectrum Disorder:Enhancing Trust,Interpretability and Reliability in AI-Driven Healthcare
16
作者 Menwa Alshammeri Noshina Tariq +2 位作者 NZ Jhanji Mamoona Humayun Muhammad Attique Khan 《Computer Modeling in Engineering & Sciences》 2026年第1期1233-1265,共33页
Artificial Intelligence(AI)is changing healthcare by helping with diagnosis.However,for doctors to trust AI tools,they need to be both accurate and easy to understand.In this study,we created a new machine learning sy... Artificial Intelligence(AI)is changing healthcare by helping with diagnosis.However,for doctors to trust AI tools,they need to be both accurate and easy to understand.In this study,we created a new machine learning system for the early detection of Autism Spectrum Disorder(ASD)in children.Our main goal was to build a model that is not only good at predicting ASD but also clear in its reasoning.For this,we combined several different models,including Random Forest,XGBoost,and Neural Networks,into a single,more powerful framework.We used two different types of datasets:(i)a standard behavioral dataset and(ii)a more complex multimodal dataset with images,audio,and physiological information.The datasets were carefully preprocessed for missing values,redundant features,and dataset imbalance to ensure fair learning.The results outperformed the state-of-the-art with a Regularized Neural Network,achieving 97.6%accuracy on behavioral data.Whereas,on the multimodal data,the accuracy is 98.2%.Other models also did well with accuracies consistently above 96%.We also used SHAP and LIME on a behavioral dataset for models’explainability. 展开更多
关键词 Autism spectrum disorder(ASD) artificial intelligence in healthcare explainable AI(XAI) ensemble learning machine learning early diagnosis model interpretability SHAP LIME predictive analytics ethical AI healthcare trustworthiness
在线阅读 下载PDF
Concrete Strength Prediction Using Machine Learning and Somersaulting Spider Optimizer
17
作者 Marwa M.Eid Amel Ali Alhussan +2 位作者 Ebrahim A.Mattar Nima Khodadadi El-Sayed M.El-Kenawy 《Computer Modeling in Engineering & Sciences》 2026年第1期465-493,共29页
Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often f... Accurate prediction of concrete compressive strength is fundamental for optimizing mix designs,improving material utilization,and ensuring structural safety in modern construction.Traditional empirical methods often fail to capture the non-linear relationships among concrete constituents,especially with the growing use of supple-mentary cementitious materials and recycled aggregates.This study presents an integrated machine learning framework for concrete strength prediction,combining advanced regression models—namely CatBoost—with metaheuristic optimization algorithms,with a particular focus on the Somersaulting Spider Optimizer(SSO).A comprehensive dataset encompassing diverse mix proportions and material types was used to evaluate baseline machine learning models,including CatBoost,XGBoost,ExtraTrees,and RandomForest.Among these,CatBoost demonstrated superior accuracy across multiple performance metrics.To further enhance predictive capability,several bio-inspired optimizers were employed for hyperparameter tuning.The SSO-CatBoost hybrid achieved the lowest mean squared error and highest correlation coefficients,outperforming other metaheuristic approaches such as Genetic Algorithm,Particle Swarm Optimization,and Grey Wolf Optimizer.Statistical significance was established through Analysis of Variance and Wilcoxon signed-rank testing,confirming the robustness of the optimized models.The proposed methodology not only delivers improved predictive performance but also offers a transparent framework for mix design optimization,supporting data-driven decision making in sustainable and resilient infrastructure development. 展开更多
关键词 Concrete strength machine learning CatBoost metaheuristic optimization somersaulting spider optimizer ensemble models
在线阅读 下载PDF
A Comprehensive Literature Review of AI-Driven Application Mapping and Scheduling Techniques for Network-on-Chip Systems
18
作者 Naveed Ahmad Muhammad Kaleem +5 位作者 Mourad Elloumi Muhammad Azhar Mushtaq Ahlem Fatnassi Mohd Fazil Anas Bilal Abdulbasit A.Darem 《Computer Modeling in Engineering & Sciences》 2026年第1期118-155,共38页
Network-on-Chip(NoC)systems are progressively deployed in connecting massively parallel megacore systems in the new computing architecture.As a result,application mapping has become an important aspect of performance ... Network-on-Chip(NoC)systems are progressively deployed in connecting massively parallel megacore systems in the new computing architecture.As a result,application mapping has become an important aspect of performance and scalability,as current trends require the distribution of computation across network nodes/points.In this paper,we survey a large number of mapping and scheduling techniques designed for NoC architectures.This time,we concentrated on 3D systems.We take a systematic literature review approach to analyze existing methods across static,dynamic,hybrid,and machine-learning-based approaches,alongside preliminary AI-based dynamic models in recent works.We classify them into several main aspects covering power-aware mapping,fault tolerance,load-balancing,and adaptive for dynamic workloads.Also,we assess the efficacy of each method against performance parameters,such as latency,throughput,response time,and error rate.Key challenges,including energy efficiency,real-time adaptability,and reinforcement learning integration,are highlighted as well.To the best of our knowledge,this is one of the recent reviews that identifies both traditional and AI-based algorithms for mapping over a modern NoC,and opens research challenges.Finally,we provide directions for future work toward improved adaptability and scalability via lightweight learned models and hierarchical mapping frameworks. 展开更多
关键词 Application mapping mapping techniques NETWORK-ON-CHIP system on chip optimisation
在线阅读 下载PDF
Multi-Objective Enhanced Cheetah Optimizer for Joint Optimization of Computation Offloading and Task Scheduling in Fog Computing
19
作者 Ahmad Zia Nazia Azim +5 位作者 Bekarystankyzy Akbayan Khalid J.Alzahrani Ateeq Ur Rehman Faheem Ullah Khan Nouf Al-Kahtani Hend Khalid Alkahtani 《Computers, Materials & Continua》 2026年第3期1559-1588,共30页
The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous c... The cloud-fog computing paradigm has emerged as a novel hybrid computing model that integrates computational resources at both fog nodes and cloud servers to address the challenges posed by dynamic and heterogeneous computing networks.Finding an optimal computational resource for task offloading and then executing efficiently is a critical issue to achieve a trade-off between energy consumption and transmission delay.In this network,the task processed at fog nodes reduces transmission delay.Still,it increases energy consumption,while routing tasks to the cloud server saves energy at the cost of higher communication delay.Moreover,the order in which offloaded tasks are executed affects the system’s efficiency.For instance,executing lower-priority tasks before higher-priority jobs can disturb the reliability and stability of the system.Therefore,an efficient strategy of optimal computation offloading and task scheduling is required for operational efficacy.In this paper,we introduced a multi-objective and enhanced version of Cheeta Optimizer(CO),namely(MoECO),to jointly optimize the computation offloading and task scheduling in cloud-fog networks to minimize two competing objectives,i.e.,energy consumption and communication delay.MoECO first assigns tasks to the optimal computational nodes and then the allocated tasks are scheduled for processing based on the task priority.The mathematical modelling of CO needs improvement in computation time and convergence speed.Therefore,MoECO is proposed to increase the search capability of agents by controlling the search strategy based on a leader’s location.The adaptive step length operator is adjusted to diversify the solution and thus improves the exploration phase,i.e.,global search strategy.Consequently,this prevents the algorithm from getting trapped in the local optimal solution.Moreover,the interaction factor during the exploitation phase is also adjusted based on the location of the prey instead of the adjacent Cheetah.This increases the exploitation capability of agents,i.e.,local search capability.Furthermore,MoECO employs a multi-objective Pareto-optimal front to simultaneously minimize designated objectives.Comprehensive simulations in MATLAB demonstrate that the proposed algorithm obtains multiple solutions via a Pareto-optimal front and achieves an efficient trade-off between optimization objectives compared to baseline methods. 展开更多
关键词 Computation offloading task scheduling cheetah optimizer fog computing optimization resource allocation internet of things
在线阅读 下载PDF
Hybrid Quantum Gate Enabled CNN Framework with Optimized Features for Human-Object Detection and Recognition
20
作者 Nouf Abdullah Almujally Tanvir Fatima Naik Bukht +3 位作者 Shuaa S.Alharbi Asaad Algarni Ahmad Jalal Jeongmin Park 《Computers, Materials & Continua》 2026年第4期2254-2271,共18页
Recognising human-object interactions(HOI)is a challenging task for traditional machine learning models,including convolutional neural networks(CNNs).Existing models show limited transferability across complex dataset... Recognising human-object interactions(HOI)is a challenging task for traditional machine learning models,including convolutional neural networks(CNNs).Existing models show limited transferability across complex datasets such as D3D-HOI and SYSU 3D HOI.The conventional architecture of CNNs restricts their ability to handle HOI scenarios with high complexity.HOI recognition requires improved feature extraction methods to overcome the current limitations in accuracy and scalability.This work proposes a Novel quantum gate-enabled hybrid CNN(QEH-CNN)for effectiveHOI recognition.Themodel enhancesCNNperformance by integrating quantumcomputing components.The framework begins with bilateral image filtering,followed bymulti-object tracking(MOT)and Felzenszwalb superpixel segmentation.A watershed algorithm refines object boundaries by cleaning merged superpixels.Feature extraction combines a histogram of oriented gradients(HOG),Global Image Statistics for Texture(GIST)descriptors,and a novel 23-joint keypoint extractionmethod using relative joint angles and joint proximitymeasures.A fuzzy optimization process refines the extracted features before feeding them into the QEH-CNNmodel.The proposed model achieves 95.06%accuracy on the 3D-D3D-HOI dataset and 97.29%on the SYSU3DHOI dataset.Theintegration of quantum computing enhances feature optimization,leading to improved accuracy and overall model efficiency. 展开更多
关键词 Pattern recognition image segmentation computer vision object detection
在线阅读 下载PDF
上一页 1 2 66 下一页 到第
使用帮助 返回顶部