期刊文献+
共找到254,654篇文章
< 1 2 250 >
每页显示 20 50 100
Support Vector-Guided Class-Incremental Learning:Discriminative Replay with Dual-Alignment Distillation
1
作者 Moyi Zhang Yixin Wang Yu Cheng 《Computers, Materials & Continua》 2026年第3期2040-2061,共22页
Modern intelligent systems,such as autonomous vehicles and face recognition,must continuously adapt to new scenarios while preserving their ability to handle previously encountered situations.However,when neural netwo... Modern intelligent systems,such as autonomous vehicles and face recognition,must continuously adapt to new scenarios while preserving their ability to handle previously encountered situations.However,when neural networks learn new classes sequentially,they suffer from catastrophic forgetting—the tendency to lose knowledge of earlier classes.This challenge,which lies at the core of class-incremental learning,severely limits the deployment of continual learning systems in real-world applications with streaming data.Existing approaches,including rehearsalbased methods and knowledge distillation techniques,have attempted to address this issue but often struggle to effectively preserve decision boundaries and discriminative features under limited memory constraints.To overcome these limitations,we propose a support vector-guided framework for class-incremental learning.The framework integrates an enhanced feature extractor with a Support Vector Machine classifier,which generates boundary-critical support vectors to guide both replay and distillation.Building on this architecture,we design a joint feature retention strategy that combines boundary proximity with feature diversity,and a Support Vector Distillation Loss that enforces dual alignment in decision and semantic spaces.In addition,triple attention modules are incorporated into the feature extractor to enhance representation power.Extensive experiments on CIFAR-100 and Tiny-ImageNet demonstrate effective improvements.On CIFAR-100 and Tiny-ImageNet with 5 tasks,our method achieves 71.68%and 58.61%average accuracy,outperforming strong baselines by 3.34%and 2.05%.These advantages are consistently observed across different task splits,highlighting the robustness and generalization of the proposed approach.Beyond benchmark evaluations,the framework also shows potential in few-shot and resource-constrained applications such as edge computing and mobile robotics. 展开更多
关键词 class-incremental learning catastrophic forgetting support vector machine knowledge distillation
在线阅读 下载PDF
Balanced Contrast Class-Incremental Learning
2
作者 Shiqi Yu Luojun Lin Yuanlong Yu 《CAAI Transactions on Intelligence Technology》 2025年第6期1867-1879,共13页
Continual learning aims to empower a model to learn new tasks continuously while reducing forgetting to retain previously learnt knowledge.In the context of receiving streaming data that are not constrained by the ind... Continual learning aims to empower a model to learn new tasks continuously while reducing forgetting to retain previously learnt knowledge.In the context of receiving streaming data that are not constrained by the independent and identically distributed(IID)assumption,continual learning efficiently transforms and leverages previously learnt knowledge through various methodologies and completes the learning of new tasks.The generalisation performance and learning efficiency of the model are enhanced in a sequence of tasks.However,the class imbalance in continual learning scenarios critically undermines model performance.In particular,in the class-incremental scenario,the class imbalance results in a bias towards new task classes while degrading the performance on previous learnt classes,leading to catastrophic forgetting.In this paper,a novel method based on balanced contrast is proposed to solve the class-incremental continual learning.The method utilises gradient balancing to mitigate the impact of class imbalance in the class-incremental scenario.The method leverages contrastive learning and gradient modifications to facilitate balanced processing of data across different classes in continual learning.The method proposed in this paper surpasses the existing baseline approaches in the class-incremental learning scenario on standard image datasets such as CIFAR-100,CIFAR-10 and mini-ImageNet.The research results reveal that the proposed method effectively mitigates catastrophic forgetting of previously learnt classes,markedly improving the efficacy of continual learning and offering a powerful solution for further advancing continual learning performance. 展开更多
关键词 artificial intelligence deep learning deep neural networks image classification
在线阅读 下载PDF
Filter Bank Networks for Few-Shot Class-Incremental Learning
3
作者 Yanzhao Zhou Binghao Liu +1 位作者 Yiran Liu Jianbin Jiao 《Computer Modeling in Engineering & Sciences》 SCIE EI 2023年第10期647-668,共22页
Deep Convolution Neural Networks(DCNNs)can capture discriminative features from large datasets.However,how to incrementally learn new samples without forgetting old ones and recognize novel classes that arise in the d... Deep Convolution Neural Networks(DCNNs)can capture discriminative features from large datasets.However,how to incrementally learn new samples without forgetting old ones and recognize novel classes that arise in the dynamically changing world,e.g.,classifying newly discovered fish species,remains an open problem.We address an even more challenging and realistic setting of this problem where new class samples are insufficient,i.e.,Few-Shot Class-Incremental Learning(FSCIL).Current FSCIL methods augment the training data to alleviate the overfitting of novel classes.By contrast,we propose Filter Bank Networks(FBNs)that augment the learnable filters to capture fine-detailed features for adapting to future new classes.In the forward pass,FBNs augment each convolutional filter to a virtual filter bank containing the canonical one,i.e.,itself,and multiple transformed versions.During back-propagation,FBNs explicitly stimulate fine-detailed features to emerge and collectively align all gradients of each filter bank to learn the canonical one.FBNs capture pattern variants that do not yet exist in the pretraining session,thus making it easy to incorporate new classes in the incremental learning phase.Moreover,FBNs introduce model-level prior knowledge to efficiently utilize the limited few-shot data.Extensive experiments on MNIST,CIFAR100,CUB200,andMini-ImageNet datasets show that FBNs consistently outperformthe baseline by a significantmargin,reporting new state-of-the-art FSCIL results.In addition,we contribute a challenging FSCIL benchmark,Fishshot1K,which contains 8261 underwater images covering 1000 ocean fish species.The code is included in the supplementary materials. 展开更多
关键词 Deep learning incremental learning few-shot learning Filter Bank Networks
在线阅读 下载PDF
Federated Class-Incremental Learning with New-Class Augmented Self-Distillation
4
作者 Zhi-Yuan Wu Tian-Liu He +4 位作者 Sheng Sun Yu-Wei Wang Min Liu Bo Gao Xue-Feng Jiang 《Journal of Computer Science & Technology》 2025年第5期1427-1437,共11页
Federated learning(FL)enables collaborative model training among participants while guaranteeing the privacy of raw data.Mainstream FL methodologies overlook the dynamic nature of real-world data,particularly its tend... Federated learning(FL)enables collaborative model training among participants while guaranteeing the privacy of raw data.Mainstream FL methodologies overlook the dynamic nature of real-world data,particularly its tendency to grow in volume and diversify in classes over time.This oversight results in FL methods suffering from catastrophic forgetting,where the trained models inadvertently discard previously learned information upon assimilating new data.In response to this challenge,we propose a novel federated class-incremental learning(FCIL)method,named Federated Classincremental Learning with New-Class Augmented Self-Distillation(FedCLASS).The core of FedCLASS is to enrich the class scores of historical models with new class scores predicted by current models and utilize the combined knowledge for self-distillation,enabling a more sufficient and precise knowledge transfer from historical models to current models.Theoretical analyses demonstrate that FedCLASS stands on reliable foundations,considering the scores of old classes predicted by historical models as conditional probabilities in the absence of new classes,and the scores of new classes predicted by current models as the conditional probabilities of class scores derived from historical models.Empirical experiments demonstrate the superiority of FedCLASS over four baseline algorithms in reducing average forgetting rate and boosting global accuracy. 展开更多
关键词 federated learning class-incremental learning knowledge distillation
原文传递
Class-incremental open-set radio-frequency fingerprints identification based on prototypes extraction and self-attention transformation
5
作者 XIE Cunxiang ZHONG Zhaogen ZHANG Limin 《Journal of Systems Engineering and Electronics》 2026年第1期112-126,共15页
In wireless sensor networks,ensuring communication security via specific emitter identification(SEI)is crucial.However,existing SEI methods are limited to closed-set scenarios and lack the ability to detect unknown de... In wireless sensor networks,ensuring communication security via specific emitter identification(SEI)is crucial.However,existing SEI methods are limited to closed-set scenarios and lack the ability to detect unknown devices and perform classincremental training.This study proposes a class-incremental open-set SEI approach.The open-set SEI model calculates radiofrequency fingerprints(RFFs)prototypes for known signals and employs a self-attention mechanism to enhance their discriminability.Detection thresholds are set through Gaussian fitting for each class.For class-incremental learning,the algorithm freezes the parameters of the previously trained model to initialize the new model.It designs specific losses:the RFFs extraction distribution difference loss and the prototype transformation distribution difference loss,which force the new model to retain old knowledge while learning new knowledge.The training loss enables learning of new class RFFs.Experimental results demonstrate that the open-set SEI model achieves state-of-theart performance and strong noise robustness.Moreover,the class-incremental learning algorithm effectively enables the model to retain old device RFFs knowledge,acquire new device RFFs knowledge,and detect unknown devices simultaneously. 展开更多
关键词 wireless sensor network specific emitter identification open-set identification class-incremental learning
在线阅读 下载PDF
Memory replay with unlabeled data for semi-supervised class-incremental learning via temporal consistency
6
作者 Qiang WANG Kele XU +2 位作者 Dawei FENG Bo DING Huaimin WANG 《Frontiers of Computer Science》 2025年第12期183-186,共4页
1 Introduction Current continual learning methods[1–4]can utilize labeled data to alleviate catastrophic forgetting effectively.However,obtaining labeled samples can be difficult and tedious as it may require expert ... 1 Introduction Current continual learning methods[1–4]can utilize labeled data to alleviate catastrophic forgetting effectively.However,obtaining labeled samples can be difficult and tedious as it may require expert knowledge.In many practical application scenarios,labeled and unlabeled samples exist simultaneously,with more unlabeled than labeled samples in streaming data[5,6].Unfortunately,existing class-incremental learning methods face limitations in effectively utilizing unlabeled data,thereby impeding their performance in incremental learning scenarios. 展开更多
关键词 alleviate catastrophic forgetting semi supervised learning temporal consistency memory replay labeled data streaming data unfortunatelyexisting unlabeled data labeled samples
原文传递
Squeezing More Past Knowledge for Online Class-Incremental Continual Learning 被引量:1
7
作者 Da Yu Mingyi Zhang +4 位作者 Mantian Li Fusheng Zha Junge Zhang Lining Sun Kaiqi Huang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2023年第3期722-736,共15页
Continual learning(CL)studies the problem of learning to accumulate knowledge over time from a stream of data.A crucial challenge is that neural networks suffer from performance degradation on previously seen data,kno... Continual learning(CL)studies the problem of learning to accumulate knowledge over time from a stream of data.A crucial challenge is that neural networks suffer from performance degradation on previously seen data,known as catastrophic forgetting,due to allowing parameter sharing.In this work,we consider a more practical online class-incremental CL setting,where the model learns new samples in an online manner and may continuously experience new classes.Moreover,prior knowledge is unavailable during training and evaluation.Existing works usually explore sample usages from a single dimension,which ignores a lot of valuable supervisory information.To better tackle the setting,we propose a novel replay-based CL method,which leverages multi-level representations produced by the intermediate process of training samples for replay and strengthens supervision to consolidate previous knowledge.Specifically,besides the previous raw samples,we store the corresponding logits and features in the memory.Furthermore,to imitate the prediction of the past model,we construct extra constraints by leveraging multi-level information stored in the memory.With the same number of samples for replay,our method can use more past knowledge to prevent interference.We conduct extensive evaluations on several popular CL datasets,and experiments show that our method consistently outperforms state-of-the-art methods with various sizes of episodic memory.We further provide a detailed analysis of these results and demonstrate that our method is more viable in practical scenarios. 展开更多
关键词 Catastrophic forgetting class-incremental learning continual learning(CL) experience replay
在线阅读 下载PDF
A memory-friendly class-incremental learning method for hand gesture recognition using HD-sEMG
8
作者 Yu Bai Le Wu +1 位作者 Shengcai Duan Xun Chen 《Medicine in Novel Technology and Devices》 2024年第2期124-132,共9页
Hand gesture recognition(HGR)plays a vital role in human-computer interaction.The integration of high-density surface electromyography(HD-sEMG)and deep neural networks(DNNs)has significantly improved the robustness an... Hand gesture recognition(HGR)plays a vital role in human-computer interaction.The integration of high-density surface electromyography(HD-sEMG)and deep neural networks(DNNs)has significantly improved the robustness and accuracy of HGR systems.These methods are typically effective for a fixed set of trained gestures.However,the need for new gesture classes over time poses a challenge.Introducing new classes to DNNs can lead to a substantial decrease in accuracy for previously learned tasks,a phenomenon known as“catastrophic forgetting,”especially when the training data for earlier tasks is not retained and retrained.This issue is exacerbated in embedded devices with limited storage,which struggle to store the large-scale data of HD-sEMG.Classincremental learning(CIL)is an effective method to reduce catastrophic forgetting.However,existing CIL methods for HGR rarely focus on reducing memory load.To address this,we propose a memory-friendly CIL method for HGR using HD-sEMG.Our approach includes a lightweight convolutional neural network,named SeparaNet,for feature representation learning,coupled with a nearest-mean-of-exemplars classifier for classifi-cation.We introduce a priority exemplar selection algorithm inspired by the herding effect to maintain a manageable set of exemplars during training.Furthermore,a task-equal-weight exemplar sampling strategy is proposed to effectively reduce memory load while preserving high recognition performance.Experimental results on two datasets demonstrate that our method significantly reduces the number of retained exemplars to only a quarter of that required by other CIL methods,accounting for less than 5%of the total samples,while still achieving comparable average accuracy. 展开更多
关键词 Myoelectric pattern recognition Memory-friendly class-incremental learning
在线阅读 下载PDF
PowerVLM:基于Federated Learning与模型剪枝的电力视觉语言大模型
9
作者 欧阳旭东 雒鹏鑫 +3 位作者 何绍洋 崔艺林 张中超 闫云凤 《全球能源互联网》 北大核心 2026年第1期101-111,共11页
智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learnin... 智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learning与模型剪枝的电力视觉语言大模型。提出了一种基于类别引导的电力视觉语言大模型PowerVLM,设计了类别引导增强模块,增强模型对电力图文数据的理解和问答能力;采用FL的强化学习训练策略,在满足数据隐私保护下,降低域间差异对模型性能的影响;最后,提出了一种基于信息决议的模型剪枝算法,可实现低训练参数的模型高效微调。分别在变电巡检、输电任务、作业安监3种典型电力场景开展实验,结果表明,该方法在电力场景多模态问答任务中的METEOR、BLEU和CIDEr等各项指标均表现优异,为电力场景智能感知提供了新的技术思路和方法支撑。 展开更多
关键词 智能电网 人工智能 视觉语言大模型 Federated learning 模型剪枝
在线阅读 下载PDF
Insights and analysis of machine learning for benzene hydrogenation to cyclohexene
10
作者 SUN Chao ZHANG Bin 《燃料化学学报(中英文)》 北大核心 2026年第2期133-139,共7页
Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face... Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research. 展开更多
关键词 machine learning heterogeneous catalysis hydrogenation of benzene XGBoost
在线阅读 下载PDF
Machine learning-based investigation of uplift resistance in special-shaped shield tunnels using numerical finite element modeling
11
作者 ZHANG Wengang YE Wenyu +2 位作者 SUN Weixin LIU Zhicheng LI Zhengchuan 《土木与环境工程学报(中英文)》 北大核心 2026年第1期1-13,共13页
The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi... The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance. 展开更多
关键词 special-shaped tunnel shield tunnel uplift resistance numerical simulation machine learning
在线阅读 下载PDF
A Survey of Federated Learning:Advances in Architecture,Synchronization,and Security Threats
12
作者 Faisal Mahmud Fahim Mahmud Rashedur M.Rahman 《Computers, Materials & Continua》 2026年第3期1-87,共87页
Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitiv... Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitive applications such as healthcare,finance,and smart systems.As the field continues to evolve,the research field has become more complex and scattered,covering different system designs,training methods,and privacy techniques.This survey is organized around the three core challenges:how the data is distributed,how models are synchronized,and how to defend against attacks.It provides a structured and up-to-date review of FL research from 2023 to 2025,offering a unified taxonomy that categorizes works by data distribution(Horizontal FL,Vertical FL,Federated Transfer Learning,and Personalized FL),training synchronization(synchronous and asynchronous FL),optimization strategies,and threat models(data leakage and poisoning attacks).In particular,we summarize the latest contributions in Vertical FL frameworks for secure multi-party learning,communication-efficient Horizontal FL,and domain-adaptive Federated Transfer Learning.Furthermore,we examine synchronization techniques addressing system heterogeneity,including straggler mitigation in synchronous FL and staleness management in asynchronous FL.The survey covers security threats in FL,such as gradient inversion,membership inference,and poisoning attacks,as well as their defense strategies that include privacy-preserving aggregation and anomaly detection.The paper concludes by outlining unresolved issues and highlighting challenges in handling personalized models,scalability,and real-world adoption. 展开更多
关键词 Federated learning(FL) horizontal federated learning(HFL) vertical federated learning(VFL) federated transfer learning(FTL) personalized federated learning synchronous federated learning(SFL) asynchronous federated learning(AFL) data leakage poisoning attacks privacy-preserving machine learning
在线阅读 下载PDF
Mitigating Attribute Inference in Split Learning via Channel Pruning and Adversarial Training
13
作者 Afnan Alhindi Saad Al-Ahmadi Mohamed Maher Ben Ismail 《Computers, Materials & Continua》 2026年第3期1465-1489,共25页
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn... Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%. 展开更多
关键词 Split learning privacy-preserving split learning distributed collaborative machine learning channel pruning adversarial learning resource-constrained devices
在线阅读 下载PDF
Task-Structured Curriculum Learning for Multi-Task Distillation:Enhancing Step-by-Step Knowledge Transfer in Language Models
14
作者 Ahmet Ezgi Aytug Onan 《Computers, Materials & Continua》 2026年第3期1647-1673,共27页
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re... Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning. 展开更多
关键词 Knowledge distillation curriculum learning language models multi-task learning step-by-step learning
在线阅读 下载PDF
Formulating an Innovative Gamified Personalized Learning Ecosystem Integrating 3D/VR Environments,Machine Learning,and Business Intelligence
15
作者 Nymfodora-Maria Raftopoulou Petros L.Pallis 《Sociology Study》 2026年第1期13-32,共20页
Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalize... Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalized Learning Ecosystem,which integrates 3D/VR environments,as well as machine learning algorithms,and business intelligence frameworks to enhance learner-centered education and inferenced decision-making.This Learning System makes use of immersive,analytically assessed virtual learning spaces,therefore facilitating real-time monitoring of not just learning performance,but also overall engagement and behavioral patterns,via a comprehensive set of sustainability-oriented ESG-aligned Key Performance Indicators(KPIs).Machine learning models support predictive analysis,personalized feedback,and hybrid recommendation mechanisms,whilst dedicated dashboards translate complex educational data into actionable insights for all Use Cases of the System(Educational Institutions,Educators and Learners).Additionally,the presented Learning System introduces a structured Mentoring and Consulting Subsystem,thence reinforcing human-centered guidance alongside automated intelligence.The Platform’s modular architecture and simulation-centered evaluation approach actively support personalized,and continuously optimized learning pathways.Thence,it exemplifies a mature,adaptive Learning Ecosystem,supporting immersive technologies,analytics,and pedagogical support,hence,contributing to contemporary digital learning innovation and sociotechnical transformation in education. 展开更多
关键词 gamified learning ecosystems learning analytics business intelligence personalized education virtual reality machine learning
在线阅读 下载PDF
A Cooperative Hybrid Learning Framework for Automated Dandruff Severity Grading
16
作者 Sin-Ye Jhong Hui-Che Hsu +3 位作者 Hsin-Hua Huang Chih-Hsien Hsia Yulius Harjoseputro Yung-Yao Chen 《Computers, Materials & Continua》 2026年第4期2272-2285,共14页
Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.S... Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.Standard classification methods fail to address these dual challenges,limiting their real-world performance.In this paper,a novel,three-phase training framework is proposed that learns a robust ordinal classifier directly from noisy labels.The approach synergistically combines a rank-based ordinal regression backbone with a cooperative,semi-supervised learning strategy to dynamically partition the data into clean and noisy subsets.A hybrid training objective is then employed,applying a supervised ordinal loss to the clean set.The noisy set is simultaneously trained using a dualobjective that combines a semi-supervised ordinal loss with a parallel,label-agnostic contrastive loss.This design allows themodel to learn fromthe entire noisy subset while using contrastive learning to mitigate the risk of error propagation frompotentially corrupt supervision.Extensive experiments on a new,large-scale,multi-site clinical dataset validate our approach.Themethod achieves state-of-the-art performance with 80.71%accuracy and a 76.86%F1-score,significantly outperforming existing approaches,including a 2.26%improvement over the strongest baseline method.This work provides not only a robust solution for a practical medical imaging problem but also a generalizable framework for other tasks plagued by noisy ordinal labels. 展开更多
关键词 Dandruff severity grading ordinal regression noisy label learning self-supervised learning contrastive learning medical image analysis
在线阅读 下载PDF
A State-of-the-Art Survey of Adversarial Reinforcement Learning for IoT Intrusion Detection
17
作者 Qasem Abu Al-Haija Shahad Al Tamimi 《Computers, Materials & Continua》 2026年第4期26-94,共69页
Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Tr... Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Training(AT)enables NIDS agents to discover and prevent newattack paths by exposing them to competing examples,thereby increasing detection accuracy,reducing False Positives(FPs),and enhancing network security.To develop robust decision-making capabilities for real-world network disruptions and hostile activity,NIDS agents are trained in adversarial scenarios to monitor the current state and notify management of any abnormal or malicious activity.The accuracy and timeliness of the IDS were crucial to the network’s availability and reliability at this time.This paper analyzes ARL applications in NIDS,revealing State-of-The-Art(SoTA)methodology,issues,and future research prospects.This includes Reinforcement Machine Learning(RML)-based NIDS,which enables an agent to interact with the environment to achieve a goal,andDeep Reinforcement Learning(DRL)-based NIDS,which can solve complex decision-making problems.Additionally,this survey study addresses cybersecurity adversarial circumstances and their importance for ARL and NIDS.Architectural design,RL algorithms,feature representation,and training methodologies are examined in the ARL-NIDS study.This comprehensive study evaluates ARL for intelligent NIDS research,benefiting cybersecurity researchers,practitioners,and policymakers.The report promotes cybersecurity defense research and innovation. 展开更多
关键词 Reinforcement learning network intrusion detection adversarial training deep learning cybersecurity defense intrusion detection system and machine learning
在线阅读 下载PDF
Machine Learning and Deep Learning for Smart Urban Transportation Systems with GPS,GIS,and Advanced Analytics:A Comprehensive Analysis
18
作者 E.Kalaivanan S.Brindha 《Journal of Harbin Institute of Technology(New Series)》 2026年第1期81-96,共16页
As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impact... As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impacting travel experiences and posing safety risks.Smart urban transportation management emerges as a strategic solution,conceptualized here as a multidimensional big data problem.The success of this strategy hinges on the effective collection of information from diverse,extensive,and heterogeneous data sources,necessitating the implementation of full⁃stack Information and Communication Technology(ICT)solutions.The main idea of the work is to investigate the current technologies of Intelligent Transportation Systems(ITS)and enhance the safety of urban transportation systems.Machine learning models,trained on historical data,can predict traffic congestion,allowing for the implementation of preventive measures.Deep learning architectures,with their ability to handle complex data representations,further refine traffic predictions,contributing to more accurate and dynamic transportation management.The background of this research underscores the challenges posed by traffic congestion in metropolitan areas and emphasizes the need for advanced technological solutions.By integrating GPS and GIS technologies with machine learning algorithms,this work aims to pay attention to the development of intelligent transportation systems that not only address current challenges but also pave the way for future advancements in urban transportation management. 展开更多
关键词 machine learning deep learning smart transportation
在线阅读 下载PDF
Predicting lymph node metastasis in colorectal cancer using caselevel multiple instance learning
19
作者 Ling-Feng Zou Xuan-Bing Wang +4 位作者 Jing-Wen Li Xin Ouyang Yi-Ying Luo Yan Luo Cheng-Long Wang 《World Journal of Gastroenterology》 2026年第1期110-125,共16页
BACKGROUND The accurate prediction of lymph node metastasis(LNM)is crucial for managing locally advanced(T3/T4)colorectal cancer(CRC).However,both traditional histopathology and standard slide-level deep learning ofte... BACKGROUND The accurate prediction of lymph node metastasis(LNM)is crucial for managing locally advanced(T3/T4)colorectal cancer(CRC).However,both traditional histopathology and standard slide-level deep learning often fail to capture the sparse and diagnostically critical features of metastatic potential.AIM To develop and validate a case-level multiple-instance learning(MIL)framework mimicking a pathologist's comprehensive review and improve T3/T4 CRC LNM prediction.METHODS The whole-slide images of 130 patients with T3/T4 CRC were retrospectively collected.A case-level MIL framework utilising the CONCH v1.5 and UNI2-h deep learning models was trained on features from all haematoxylin and eosinstained primary tumour slides for each patient.These pathological features were subsequently integrated with clinical data,and model performance was evaluated using the area under the curve(AUC).RESULTS The case-level framework demonstrated superior LNM prediction over slide-level training,with the CONCH v1.5 model achieving a mean AUC(±SD)of 0.899±0.033 vs 0.814±0.083,respectively.Integrating pathology features with clinical data further enhanced performance,yielding a top model with a mean AUC of 0.904±0.047,in sharp contrast to a clinical-only model(mean AUC 0.584±0.084).Crucially,a pathologist’s review confirmed that the model-identified high-attention regions correspond to known high-risk histopathological features.CONCLUSION A case-level MIL framework provides a superior approach for predicting LNM in advanced CRC.This method shows promise for risk stratification and therapy decisions,requiring further validation. 展开更多
关键词 Colorectal cancer Lymph node metastasis Deep learning Multiple instance learning HISTOPATHOLOGY
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部