期刊文献+
共找到254,663篇文章
< 1 2 250 >
每页显示 20 50 100
Denoising graph neural network based on zero-shot learning for Gibbs phenomenon in high-order DG applications
1
作者 Wei AN Jiawen LIU +3 位作者 Wenxuan OUYANG Haoyu RU Xuejun LIU Hongqiang LYU 《Chinese Journal of Aeronautics》 2025年第3期234-248,共15页
With the availability of high-performance computing technology and the development of advanced numerical simulation methods, Computational Fluid Dynamics (CFD) is becoming more and more practical and efficient in engi... With the availability of high-performance computing technology and the development of advanced numerical simulation methods, Computational Fluid Dynamics (CFD) is becoming more and more practical and efficient in engineering. As one of the high-precision representative algorithms, the high-order Discontinuous Galerkin Method (DGM) has not only attracted widespread attention from scholars in the CFD research community, but also received strong development. However, when DGM is extended to high-speed aerodynamic flow field calculations, non-physical numerical Gibbs oscillations near shock waves often significantly affect the numerical accuracy and even cause calculation failure. Data driven approaches based on machine learning techniques can be used to learn the characteristics of Gibbs noise, which motivates us to use it in high-speed DG applications. To achieve this goal, labeled data need to be generated in order to train the machine learning models. This paper proposes a new method for denoising modeling of Gibbs phenomenon using a machine learning technique, the zero-shot learning strategy, to eliminate acquiring large amounts of CFD data. The model adopts a graph convolutional network combined with graph attention mechanism to learn the denoising paradigm from synthetic Gibbs noise data and generalize to DGM numerical simulation data. Numerical simulation results show that the Gibbs denoising model proposed in this paper can suppress the numerical oscillation near shock waves in the high-order DGM. Our work automates the extension of DGM to high-speed aerodynamic flow field calculations with higher generalization and lower cost. 展开更多
关键词 Computational fluid dynamics High-order discon tinuous Galerkin method Gibbs phenomenon Graph neural networks zero-shot learning
原文传递
A Dual Discriminator Method for Generalized Zero-Shot Learning
2
作者 Tianshu Wei Jinjie Huang 《Computers, Materials & Continua》 SCIE EI 2024年第4期1599-1612,共14页
Zero-shot learning enables the recognition of new class samples by migrating models learned from semanticfeatures and existing sample features to things that have never been seen before. The problems of consistencyof ... Zero-shot learning enables the recognition of new class samples by migrating models learned from semanticfeatures and existing sample features to things that have never been seen before. The problems of consistencyof different types of features and domain shift problems are two of the critical issues in zero-shot learning. Toaddress both of these issues, this paper proposes a new modeling structure. The traditional approach mappedsemantic features and visual features into the same feature space;based on this, a dual discriminator approachis used in the proposed model. This dual discriminator approach can further enhance the consistency betweensemantic and visual features. At the same time, this approach can also align unseen class semantic features andtraining set samples, providing a portion of information about the unseen classes. In addition, a new feature fusionmethod is proposed in the model. This method is equivalent to adding perturbation to the seen class features,which can reduce the degree to which the classification results in the model are biased towards the seen classes.At the same time, this feature fusion method can provide part of the information of the unseen classes, improvingits classification accuracy in generalized zero-shot learning and reducing domain bias. The proposed method isvalidated and compared with othermethods on four datasets, and fromthe experimental results, it can be seen thatthe method proposed in this paper achieves promising results. 展开更多
关键词 Generalized zero-shot learning modality consistent DISCRIMINATOR domain shift problem feature fusion
在线阅读 下载PDF
基于反向投影的zero-shot learning目标分类算法研究 被引量:1
3
作者 冯鹏 庹红娅 +2 位作者 乔凌峰 王洁欣 敬忠良 《计算机应用研究》 CSCD 北大核心 2017年第11期3291-3294,共4页
Zero-shot learning(ZSL)是针对没有训练样本的类别进行分类的问题。传统回归方法的核心是将视觉特征投影到语义空间,没有充分利用视觉特征自身包含的样本信息,同时训练计算量大。提出基于反向投影的ZSL目标分类方法,将类别原型投影到... Zero-shot learning(ZSL)是针对没有训练样本的类别进行分类的问题。传统回归方法的核心是将视觉特征投影到语义空间,没有充分利用视觉特征自身包含的样本信息,同时训练计算量大。提出基于反向投影的ZSL目标分类方法,将类别原型投影到视觉空间,利用视觉特征的语义性学习出映射函数,参数优化过程仅通过解析解就可以获得。在两个基准数据集的实验结果表明,提出的反向投影方法分类结果较传统回归方法和其他现有方法有大幅提升,并且训练时间大大减少,可以更好地推广到未知类别的分类问题上。 展开更多
关键词 zero-shot learning 目标分类 反向投影 解析解
在线阅读 下载PDF
A Simple Framework to Generalized Zero-Shot Learning for Fault Diagnosis of Industrial Processes 被引量:8
4
作者 Jiacheng Huang Zuxin Li Zhe Zhou 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2023年第6期1504-1506,共3页
Dear Editor, This letter provides a simple framework to generalized zero-shot learning for fault diagnosis. For industrial process monitoring, supervised learning and zero-shot learning(ZSL) can only deal with seen an... Dear Editor, This letter provides a simple framework to generalized zero-shot learning for fault diagnosis. For industrial process monitoring, supervised learning and zero-shot learning(ZSL) can only deal with seen and unseen faults, respectively. However, in the online monitoring stage of the actual industrial process, both seen and unseen faults may occur. This makes supervised learning and zero-shot learning impractical in industrial process monitoring. 展开更多
关键词 SEE LETTER learning
在线阅读 下载PDF
Learning from Scarcity:A Review of Deep Learning Strategies for Cold-Start Energy Time-Series Forecasting
5
作者 Jihoon Moon 《Computer Modeling in Engineering & Sciences》 2026年第1期26-76,共51页
Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-iti... Predicting the behavior of renewable energy systems requires models capable of generating accurate forecasts from limited historical data,a challenge that becomes especially pronounced when commissioning new facil-ities where operational records are scarce.This review aims to synthesize recent progress in data-efficient deep learning approaches for addressing such“cold-start”forecasting problems.It primarily covers three interrelated domains—solar photovoltaic(PV),wind power,and electrical load forecasting—where data scarcity and operational variability are most critical,while also including representative studies on hydropower and carbon emission prediction to provide a broader systems perspective.To this end,we examined trends from over 150 predominantly peer-reviewed studies published between 2019 and mid-2025,highlighting advances in zero-shot and few-shot meta-learning frameworks that enable rapid model adaptation with minimal labeled data.Moreover,transfer learning approaches combined with spatiotemporal graph neural networks have been employed to transfer knowledge from existing energy assets to new,data-sparse environments,effectively capturing hidden dependencies among geographic features,meteorological dynamics,and grid structures.Synthetic data generation has further proven valuable for expanding training samples and mitigating overfitting in cold-start scenarios.In addition,large language models and explainable artificial intelligence(XAI)—notably conversational XAI systems—have been used to interpret and communicate complex model behaviors in accessible terms,fostering operator trust from the earliest deployment stages.By consolidating methodological advances,unresolved challenges,and open-source resources,this review provides a coherent overview of deep learning strategies that can shorten the data-sparse ramp-up period of new energy infrastructures and accelerate the transition toward resilient,low-carbon electricity grids. 展开更多
关键词 Cold-start forecasting zero-shot learning few-shot meta-learning transfer learning spatiotemporal graph neural networks energy time series large language models explainable artificial intelligence(XAI)
在线阅读 下载PDF
GMCoT:a graph-augmented multimodal chain-of-thought reasoning framework for multi-label zero-shot learning
6
作者 Xiang WEN Haobo WANG +2 位作者 Ke CHEN Tianlei HU Gang CHEN 《Frontiers of Information Technology & Electronic Engineering》 2025年第12期2623-2637,共15页
In recent years,multi-label zero-shot learning(ML-ZSL)has garnered increasing attention because of its wide range of potential applications,such as image annotation,text classification,and bioinformatics.The central c... In recent years,multi-label zero-shot learning(ML-ZSL)has garnered increasing attention because of its wide range of potential applications,such as image annotation,text classification,and bioinformatics.The central challenge in ML-ZSL lies in predicting multiple labels for unseen classes without requiring any labeled training data,which contrasts with conventional supervised learning paradigms.However,existing methods face several significant challenges.These include the substantial semantic gap between different modalities,which impedes effective knowledge transfer,and the intricate and typically complex relationships among multiple labels,making it difficult to model them in a meaningful and accurate manner.To overcome these challenges,we propose a graph-augmented multimodal chain-of-thought(GMCoT)reasoning approach.The proposed method combines the strengths of multimodal large language models with graph-based structures,significantly enhancing the reasoning process involved in multi-label prediction.First,a novel multimodal chain-of-thought reasoning framework is presented which imitates human-like step-by-step reasoning to produce multi-label predictions.Second,a technique is presented for integrating label graphs into the reasoning process.This technique enables the capture of complex semantic relationships among labels,thereby improving the accuracy and consistency of multi-label generation.Comprehensive experiments on benchmark datasets demonstrate that the proposed GMCoT approach outperforms state-of-the-art methods in ML-ZSL. 展开更多
关键词 Chain-of-thought Multi-label zero-shot learning Multimodal reasoning Large language model
原文传递
BDA:Bi-directional attention for zero-shot learning
7
作者 Junseok Lee Jinming Cao +3 位作者 Yifang Yin Jihie Kim Roger Zimmermann Seongsik Park 《Computational Visual Media》 2025年第5期983-1003,共21页
Zero-shot learning(ZSL)is an important and rapidly growing area of machine learning that aims to recognize new classes without prior training data.Despite its significance,ZSL has faced challenges with overfitting in ... Zero-shot learning(ZSL)is an important and rapidly growing area of machine learning that aims to recognize new classes without prior training data.Despite its significance,ZSL has faced challenges with overfitting in embedding-based methods and limitations in traditional one-directional attention(ODA)based approaches.To bridge these gaps,this paper proposes the use of bi-directional attention(BDA)to integrate insights from both embedding and attention-based approaches.The proposed BDA system consists of a bi-directional attention network(BDAN)and a synthesized visual embedding network(SVEN)that facilitates visual-semantic interaction for ZSL classification.More specifically,the BDAN employs region self-attention(RSA),semantic synthesis attention(SSA),and visual synthesis attention(VSA)to overcome the overfitting issue in embedding methods and enhance transferability,to associate visual features with semantic property information,and to learn locally improved visual features.Extensive testing on CUB,SUN,and AWA2 datasets confirm the superiority of our proposed method over traditional approaches. 展开更多
关键词 zero-shot learning(ZSL) bi-directional attention(BDA) transferability INTERACTION
原文传递
PowerVLM:基于Federated Learning与模型剪枝的电力视觉语言大模型
8
作者 欧阳旭东 雒鹏鑫 +3 位作者 何绍洋 崔艺林 张中超 闫云凤 《全球能源互联网》 北大核心 2026年第1期101-111,共11页
智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learnin... 智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learning与模型剪枝的电力视觉语言大模型。提出了一种基于类别引导的电力视觉语言大模型PowerVLM,设计了类别引导增强模块,增强模型对电力图文数据的理解和问答能力;采用FL的强化学习训练策略,在满足数据隐私保护下,降低域间差异对模型性能的影响;最后,提出了一种基于信息决议的模型剪枝算法,可实现低训练参数的模型高效微调。分别在变电巡检、输电任务、作业安监3种典型电力场景开展实验,结果表明,该方法在电力场景多模态问答任务中的METEOR、BLEU和CIDEr等各项指标均表现优异,为电力场景智能感知提供了新的技术思路和方法支撑。 展开更多
关键词 智能电网 人工智能 视觉语言大模型 Federated learning 模型剪枝
在线阅读 下载PDF
Insights and analysis of machine learning for benzene hydrogenation to cyclohexene
9
作者 SUN Chao ZHANG Bin 《燃料化学学报(中英文)》 北大核心 2026年第2期133-139,共7页
Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face... Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research. 展开更多
关键词 machine learning heterogeneous catalysis hydrogenation of benzene XGBoost
在线阅读 下载PDF
Machine learning-based investigation of uplift resistance in special-shaped shield tunnels using numerical finite element modeling
10
作者 ZHANG Wengang YE Wenyu +2 位作者 SUN Weixin LIU Zhicheng LI Zhengchuan 《土木与环境工程学报(中英文)》 北大核心 2026年第1期1-13,共13页
The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi... The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance. 展开更多
关键词 special-shaped tunnel shield tunnel uplift resistance numerical simulation machine learning
在线阅读 下载PDF
A Survey of Federated Learning:Advances in Architecture,Synchronization,and Security Threats
11
作者 Faisal Mahmud Fahim Mahmud Rashedur M.Rahman 《Computers, Materials & Continua》 2026年第3期1-87,共87页
Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitiv... Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitive applications such as healthcare,finance,and smart systems.As the field continues to evolve,the research field has become more complex and scattered,covering different system designs,training methods,and privacy techniques.This survey is organized around the three core challenges:how the data is distributed,how models are synchronized,and how to defend against attacks.It provides a structured and up-to-date review of FL research from 2023 to 2025,offering a unified taxonomy that categorizes works by data distribution(Horizontal FL,Vertical FL,Federated Transfer Learning,and Personalized FL),training synchronization(synchronous and asynchronous FL),optimization strategies,and threat models(data leakage and poisoning attacks).In particular,we summarize the latest contributions in Vertical FL frameworks for secure multi-party learning,communication-efficient Horizontal FL,and domain-adaptive Federated Transfer Learning.Furthermore,we examine synchronization techniques addressing system heterogeneity,including straggler mitigation in synchronous FL and staleness management in asynchronous FL.The survey covers security threats in FL,such as gradient inversion,membership inference,and poisoning attacks,as well as their defense strategies that include privacy-preserving aggregation and anomaly detection.The paper concludes by outlining unresolved issues and highlighting challenges in handling personalized models,scalability,and real-world adoption. 展开更多
关键词 Federated learning(FL) horizontal federated learning(HFL) vertical federated learning(VFL) federated transfer learning(FTL) personalized federated learning synchronous federated learning(SFL) asynchronous federated learning(AFL) data leakage poisoning attacks privacy-preserving machine learning
在线阅读 下载PDF
Mitigating Attribute Inference in Split Learning via Channel Pruning and Adversarial Training
12
作者 Afnan Alhindi Saad Al-Ahmadi Mohamed Maher Ben Ismail 《Computers, Materials & Continua》 2026年第3期1465-1489,共25页
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn... Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%. 展开更多
关键词 Split learning privacy-preserving split learning distributed collaborative machine learning channel pruning adversarial learning resource-constrained devices
在线阅读 下载PDF
Task-Structured Curriculum Learning for Multi-Task Distillation:Enhancing Step-by-Step Knowledge Transfer in Language Models
13
作者 Ahmet Ezgi Aytug Onan 《Computers, Materials & Continua》 2026年第3期1647-1673,共27页
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re... Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning. 展开更多
关键词 Knowledge distillation curriculum learning language models multi-task learning step-by-step learning
在线阅读 下载PDF
Formulating an Innovative Gamified Personalized Learning Ecosystem Integrating 3D/VR Environments,Machine Learning,and Business Intelligence
14
作者 Nymfodora-Maria Raftopoulou Petros L.Pallis 《Sociology Study》 2026年第1期13-32,共20页
Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalize... Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalized Learning Ecosystem,which integrates 3D/VR environments,as well as machine learning algorithms,and business intelligence frameworks to enhance learner-centered education and inferenced decision-making.This Learning System makes use of immersive,analytically assessed virtual learning spaces,therefore facilitating real-time monitoring of not just learning performance,but also overall engagement and behavioral patterns,via a comprehensive set of sustainability-oriented ESG-aligned Key Performance Indicators(KPIs).Machine learning models support predictive analysis,personalized feedback,and hybrid recommendation mechanisms,whilst dedicated dashboards translate complex educational data into actionable insights for all Use Cases of the System(Educational Institutions,Educators and Learners).Additionally,the presented Learning System introduces a structured Mentoring and Consulting Subsystem,thence reinforcing human-centered guidance alongside automated intelligence.The Platform’s modular architecture and simulation-centered evaluation approach actively support personalized,and continuously optimized learning pathways.Thence,it exemplifies a mature,adaptive Learning Ecosystem,supporting immersive technologies,analytics,and pedagogical support,hence,contributing to contemporary digital learning innovation and sociotechnical transformation in education. 展开更多
关键词 gamified learning ecosystems learning analytics business intelligence personalized education virtual reality machine learning
在线阅读 下载PDF
A Cooperative Hybrid Learning Framework for Automated Dandruff Severity Grading
15
作者 Sin-Ye Jhong Hui-Che Hsu +3 位作者 Hsin-Hua Huang Chih-Hsien Hsia Yulius Harjoseputro Yung-Yao Chen 《Computers, Materials & Continua》 2026年第4期2272-2285,共14页
Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.S... Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.Standard classification methods fail to address these dual challenges,limiting their real-world performance.In this paper,a novel,three-phase training framework is proposed that learns a robust ordinal classifier directly from noisy labels.The approach synergistically combines a rank-based ordinal regression backbone with a cooperative,semi-supervised learning strategy to dynamically partition the data into clean and noisy subsets.A hybrid training objective is then employed,applying a supervised ordinal loss to the clean set.The noisy set is simultaneously trained using a dualobjective that combines a semi-supervised ordinal loss with a parallel,label-agnostic contrastive loss.This design allows themodel to learn fromthe entire noisy subset while using contrastive learning to mitigate the risk of error propagation frompotentially corrupt supervision.Extensive experiments on a new,large-scale,multi-site clinical dataset validate our approach.Themethod achieves state-of-the-art performance with 80.71%accuracy and a 76.86%F1-score,significantly outperforming existing approaches,including a 2.26%improvement over the strongest baseline method.This work provides not only a robust solution for a practical medical imaging problem but also a generalizable framework for other tasks plagued by noisy ordinal labels. 展开更多
关键词 Dandruff severity grading ordinal regression noisy label learning self-supervised learning contrastive learning medical image analysis
在线阅读 下载PDF
A State-of-the-Art Survey of Adversarial Reinforcement Learning for IoT Intrusion Detection
16
作者 Qasem Abu Al-Haija Shahad Al Tamimi 《Computers, Materials & Continua》 2026年第4期26-94,共69页
Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Tr... Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Training(AT)enables NIDS agents to discover and prevent newattack paths by exposing them to competing examples,thereby increasing detection accuracy,reducing False Positives(FPs),and enhancing network security.To develop robust decision-making capabilities for real-world network disruptions and hostile activity,NIDS agents are trained in adversarial scenarios to monitor the current state and notify management of any abnormal or malicious activity.The accuracy and timeliness of the IDS were crucial to the network’s availability and reliability at this time.This paper analyzes ARL applications in NIDS,revealing State-of-The-Art(SoTA)methodology,issues,and future research prospects.This includes Reinforcement Machine Learning(RML)-based NIDS,which enables an agent to interact with the environment to achieve a goal,andDeep Reinforcement Learning(DRL)-based NIDS,which can solve complex decision-making problems.Additionally,this survey study addresses cybersecurity adversarial circumstances and their importance for ARL and NIDS.Architectural design,RL algorithms,feature representation,and training methodologies are examined in the ARL-NIDS study.This comprehensive study evaluates ARL for intelligent NIDS research,benefiting cybersecurity researchers,practitioners,and policymakers.The report promotes cybersecurity defense research and innovation. 展开更多
关键词 Reinforcement learning network intrusion detection adversarial training deep learning cybersecurity defense intrusion detection system and machine learning
在线阅读 下载PDF
Machine Learning and Deep Learning for Smart Urban Transportation Systems with GPS,GIS,and Advanced Analytics:A Comprehensive Analysis
17
作者 E.Kalaivanan S.Brindha 《Journal of Harbin Institute of Technology(New Series)》 2026年第1期81-96,共16页
As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impact... As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impacting travel experiences and posing safety risks.Smart urban transportation management emerges as a strategic solution,conceptualized here as a multidimensional big data problem.The success of this strategy hinges on the effective collection of information from diverse,extensive,and heterogeneous data sources,necessitating the implementation of full⁃stack Information and Communication Technology(ICT)solutions.The main idea of the work is to investigate the current technologies of Intelligent Transportation Systems(ITS)and enhance the safety of urban transportation systems.Machine learning models,trained on historical data,can predict traffic congestion,allowing for the implementation of preventive measures.Deep learning architectures,with their ability to handle complex data representations,further refine traffic predictions,contributing to more accurate and dynamic transportation management.The background of this research underscores the challenges posed by traffic congestion in metropolitan areas and emphasizes the need for advanced technological solutions.By integrating GPS and GIS technologies with machine learning algorithms,this work aims to pay attention to the development of intelligent transportation systems that not only address current challenges but also pave the way for future advancements in urban transportation management. 展开更多
关键词 machine learning deep learning smart transportation
在线阅读 下载PDF
Predicting lymph node metastasis in colorectal cancer using caselevel multiple instance learning
18
作者 Ling-Feng Zou Xuan-Bing Wang +4 位作者 Jing-Wen Li Xin Ouyang Yi-Ying Luo Yan Luo Cheng-Long Wang 《World Journal of Gastroenterology》 2026年第1期110-125,共16页
BACKGROUND The accurate prediction of lymph node metastasis(LNM)is crucial for managing locally advanced(T3/T4)colorectal cancer(CRC).However,both traditional histopathology and standard slide-level deep learning ofte... BACKGROUND The accurate prediction of lymph node metastasis(LNM)is crucial for managing locally advanced(T3/T4)colorectal cancer(CRC).However,both traditional histopathology and standard slide-level deep learning often fail to capture the sparse and diagnostically critical features of metastatic potential.AIM To develop and validate a case-level multiple-instance learning(MIL)framework mimicking a pathologist's comprehensive review and improve T3/T4 CRC LNM prediction.METHODS The whole-slide images of 130 patients with T3/T4 CRC were retrospectively collected.A case-level MIL framework utilising the CONCH v1.5 and UNI2-h deep learning models was trained on features from all haematoxylin and eosinstained primary tumour slides for each patient.These pathological features were subsequently integrated with clinical data,and model performance was evaluated using the area under the curve(AUC).RESULTS The case-level framework demonstrated superior LNM prediction over slide-level training,with the CONCH v1.5 model achieving a mean AUC(±SD)of 0.899±0.033 vs 0.814±0.083,respectively.Integrating pathology features with clinical data further enhanced performance,yielding a top model with a mean AUC of 0.904±0.047,in sharp contrast to a clinical-only model(mean AUC 0.584±0.084).Crucially,a pathologist’s review confirmed that the model-identified high-attention regions correspond to known high-risk histopathological features.CONCLUSION A case-level MIL framework provides a superior approach for predicting LNM in advanced CRC.This method shows promise for risk stratification and therapy decisions,requiring further validation. 展开更多
关键词 Colorectal cancer Lymph node metastasis Deep learning Multiple instance learning HISTOPATHOLOGY
暂未订购
A comprehensive analysis of artificial intelligence,machine learning,deep learning and computer vision in food science
19
作者 Premkumar Borugadda Hemantha Kumar Kalluri 《Journal of Future Foods》 2026年第6期975-991,共17页
Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food proce... Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food processing from seed to harvest,storage,preparation,and consumption.This current paper seeks to demystify the importance of artificial intelligence,machine learning(ML),deep learning(DL),and computer vision(CV)in ensuring food safety and quality.By stressing the importance of these technologies,the audience will feel reassured and confident in their potential.These are very handy for such problems,giving assurance over food safety.CV is incredibly noble in today's generation because it improves food processing quality and positively impacts firms and researchers.Thus,at the present production stage,rich in image processing and computer visioning is incorporated into all facets of food production.In this field,DL and ML are implemented to identify the type of food in addition to quality.Concerning data and result-oriented perceptions,one has found similarities regarding various approaches.As a result,the findings of this study will be helpful for scholars looking for a proper approach to identify the quality of food offered.It helps to indicate which food products have been discussed by other scholars and lets the reader know papers by other scholars inclined to research further.Also,DL is accurately integrated with identifying the quality and safety of foods in the market.This paper describes the current practices and concerns of ML,DL,and probable trends for its future development. 展开更多
关键词 Artificial intelligence Computer vision Deep learning Food quality Food recognition Machine learning
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部