期刊文献+
共找到19,384篇文章
< 1 2 250 >
每页显示 20 50 100
Precision organoid segmentation technique(POST):accurate organoid segmentation in challenging bright-field images 被引量:1
1
作者 Xuan Du Yuchen Li +5 位作者 Jiaping Song Zilin Zhang Jing Zhang Yanhui Li Zaozao Chen Zhongze Gu 《Bio-Design and Manufacturing》 2026年第1期80-93,I0013-I0016,共18页
Organoids possess immense potential for unraveling the intricate functions of human tissues and facilitating preclinical disease treatment.Their applications span from high-throughput drug screening to the modeling of... Organoids possess immense potential for unraveling the intricate functions of human tissues and facilitating preclinical disease treatment.Their applications span from high-throughput drug screening to the modeling of complex diseases,with some even achieving clinical translation.Changes in the overall size,shape,boundary,and other morphological features of organoids provide a noninvasive method for assessing organoid drug sensitivity.However,the precise segmentation of organoids in bright-field microscopy images is made difficult by the complexity of the organoid morphology and interference,including overlapping organoids,bubbles,dust particles,and cell fragments.This paper introduces the precision organoid segmentation technique(POST),which is a deep-learning algorithm for segmenting challenging organoids under simple bright-field imaging conditions.Unlike existing methods,POST accurately segments each organoid and eliminates various artifacts encountered during organoid culturing and imaging.Furthermore,it is sensitive to and aligns with measurements of organoid activity in drug sensitivity experiments.POST is expected to be a valuable tool for drug screening using organoids owing to its capability of automatically and rapidly eliminating interfering substances and thereby streamlining the organoid analysis and drug screening process. 展开更多
关键词 Organoid Drug screening Deep learning Image segmentation
暂未订购
Global-local feature optimization based RGB-IR fusion object detection on drone view 被引量:1
2
作者 Zhaodong CHEN Hongbing JI Yongquan ZHANG 《Chinese Journal of Aeronautics》 2026年第1期436-453,共18页
Visible and infrared(RGB-IR)fusion object detection plays an important role in security,disaster relief,etc.In recent years,deep-learning-based RGB-IR fusion detection methods have been developing rapidly,but still st... Visible and infrared(RGB-IR)fusion object detection plays an important role in security,disaster relief,etc.In recent years,deep-learning-based RGB-IR fusion detection methods have been developing rapidly,but still struggle to deal with the complex and changing scenarios captured by drones,mainly due to two reasons:(A)RGB-IR fusion detectors are susceptible to inferior inputs that degrade performance and stability.(B)RGB-IR fusion detectors are susceptible to redundant features that reduce accuracy and efficiency.In this paper,an innovative RGB-IR fusion detection framework based on global-local feature optimization,named GLFDet,is proposed to improve the detection performance and efficiency of drone-captured objects.The key components of GLFDet include a Global Feature Optimization(GFO)module,a Local Feature Optimization(LFO)module and a Channel Separation Fusion(CSF)module.Specifically,GFO calculates the information content of the input image from the frequency domain and optimizes the features holistically.Then,LFO dynamically selects high-value features and filters out low-value features before fusion,which significantly improves the efficiency of fusion.Finally,CSF fuses the RGB and IR features across the corresponding channels,which avoids the rearrangement of the channel relationships and enhances the model stability.Extensive experimental results show that the proposed method achieves the best performance on three popular RGB-IR datasets Drone Vehicle,VEDAI,and LLVIP.In addition,GLFDet is more lightweight than other comparable models,making it more appealing to edge devices such as drones.The code is available at https://github.com/lao chen330/GLFDet. 展开更多
关键词 Object detection Deep learning RGB-IR fusion DRONES Global feature Local feature
原文传递
Insights and analysis of machine learning for benzene hydrogenation to cyclohexene
3
作者 SUN Chao ZHANG Bin 《燃料化学学报(中英文)》 北大核心 2026年第2期133-139,共7页
Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face... Cyclohexene is an important raw material in the production of nylon.Selective hydrogenation of benzene is a key method for preparing cyclohexene.However,the Ru catalysts used in current industrial processes still face challenges,including high metal usage,high process costs,and low cyclohexene yield.This study utilizes existing literature data combined with machine learning methods to analyze the factors influencing benzene conversion,cyclohexene selectivity,and yield in the benzene hydrogenation to cyclohexene reaction.It constructs predictive models based on XGBoost and Random Forest algorithms.After analysis,it was found that reaction time,Ru content,and space velocity are key factors influencing cyclohexene yield,selectivity,and benzene conversion.Shapley Additive Explanations(SHAP)analysis and feature importance analysis further revealed the contribution of each variable to the reaction outcomes.Additionally,we randomly generated one million variable combinations using the Dirichlet distribution to attempt to predict high-yield catalyst formulations.This paper provides new insights into the application of machine learning in heterogeneous catalysis and offers some reference for further research. 展开更多
关键词 machine learning heterogeneous catalysis hydrogenation of benzene XGBoost
在线阅读 下载PDF
Machine learning-based investigation of uplift resistance in special-shaped shield tunnels using numerical finite element modeling
4
作者 ZHANG Wengang YE Wenyu +2 位作者 SUN Weixin LIU Zhicheng LI Zhengchuan 《土木与环境工程学报(中英文)》 北大核心 2026年第1期1-13,共13页
The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi... The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance. 展开更多
关键词 special-shaped tunnel shield tunnel uplift resistance numerical simulation machine learning
在线阅读 下载PDF
The Analysis of Gauss Radial Basis Functions and Its Application in Locating Olivine on the Moon
5
作者 SONG Shicang SONG Xiaoyuan SONG Shuhan 《应用数学》 北大核心 2026年第1期173-181,共9页
Gauss radial basis functions(GRBF)are frequently employed in data fitting and machine learning.Their linear independence property can theoretically guarantee the avoidance of data redundancy.In this paper,one of the m... Gauss radial basis functions(GRBF)are frequently employed in data fitting and machine learning.Their linear independence property can theoretically guarantee the avoidance of data redundancy.In this paper,one of the main contributions is proving this property using linear algebra instead of profound knowledge.This makes it easy to read and understand this fundamental fact.The proof of linear independence of a set of Gauss functions relies on the constructing method for one-dimensional space and on the deducing method for higher dimensions.Additionally,under the condition of preserving the same moments between the original function and interpolating function,both the interpolating existence and uniqueness are proven for GRBF in one-dimensional space.The final work demonstrates the application of the GRBF method to locate lunar olivine.By combining preprocessed data using GRBF with the removing envelope curve method,a program is created to find the position of lunar olivine based on spectrum data,and the numerical experiment shows that it is an effective scheme. 展开更多
关键词 Gauss function Radial basis function Machine learning Lunar olivine locating Data fitting
在线阅读 下载PDF
PowerVLM:基于Federated Learning与模型剪枝的电力视觉语言大模型
6
作者 欧阳旭东 雒鹏鑫 +3 位作者 何绍洋 崔艺林 张中超 闫云凤 《全球能源互联网》 北大核心 2026年第1期101-111,共11页
智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learnin... 智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learning与模型剪枝的电力视觉语言大模型。提出了一种基于类别引导的电力视觉语言大模型PowerVLM,设计了类别引导增强模块,增强模型对电力图文数据的理解和问答能力;采用FL的强化学习训练策略,在满足数据隐私保护下,降低域间差异对模型性能的影响;最后,提出了一种基于信息决议的模型剪枝算法,可实现低训练参数的模型高效微调。分别在变电巡检、输电任务、作业安监3种典型电力场景开展实验,结果表明,该方法在电力场景多模态问答任务中的METEOR、BLEU和CIDEr等各项指标均表现优异,为电力场景智能感知提供了新的技术思路和方法支撑。 展开更多
关键词 智能电网 人工智能 视觉语言大模型 Federated Learning 模型剪枝
在线阅读 下载PDF
A review of covalent organic framework materials for CO_(2) adsorption
7
作者 Zeng Fanqiang Ji Changchun +1 位作者 Wang Yudong Li Yousheng 《新型炭材料(中英文)》 北大核心 2026年第1期44-66,共23页
The capture of atmospheric carbon dioxide by adsorbents is an important strategy to deal with the greenhouse effect.Compared with traditional CO_(2) adsorption materials like activated carbon,silica gel,and zeolite mo... The capture of atmospheric carbon dioxide by adsorbents is an important strategy to deal with the greenhouse effect.Compared with traditional CO_(2) adsorption materials like activated carbon,silica gel,and zeolite molecular sieves,covalent organic frameworks(COFs)have excellent thermal and chemical stabilities and can be produced in many different forms.Using their different possible construction units,ordered structures for specific applications can be produced,giving them broad prospects in fields such as gas storage.This review analyzes the different types of COFs that have been synthesized and their different methods of CO_(2) capture.It then discusses different ways to increase CO_(2) adsorption by changing the internal structure of COFs and modifying their surfaces.The limitations of COF-derived carbon materials in CO_(2) capture are reviewed and,finally,the key role of machine learning and computational simulation in improving CO_(2) adsorption is mentioned,and the current status and future possible uses of COFs are summarized. 展开更多
关键词 CO_(2)capture Modification of COFs COF-derived carbon materials Machine learning Computational simulation
在线阅读 下载PDF
A Survey of Federated Learning:Advances in Architecture,Synchronization,and Security Threats
8
作者 Faisal Mahmud Fahim Mahmud Rashedur M.Rahman 《Computers, Materials & Continua》 2026年第3期1-87,共87页
Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitiv... Federated Learning(FL)has become a leading decentralized solution that enables multiple clients to train a model in a collaborative environment without directly sharing raw data,making it suitable for privacy-sensitive applications such as healthcare,finance,and smart systems.As the field continues to evolve,the research field has become more complex and scattered,covering different system designs,training methods,and privacy techniques.This survey is organized around the three core challenges:how the data is distributed,how models are synchronized,and how to defend against attacks.It provides a structured and up-to-date review of FL research from 2023 to 2025,offering a unified taxonomy that categorizes works by data distribution(Horizontal FL,Vertical FL,Federated Transfer Learning,and Personalized FL),training synchronization(synchronous and asynchronous FL),optimization strategies,and threat models(data leakage and poisoning attacks).In particular,we summarize the latest contributions in Vertical FL frameworks for secure multi-party learning,communication-efficient Horizontal FL,and domain-adaptive Federated Transfer Learning.Furthermore,we examine synchronization techniques addressing system heterogeneity,including straggler mitigation in synchronous FL and staleness management in asynchronous FL.The survey covers security threats in FL,such as gradient inversion,membership inference,and poisoning attacks,as well as their defense strategies that include privacy-preserving aggregation and anomaly detection.The paper concludes by outlining unresolved issues and highlighting challenges in handling personalized models,scalability,and real-world adoption. 展开更多
关键词 Federated learning(FL) horizontal federated learning(HFL) vertical federated learning(VFL) federated transfer learning(FTL) personalized federated learning synchronous federated learning(SFL) asynchronous federated learning(AFL) data leakage poisoning attacks privacy-preserving machine learning
在线阅读 下载PDF
Mitigating Attribute Inference in Split Learning via Channel Pruning and Adversarial Training
9
作者 Afnan Alhindi Saad Al-Ahmadi Mohamed Maher Ben Ismail 《Computers, Materials & Continua》 2026年第3期1465-1489,共25页
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn... Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%. 展开更多
关键词 Split learning privacy-preserving split learning distributed collaborative machine learning channel pruning adversarial learning resource-constrained devices
在线阅读 下载PDF
A State-of-the-Art Survey of Adversarial Reinforcement Learning for IoT Intrusion Detection
10
作者 Qasem Abu Al-Haija Shahad Al Tamimi 《Computers, Materials & Continua》 2026年第4期26-94,共69页
Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Tr... Adversarial Reinforcement Learning(ARL)models for intelligent devices and Network Intrusion Detection Systems(NIDS)improve systemresilience against sophisticated cyber-attacks.As a core component of ARL,Adversarial Training(AT)enables NIDS agents to discover and prevent newattack paths by exposing them to competing examples,thereby increasing detection accuracy,reducing False Positives(FPs),and enhancing network security.To develop robust decision-making capabilities for real-world network disruptions and hostile activity,NIDS agents are trained in adversarial scenarios to monitor the current state and notify management of any abnormal or malicious activity.The accuracy and timeliness of the IDS were crucial to the network’s availability and reliability at this time.This paper analyzes ARL applications in NIDS,revealing State-of-The-Art(SoTA)methodology,issues,and future research prospects.This includes Reinforcement Machine Learning(RML)-based NIDS,which enables an agent to interact with the environment to achieve a goal,andDeep Reinforcement Learning(DRL)-based NIDS,which can solve complex decision-making problems.Additionally,this survey study addresses cybersecurity adversarial circumstances and their importance for ARL and NIDS.Architectural design,RL algorithms,feature representation,and training methodologies are examined in the ARL-NIDS study.This comprehensive study evaluates ARL for intelligent NIDS research,benefiting cybersecurity researchers,practitioners,and policymakers.The report promotes cybersecurity defense research and innovation. 展开更多
关键词 Reinforcement learning network intrusion detection adversarial training deep learning cybersecurity defense intrusion detection system and machine learning
在线阅读 下载PDF
Task-Structured Curriculum Learning for Multi-Task Distillation:Enhancing Step-by-Step Knowledge Transfer in Language Models
11
作者 Ahmet Ezgi Aytug Onan 《Computers, Materials & Continua》 2026年第3期1647-1673,共27页
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re... Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning. 展开更多
关键词 Knowledge distillation curriculum learning language models multi-task learning step-by-step learning
在线阅读 下载PDF
A Cooperative Hybrid Learning Framework for Automated Dandruff Severity Grading
12
作者 Sin-Ye Jhong Hui-Che Hsu +3 位作者 Hsin-Hua Huang Chih-Hsien Hsia Yulius Harjoseputro Yung-Yao Chen 《Computers, Materials & Continua》 2026年第4期2272-2285,共14页
Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.S... Automated grading of dandruff severity is a clinically significant but challenging task due to the inherent ordinal nature of severity levels and the high prevalence of label noise from subjective expert annotations.Standard classification methods fail to address these dual challenges,limiting their real-world performance.In this paper,a novel,three-phase training framework is proposed that learns a robust ordinal classifier directly from noisy labels.The approach synergistically combines a rank-based ordinal regression backbone with a cooperative,semi-supervised learning strategy to dynamically partition the data into clean and noisy subsets.A hybrid training objective is then employed,applying a supervised ordinal loss to the clean set.The noisy set is simultaneously trained using a dualobjective that combines a semi-supervised ordinal loss with a parallel,label-agnostic contrastive loss.This design allows themodel to learn fromthe entire noisy subset while using contrastive learning to mitigate the risk of error propagation frompotentially corrupt supervision.Extensive experiments on a new,large-scale,multi-site clinical dataset validate our approach.Themethod achieves state-of-the-art performance with 80.71%accuracy and a 76.86%F1-score,significantly outperforming existing approaches,including a 2.26%improvement over the strongest baseline method.This work provides not only a robust solution for a practical medical imaging problem but also a generalizable framework for other tasks plagued by noisy ordinal labels. 展开更多
关键词 Dandruff severity grading ordinal regression noisy label learning self-supervised learning contrastive learning medical image analysis
在线阅读 下载PDF
Formulating an Innovative Gamified Personalized Learning Ecosystem Integrating 3D/VR Environments,Machine Learning,and Business Intelligence
13
作者 Nymfodora-Maria Raftopoulou Petros L.Pallis 《Sociology Study》 2026年第1期13-32,共20页
Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalize... Latest digital advancements have intensified the necessity for adaptive,data-driven and socially-centered learning ecosystems.This paper presents the formulation of a cross-platform,innovative,gamified and personalized Learning Ecosystem,which integrates 3D/VR environments,as well as machine learning algorithms,and business intelligence frameworks to enhance learner-centered education and inferenced decision-making.This Learning System makes use of immersive,analytically assessed virtual learning spaces,therefore facilitating real-time monitoring of not just learning performance,but also overall engagement and behavioral patterns,via a comprehensive set of sustainability-oriented ESG-aligned Key Performance Indicators(KPIs).Machine learning models support predictive analysis,personalized feedback,and hybrid recommendation mechanisms,whilst dedicated dashboards translate complex educational data into actionable insights for all Use Cases of the System(Educational Institutions,Educators and Learners).Additionally,the presented Learning System introduces a structured Mentoring and Consulting Subsystem,thence reinforcing human-centered guidance alongside automated intelligence.The Platform’s modular architecture and simulation-centered evaluation approach actively support personalized,and continuously optimized learning pathways.Thence,it exemplifies a mature,adaptive Learning Ecosystem,supporting immersive technologies,analytics,and pedagogical support,hence,contributing to contemporary digital learning innovation and sociotechnical transformation in education. 展开更多
关键词 gamified learning ecosystems learning analytics business intelligence personalized education virtual reality machine learning
在线阅读 下载PDF
TeachSecure-CTI:Adaptive Cybersecurity Curriculum Generation Using Threat Dynamics and AI
14
作者 Alaa Tolah 《Computers, Materials & Continua》 2026年第4期1698-1734,共37页
The rapidly evolving cybersecurity threat landscape exposes a critical flaw in traditional educational programs where static curricula cannot adapt swiftly to novel attack vectors.This creates a significant gap betwee... The rapidly evolving cybersecurity threat landscape exposes a critical flaw in traditional educational programs where static curricula cannot adapt swiftly to novel attack vectors.This creates a significant gap between theoretical knowledge and the practical defensive capabilities needed in the field.To address this,we propose TeachSecure-CTI,a novel framework for adaptive cybersecurity curriculumgeneration that integrates real-time Cyber Threat Intelligence(CTI)with AI-driven personalization.Our framework employs a layered architecture featuring a CTI ingestion and clusteringmodule,natural language processing for semantic concept extraction,and a reinforcement learning agent for adaptive content sequencing.Bydynamically aligning learningmaterialswithboththe evolving threat environment and individual learner profiles,TeachSecure-CTI ensures content remains current,relevant,and tailored.A 12-week study with 150 students across three institutions demonstrated that the framework improves learning gains by 34%,significantly exceeding the 12%–21%reported in recent literature.The system achieved 84.8%personalization accuracy,85.9%recognition accuracy for MITRE ATT&CK tactics,and a 31%faster competency development rate compared to static curricula.These findings have implications beyond academia,extending to workforce development,cyber range training,and certification programs.By bridging the gap between dynamic threats and static educational materials,TeachSecure-CTI offers an empirically validated,scalable solution for cultivating cybersecurity professionals capable of responding to modern threats. 展开更多
关键词 Adaptive learning cybersecurity education threat intelligence artificial intelligence curriculumgeneration personalised learning
在线阅读 下载PDF
A comprehensive analysis of artificial intelligence,machine learning,deep learning and computer vision in food science
15
作者 Premkumar Borugadda Hemantha Kumar Kalluri 《Journal of Future Foods》 2026年第6期975-991,共17页
Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food proce... Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food processing from seed to harvest,storage,preparation,and consumption.This current paper seeks to demystify the importance of artificial intelligence,machine learning(ML),deep learning(DL),and computer vision(CV)in ensuring food safety and quality.By stressing the importance of these technologies,the audience will feel reassured and confident in their potential.These are very handy for such problems,giving assurance over food safety.CV is incredibly noble in today's generation because it improves food processing quality and positively impacts firms and researchers.Thus,at the present production stage,rich in image processing and computer visioning is incorporated into all facets of food production.In this field,DL and ML are implemented to identify the type of food in addition to quality.Concerning data and result-oriented perceptions,one has found similarities regarding various approaches.As a result,the findings of this study will be helpful for scholars looking for a proper approach to identify the quality of food offered.It helps to indicate which food products have been discussed by other scholars and lets the reader know papers by other scholars inclined to research further.Also,DL is accurately integrated with identifying the quality and safety of foods in the market.This paper describes the current practices and concerns of ML,DL,and probable trends for its future development. 展开更多
关键词 Artificial intelligence Computer vision Deep learning Food quality Food recognition Machine learning
在线阅读 下载PDF
Harnessing artificial intelligence for the assessment of liver fibrosis and steatosis via multiparametric ultrasound
16
作者 Nicholas Viceconti Silvia Andaloro +8 位作者 Mattia Paratore Sara Miliani Giulia D’Acunzo Giuseppe Cerniglia Fabrizio Mancuso Elena Melita Antonio Gasbarrini Laura Riccardi Matteo Garcovich 《World Journal of Gastroenterology》 2026年第2期59-76,共18页
Artificial intelligence(AI)is revolutionizing medical imaging,particularly in chronic liver diseases assessment.AI technologies,including machine learning and deep learning,are increasingly integrated with multiparame... Artificial intelligence(AI)is revolutionizing medical imaging,particularly in chronic liver diseases assessment.AI technologies,including machine learning and deep learning,are increasingly integrated with multiparametric ultrasound(US)techniques to provide more accurate,objective,and non-invasive evaluations of liver fibrosis and steatosis.Analyzing large datasets from US images,AI enhances diagnostic precision,enabling better quantification of liver stiffness and fat content,which are essential for diagnosing and staging liver fibrosis and steatosis.Combining advanced US modalities,such as elastography and doppler imaging with AI,has demonstrated improved sensitivity in identifying different stages of liver disease and distinguishing various degrees of steatotic liver.These advancements also contribute to greater reproducibility and reduced operator dependency,addressing some of the limitations of traditional methods.The clinical implications of AI in liver disease are vast,ranging from early detection to predicting disease progression and evaluating treatment response.Despite these promising developments,challenges such as the need for large-scale datasets,algorithm transparency,and clinical validation remain.The aim of this review is to explore the current applications and future potential of AI in liver fibrosis and steatosis assessment using multiparametric US,highlighting the technological advances and clinical relevance of this emerging field. 展开更多
关键词 Artificial intelligence Multiparametric ultrasound LIVER FIBROSIS STEATOSIS Shear wave elastography Attenuation imaging Machine learning Deep learning
在线阅读 下载PDF
Improved expert system of rockburst intensity level prediction based on machine learning and data-driven:Supported by 1114 rockburst cases in 197 rock underground projects
17
作者 PANG Hong-li GONG Feng-qiang +1 位作者 GAO Ming-zhong DAI Jin-hao 《Journal of Central South University》 2026年第1期335-356,共22页
Accurate prediction of rockburst intensity levels is crucial for ensuring the safety of deep hard rock engineering construction.This paper introduced an expert system for rockburst intensity level prediction that empl... Accurate prediction of rockburst intensity levels is crucial for ensuring the safety of deep hard rock engineering construction.This paper introduced an expert system for rockburst intensity level prediction that employs machine learning algorithms as the basis for its inference rules.The system comprises four modules:a database,a repository,an inference engine,and an interpreter.A database containing 1114 rockburst cases was used to construct 357 datasets that serve as the repository for the expert system.Additionally,19 types of machine learning algorithms were used to establish 6783 micro-models to construct cognitive rules within the inference engine.By integrating probability theory and marginal analysis,a fuzzy scoring method based on the SoftMax function was developed and applied to the interpreter for rockburst intensity level prediction,effectively restoring the continuity of rockburst characteristics.The research results indicate that ensemble algorithms based on decision trees are more effective in capturing the characteristics of rockburst.Key factors for accurate prediction of rockburst intensity include uniaxial compressive strength,elastic energy index,the maximum principal stress,tangential stress,and their composite indicators.The accuracy of the proposed rockburst intensity level prediction expert system was verified using 20 engineering rockburst cases,with predictions aligning closely with the actual rockburst intensity levels. 展开更多
关键词 rock mechanics ROCKBURST rockburst intensity level prediction expert system machine learning supervised learning
在线阅读 下载PDF
A Review on Fault Diagnosis Methods of Gas Turbine
18
作者 Tao Zhang Hailun Wang +1 位作者 Tianyue Wang Tian Tian 《Computers, Materials & Continua》 2026年第3期88-116,共29页
The critical components of gas turbines suffer from prolonged exposure to factors such as thermal oxidation,mechanical wear,and airflow disturbances during prolonged operation.These conditions can lead to a series of ... The critical components of gas turbines suffer from prolonged exposure to factors such as thermal oxidation,mechanical wear,and airflow disturbances during prolonged operation.These conditions can lead to a series of issues,including mechanical faults,air path malfunctions,and combustion irregularities.Traditional modelbased approaches face inherent limitations due to their inability to handle nonlinear problems,natural factors,measurement uncertainties,fault coupling,and implementation challenges.The development of artificial intelligence algorithms has provided an effective solution to these issues,sparking extensive research into data-driven fault diagnosis methodologies.The review mechanism involved searching IEEE Xplore,ScienceDirect,and Web of Science for peerreviewed articles published between 2019 and 2025,focusing on multi-fault diagnosis techniques.A total of 220 papers were identified,with 123 meeting the inclusion criteria.This paper provides a comprehensive review of diagnostic methodologies,detailing their operational principles and distinctive features.It analyzes current research hotspots and challenges while forecasting future trends.The study systematically evaluates the strengths and limitations of various fault diagnosis techniques,revealing their practical applicability and constraints through comparative analysis.Furthermore,this paper looks forward to the future development direction of this field and provides a valuable reference for the optimization and development of gas turbine fault diagnosis technology in the future. 展开更多
关键词 Fault diagnosis machine learning gas turbine artificial intelligence deep learning
在线阅读 下载PDF
Intelligent Semantic Segmentation with Vision Transformers for Aerial Vehicle Monitoring
19
作者 Moneerah Alotaibi 《Computers, Materials & Continua》 2026年第1期1629-1648,共20页
Advanced traffic monitoring systems encounter substantial challenges in vehicle detection and classification due to the limitations of conventional methods,which often demand extensive computational resources and stru... Advanced traffic monitoring systems encounter substantial challenges in vehicle detection and classification due to the limitations of conventional methods,which often demand extensive computational resources and struggle with diverse data acquisition techniques.This research presents a novel approach for vehicle classification and recognition in aerial image sequences,integrating multiple advanced techniques to enhance detection accuracy.The proposed model begins with preprocessing using Multiscale Retinex(MSR)to enhance image quality,followed by Expectation-Maximization(EM)Segmentation for precise foreground object identification.Vehicle detection is performed using the state-of-the-art YOLOv10 framework,while feature extraction incorporates Maximally Stable Extremal Regions(MSER),Dense Scale-Invariant Feature Transform(Dense SIFT),and Zernike Moments Features to capture distinct object characteristics.Feature optimization is further refined through a Hybrid Swarm-based Optimization algorithm,ensuring optimal feature selection for improved classification performance.The final classification is conducted using a Vision Transformer,leveraging its robust learning capabilities for enhanced accuracy.Experimental evaluations on benchmark datasets,including UAVDT and the Unmanned Aerial Vehicle Intruder Dataset(UAVID),demonstrate the superiority of the proposed approach,achieving an accuracy of 94.40%on UAVDT and 93.57%on UAVID.The results highlight the efficacy of the model in significantly enhancing vehicle detection and classification in aerial imagery,outperforming existing methodologies and offering a statistically validated improvement for intelligent traffic monitoring systems compared to existing approaches. 展开更多
关键词 Machine learning semantic segmentation remote sensors deep learning object monitoring system
在线阅读 下载PDF
LinguTimeX a Framework for Multilingual CTC Detection Using Explainable AI and Natural Language Processing
20
作者 Omar Darwish Shorouq Al-Eidi +4 位作者 Abdallah Al-Shorman Majdi Maabreh Anas Alsobeh Plamen Zahariev Yahya Tashtoush 《Computers, Materials & Continua》 2026年第1期2231-2251,共21页
Covert timing channels(CTC)exploit network resources to establish hidden communication pathways,posing signi cant risks to data security and policy compliance.erefore,detecting such hidden and dangerous threats remain... Covert timing channels(CTC)exploit network resources to establish hidden communication pathways,posing signi cant risks to data security and policy compliance.erefore,detecting such hidden and dangerous threats remains one of the security challenges. is paper proposes LinguTimeX,a new framework that combines natural language processing with arti cial intelligence,along with explainable Arti cial Intelligence(AI)not only to detect CTC but also to provide insights into the decision process.LinguTimeX performs multidimensional feature extraction by fusing linguistic attributes with temporal network patterns to identify covert channels precisely.LinguTimeX demonstrates strong e ectiveness in detecting CTC across multiple languages;namely English,Arabic,and Chinese.Speci cally,the LSTM and RNN models achieved F1 scores of 90%on the English dataset,89%on the Arabic dataset,and 88%on the Chinese dataset,showcasing their superior performance and ability to generalize across multiple languages. is highlights their robustness in detecting CTCs within security systems,regardless of the language or cultural context of the data.In contrast,the DeepForest model produced F1-scores ranging from 86%to 87%across the same datasets,further con rming its e ectiveness in CTC detection.Although other algorithms also showed reasonable accuracy,the LSTM and RNN models consistently outperformed them in multilingual settings,suggesting that deep learning models might be better suited for this particular problem. 展开更多
关键词 Arabic language Chinese language covert timing channel CYBERSECURITY deep learning English language language processing machine learning
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部