期刊文献+
共找到53,146篇文章
< 1 2 250 >
每页显示 20 50 100
Microseismic signal processing and rockburst disaster identification:A multi-task deep learning and machine learning approach
1
作者 Chunchi Ma Weihao Xu +3 位作者 Xuefeng Ran Tianbin Li Hang Zhang Dongwei Xing 《Journal of Rock Mechanics and Geotechnical Engineering》 2026年第1期441-456,共16页
Underground engineering projects such as deep tunnel excavation often encounter rockburst disasters accompanied by numerous microseismic events.Rapid interpretation of microseismic signals is crucial for the timely id... Underground engineering projects such as deep tunnel excavation often encounter rockburst disasters accompanied by numerous microseismic events.Rapid interpretation of microseismic signals is crucial for the timely identification of rockbursts.However,conventional processing encompasses multi-step workflows,including classification,denoising,picking,locating,and computational analysis,coupled with manual intervention,which collectively compromise the reliability of early warnings.To address these challenges,this study innovatively proposes the“microseismic stethoscope"-a multi-task machine learning and deep learning model designed for the automated processing of massive microseismic signals.This model efficiently extracts three key parameters that are necessary for recognizing rockburst disasters:rupture location,microseismic energy,and moment magnitude.Specifically,the model extracts raw waveform features from three dedicated sub-networks:a classifier for source zone classification,and two regressors for microseismic energy and moment magnitude estimation.This model demonstrates superior efficiency compared to traditional processing and semi-automated processing,reducing per-event processing time from 0.71 s to 0.49 s to merely 0.036 s.It concurrently achieves 98%accuracy in source zone classification,with microseismic energy and moment magnitude estimation errors of 0.13 and 0.05,respectively.This model has been well applied and validated in the Daxiagu Tunnel case in Sichuan,China.The application results indicate that the model is as accurate as traditional methods in determining source parameters,and thus can be used to identify potential geomechanical processes of rockburst disasters.By enhancing the signal processing reliability of microseismic events,the proposed model in this study presents a significant advancement in the identification of rockburst disasters. 展开更多
关键词 Underground engineering Microseismic signal processing deep learning multi-task Rockburst identification
在线阅读 下载PDF
Spatio-Temporal Cellular Network Traffic Prediction Using Multi-Task Deep Learning for AI-Enabled 6G 被引量:1
2
作者 Xiaochuan Sun Biao Wei +3 位作者 Jiahui Gao Difei Cao Zhigang Li Yingqi Li 《Journal of Beijing Institute of Technology》 EI CAS 2022年第5期441-453,共13页
Spatio-temporal cellular network traffic prediction at wide-area level plays an important role in resource reconfiguration,traffic scheduling and intrusion detection,thus potentially supporting connected intelligence ... Spatio-temporal cellular network traffic prediction at wide-area level plays an important role in resource reconfiguration,traffic scheduling and intrusion detection,thus potentially supporting connected intelligence of the sixth generation of mobile communications technology(6G).However,the existing studies just focus on the spatio-temporal modeling of traffic data of single network service,such as short message,call,or Internet.It is not conducive to accurate prediction of traffic data,characterised by diverse network service,spatio-temporality and supersize volume.To address this issue,a novel multi-task deep learning framework is developed for citywide cellular network traffic prediction.Functionally,this framework mainly consists of a dual modular feature sharing layer and a multi-task learning layer(DMFS-MT).The former aims at mining long-term spatio-temporal dependencies and local spatio-temporal fluctuation trends in data,respectively,via a new combination of convolutional gated recurrent unit(ConvGRU)and 3-dimensional convolutional neural network(3D-CNN).For the latter,each task is performed for predicting service-specific traffic data based on a fully connected network.On the real-world Telecom Italia dataset,simulation results demonstrate the effectiveness of our proposal through prediction performance measure,spatial pattern comparison and statistical distribution verification. 展开更多
关键词 the sixth generation of mobile communications technology(6G) cellular network traffic multi-task deep learning spatio-temporality
在线阅读 下载PDF
Multi-Task Deep Learning with Task Attention for Post-Click Conversion Rate Prediction
3
作者 Hongxin Luo Xiaobing Zhou +1 位作者 Haiyan Ding Liqing Wang 《Intelligent Automation & Soft Computing》 SCIE 2023年第6期3583-3593,共11页
Online advertising has gained much attention on various platforms as a hugely lucrative market.In promoting content and advertisements in real life,the acquisition of user target actions is usually a multi-step proces... Online advertising has gained much attention on various platforms as a hugely lucrative market.In promoting content and advertisements in real life,the acquisition of user target actions is usually a multi-step process,such as impres-sion→click→conversion,which means the process from the delivery of the recommended item to the user’s click to the final conversion.Due to data sparsity or sample selection bias,it is difficult for the trained model to achieve the business goal of the target campaign.Multi-task learning,a classical solution to this pro-blem,aims to generalize better on the original task given several related tasks by exploiting the knowledge between tasks to share the same feature and label space.Adaptively learned task relations bring better performance to make full use of the correlation between tasks.We train a general model capable of captur-ing the relationships between various tasks on all existing active tasks from a meta-learning perspective.In addition,this paper proposes a Multi-task Attention Network(MAN)to identify commonalities and differences between tasks in the feature space.The model performance is improved by explicitly learning the stacking of task relationships in the label space.To illustrate the effectiveness of our method,experiments are conducted on Alibaba Click and Conversion Pre-diction(Ali-CCP)dataset.Experimental results show that the method outperforms the state-of-the-art multi-task learning methods. 展开更多
关键词 multi-task learning recommend system ATTENTION META-learning
在线阅读 下载PDF
A Multi-Task Deep Learning Framework for Simultaneous Detection of Thoracic Pathology through Image Classification
4
作者 Nada Al Zahrani Ramdane Hedjar +4 位作者 Mohamed Mekhtiche Mohamed Bencherif Taha Al Fakih Fattoh Al-Qershi Muna Alrazghan 《Journal of Computer and Communications》 2024年第4期153-170,共18页
Thoracic diseases pose significant risks to an individual's chest health and are among the most perilous medical diseases. They can impact either one or both lungs, which leads to a severe impairment of a person’... Thoracic diseases pose significant risks to an individual's chest health and are among the most perilous medical diseases. They can impact either one or both lungs, which leads to a severe impairment of a person’s ability to breathe normally. Some notable examples of such diseases encompass pneumonia, lung cancer, coronavirus disease 2019 (COVID-19), tuberculosis, and chronic obstructive pulmonary disease (COPD). Consequently, early and precise detection of these diseases is paramount during the diagnostic process. Traditionally, the primary methods employed for the detection involve the use of X-ray imaging or computed tomography (CT) scans. Nevertheless, due to the scarcity of proficient radiologists and the inherent similarities between these diseases, the accuracy of detection can be compromised, leading to imprecise or erroneous results. To address this challenge, scientists have turned to computer-based solutions, aiming for swift and accurate diagnoses. The primary objective of this study is to develop two machine learning models, utilizing single-task and multi-task learning frameworks, to enhance classification accuracy. Within the multi-task learning architecture, two principal approaches exist soft parameter sharing and hard parameter sharing. Consequently, this research adopts a multi-task deep learning approach that leverages CNNs to achieve improved classification performance for the specified tasks. These tasks, focusing on pneumonia and COVID-19, are processed and learned simultaneously within a multi-task model. To assess the effectiveness of the trained model, it is rigorously validated using three different real-world datasets for training and testing. 展开更多
关键词 PNEUMONIA Thoracic Pathology COVID-19 deep learning multi-task learning
暂未订购
Task-Structured Curriculum Learning for Multi-Task Distillation:Enhancing Step-by-Step Knowledge Transfer in Language Models
5
作者 Ahmet Ezgi Aytug Onan 《Computers, Materials & Continua》 2026年第3期1647-1673,共27页
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re... Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning. 展开更多
关键词 Knowledge distillation curriculum learning language models multi-task learning step-by-step learning
在线阅读 下载PDF
A comprehensive analysis of artificial intelligence,machine learning,deep learning and computer vision in food science
6
作者 Premkumar Borugadda Hemantha Kumar Kalluri 《Journal of Future Foods》 2026年第6期975-991,共17页
Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food proce... Providing safe and quality food is crucial for every household and is of extreme significance in the growth of any society.It is a complex procedure that deals with all issues focusing on the development of food processing from seed to harvest,storage,preparation,and consumption.This current paper seeks to demystify the importance of artificial intelligence,machine learning(ML),deep learning(DL),and computer vision(CV)in ensuring food safety and quality.By stressing the importance of these technologies,the audience will feel reassured and confident in their potential.These are very handy for such problems,giving assurance over food safety.CV is incredibly noble in today's generation because it improves food processing quality and positively impacts firms and researchers.Thus,at the present production stage,rich in image processing and computer visioning is incorporated into all facets of food production.In this field,DL and ML are implemented to identify the type of food in addition to quality.Concerning data and result-oriented perceptions,one has found similarities regarding various approaches.As a result,the findings of this study will be helpful for scholars looking for a proper approach to identify the quality of food offered.It helps to indicate which food products have been discussed by other scholars and lets the reader know papers by other scholars inclined to research further.Also,DL is accurately integrated with identifying the quality and safety of foods in the market.This paper describes the current practices and concerns of ML,DL,and probable trends for its future development. 展开更多
关键词 Artificial intelligence Computer vision deep learning Food quality Food recognition Machine learning
在线阅读 下载PDF
Machine Learning and Deep Learning for Smart Urban Transportation Systems with GPS,GIS,and Advanced Analytics:A Comprehensive Analysis
7
作者 E.Kalaivanan S.Brindha 《Journal of Harbin Institute of Technology(New Series)》 2026年第1期81-96,共16页
As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impact... As urbanization continues to accelerate,the challenges associated with managing transportation in metropolitan areas become increasingly complex.The surge in population density contributes to traffic congestion,impacting travel experiences and posing safety risks.Smart urban transportation management emerges as a strategic solution,conceptualized here as a multidimensional big data problem.The success of this strategy hinges on the effective collection of information from diverse,extensive,and heterogeneous data sources,necessitating the implementation of full⁃stack Information and Communication Technology(ICT)solutions.The main idea of the work is to investigate the current technologies of Intelligent Transportation Systems(ITS)and enhance the safety of urban transportation systems.Machine learning models,trained on historical data,can predict traffic congestion,allowing for the implementation of preventive measures.Deep learning architectures,with their ability to handle complex data representations,further refine traffic predictions,contributing to more accurate and dynamic transportation management.The background of this research underscores the challenges posed by traffic congestion in metropolitan areas and emphasizes the need for advanced technological solutions.By integrating GPS and GIS technologies with machine learning algorithms,this work aims to pay attention to the development of intelligent transportation systems that not only address current challenges but also pave the way for future advancements in urban transportation management. 展开更多
关键词 machine learning deep learning smart transportation
在线阅读 下载PDF
A Comparative Benchmark of Machine and Deep Learning for Cyberattack Detection in IoT Networks
8
作者 Enzo Hoummady Fehmi Jaafar 《Computers, Materials & Continua》 2026年第4期1070-1092,共23页
With the proliferation of Internet of Things(IoT)devices,securing these interconnected systems against cyberattacks has become a critical challenge.Traditional security paradigms often fail to cope with the scale and ... With the proliferation of Internet of Things(IoT)devices,securing these interconnected systems against cyberattacks has become a critical challenge.Traditional security paradigms often fail to cope with the scale and diversity of IoT network traffic.This paper presents a comparative benchmark of classic machine learning(ML)and state-of-the-art deep learning(DL)algorithms for IoT intrusion detection.Our methodology employs a twophased approach:a preliminary pilot study using a custom-generated dataset to establish baselines,followed by a comprehensive evaluation on the large-scale CICIoTDataset2023.We benchmarked algorithms including Random Forest,XGBoost,CNN,and StackedLSTM.The results indicate that while top-performingmodels frombothcategories achieve over 99%classification accuracy,this metric masks a crucial performance trade-off.We demonstrate that treebased ML ensembles exhibit superior precision(91%)in identifying benign traffic,making them effective at reducing false positives.Conversely,DL models demonstrate superior recall(96%),making them better suited for minimizing the interruption of legitimate traffic.We conclude that the selection of an optimal model is not merely a matter of maximizing accuracy but is a strategic choice dependent on the specific security priority either minimizing false alarms or ensuring service availability.Thiswork provides a practical framework for deploying context-aware security solutions in diverse IoT environments. 展开更多
关键词 Internet of Things deep learning abnormal network traffic cyberattacks machine learning
在线阅读 下载PDF
A novel method for EPID transmission dose generation using Monte Carlo simulation and deep learning
9
作者 Tao Qiu Ning Gao +3 位作者 Yan-Kui Chang Xi Pei Huan-Li Luo Fu Jin 《Nuclear Science and Techniques》 2026年第4期41-52,共12页
This study aimed to integrate Monte Carlo(MC)simulation with deep learning(DL)-based denoising techniques to achieve fast and accurate prediction of high-quality electronic portal imaging device(EPID)transmission dose... This study aimed to integrate Monte Carlo(MC)simulation with deep learning(DL)-based denoising techniques to achieve fast and accurate prediction of high-quality electronic portal imaging device(EPID)transmission dose(TD)for patientspecific quality assurance(PSQA).A total of 100 lung cases were used to obtain the noisy EPID TD by the ARCHER MC code under four kinds of particle numbers(1×10^(6),1×10^(7),1×10^(8)and 1×10^(9)),and the original EPID TD was denoised by the SUNet neural network.The denoised EPID TD was assessed both qualitatively and quantitatively using the structural similarity(SSIM),peak signal-to-noise ratio(PSNR),and gamma passing rate(GPR)with respect to 1×10^(9)as a reference.The computation times for both the MC simulation and DL-based denoising were recorded.As the number of particles increased,both the quality of the noisy EPID TD and computation time increased significantly(1×10^(6):1.12 s,1×10^(7):1.72 s,1×10^(8):8.62 s,and 1×10^(9):73.89 s).In contrast,the DL-based denoising time remained at 0.13-0.16 s.The denoised EPID TD shows a smoother visual appearance and profile curves,but differences between 1×10^(6)and 1×10^(9)still remain.SSIM improves from 0.61 to 0.95 for 1×10^(6),0.70 to 0.96 for 1×10^(7),and 0.90 to 0.97 for 1×10^(8).PSNR increases by>20%for 1×10^(6)and 1×10^(7),and>10%for 1×10^(8).GPR improves from 48.47%to 89.10%for 1×10^(6),61.04%to 94.35%for 1×10^(7),and 91.88%to 99.55%for 1×10^(8).The method that combines MC simulation with DL-based denoising for EPID TD generation can accelerate TD prediction and maintain high accuracy,offering a promising solution for efficient PSQA. 展开更多
关键词 PSQA EPID Monte Carlo deep learning
在线阅读 下载PDF
Forecasting solar cycles using the time-series dense encoder deep learning model
10
作者 Cui Zhao Shangbin Yang +1 位作者 Jianguo Liu Shiyuan Liu 《Astronomical Techniques and Instruments》 2026年第1期43-54,共12页
The solar cycle(SC),a phenomenon caused by the quasi-periodic regular activities in the Sun,occurs approximately every 11 years.Intense solar activity can disrupt the Earth’s ionosphere,affecting communication and na... The solar cycle(SC),a phenomenon caused by the quasi-periodic regular activities in the Sun,occurs approximately every 11 years.Intense solar activity can disrupt the Earth’s ionosphere,affecting communication and navigation systems.Consequently,accurately predicting the intensity of the SC holds great significance,but predicting the SC involves a long-term time series,and many existing time series forecasting methods have fallen short in terms of accuracy and efficiency.The Time-series Dense Encoder model is a deep learning solution tailored for long time series prediction.Based on a multi-layer perceptron structure,it outperforms the best previously existing models in accuracy,while being efficiently trainable on general datasets.We propose a method based on this model for SC forecasting.Using a trained model,we predict the test set from SC 19 to SC 25 with an average mean absolute percentage error of 32.02,root mean square error of 30.3,mean absolute error of 23.32,and R^(2)(coefficient of determination)of 0.76,outperforming other deep learning models in terms of accuracy and training efficiency on sunspot number datasets.Subsequently,we use it to predict the peaks of SC 25 and SC 26.For SC 25,the peak time has ended,but a stronger peak is predicted for SC 26,of 199.3,within a range of 170.8-221.9,projected to occur during April 2034. 展开更多
关键词 Solar cycle Forecasting TIDE deep learning
在线阅读 下载PDF
Deep Learning-Assisted Organogel Pressure Sensor for Alphabet Recognition and Bio-Mechanical Motion Monitoring
11
作者 Kusum Sharma Kousik Bhunia +5 位作者 Subhajit Chatterjee Muthukumar Perumalsamy Anandhan Ayyappan Saj Theophilus Bhatti Yung‑Cheol Byun Sang-Jae Kim 《Nano-Micro Letters》 2026年第2期644-663,共20页
Wearable sensors integrated with deep learning techniques have the potential to revolutionize seamless human-machine interfaces for real-time health monitoring,clinical diagnosis,and robotic applications.Nevertheless,... Wearable sensors integrated with deep learning techniques have the potential to revolutionize seamless human-machine interfaces for real-time health monitoring,clinical diagnosis,and robotic applications.Nevertheless,it remains a critical challenge to simultaneously achieve desirable mechanical and electrical performance along with biocompatibility,adhesion,self-healing,and environmental robustness with excellent sensing metrics.Herein,we report a multifunctional,anti-freezing,selfadhesive,and self-healable organogel pressure sensor composed of cobalt nanoparticle encapsulated nitrogen-doped carbon nanotubes(CoN CNT)embedded in a polyvinyl alcohol-gelatin(PVA/GLE)matrix.Fabricated using a binary solvent system of water and ethylene glycol(EG),the CoN CNT/PVA/GLE organogel exhibits excellent flexibility,biocompatibility,and temperature tolerance with remarkable environmental stability.Electrochemical impedance spectroscopy confirms near-stable performance across a broad humidity range(40%-95%RH).Freeze-tolerant conductivity under sub-zero conditions(-20℃)is attributed to the synergistic role of CoN CNT and EG,preserving mobility and network integrity.The Co N CNT/PVA/GLE organogel sensor exhibits high sensitivity of 5.75 k Pa^(-1)in the detection range from 0 to 20 k Pa,ideal for subtle biomechanical motion detection.A smart human-machine interface for English letter recognition using deep learning achieved 98%accuracy.The organogel sensor utility was extended to detect human gestures like finger bending,wrist motion,and throat vibration during speech. 展开更多
关键词 Wearable ORGANOGEL deep learning Pressure sensor Bio-mechanical motion
在线阅读 下载PDF
A Hybrid Approach to Software Testing Efficiency:Stacked Ensembles and Deep Q-Learning for Test Case Prioritization and Ranking
12
作者 Anis Zarrad Thomas Armstrong Jaber Jemai 《Computers, Materials & Continua》 2026年第3期1726-1746,共21页
Test case prioritization and ranking play a crucial role in software testing by improving fault detection efficiency and ensuring software reliability.While prioritization selects the most relevant test cases for opti... Test case prioritization and ranking play a crucial role in software testing by improving fault detection efficiency and ensuring software reliability.While prioritization selects the most relevant test cases for optimal coverage,ranking further refines their execution order to detect critical faults earlier.This study investigates machine learning techniques to enhance both prioritization and ranking,contributing to more effective and efficient testing processes.We first employ advanced feature engineering alongside ensemble models,including Gradient Boosted,Support Vector Machines,Random Forests,and Naive Bayes classifiers to optimize test case prioritization,achieving an accuracy score of 0.98847 and significantly improving the Average Percentage of Fault Detection(APFD).Subsequently,we introduce a deep Q-learning framework combined with a Genetic Algorithm(GA)to refine test case ranking within priority levels.This approach achieves a rank accuracy of 0.9172,demonstrating robust performance despite the increasing computational demands of specialized variation operators.Our findings highlight the effectiveness of stacked ensemble learning and reinforcement learning in optimizing test case prioritization and ranking.This integrated approach improves testing efficiency,reduces late-stage defects,and improves overall software stability.The study provides valuable information for AI-driven testing frameworks,paving the way for more intelligent and adaptive software quality assurance methodologies. 展开更多
关键词 Software testing test case prioritization test case ranking machine learning reinforcement learning deep Q-learning
在线阅读 下载PDF
Fatigue Detection with Multimodal Physiological Signals via Uncertainty-Aware Deep Transfer Learning
13
作者 Kourosh Kakhi Hamzeh Asgharnezhad +2 位作者 Abbas Khosravi Roohallah Alizadehsani U.Rajendra Acharya 《Journal of Bionic Engineering》 2026年第1期472-487,共16页
Accurate detection of driver fatigue is essential for improving road safety.This study investigates the effectiveness of using multimodal physiological signals for fatigue detection while incorporating uncertainty qua... Accurate detection of driver fatigue is essential for improving road safety.This study investigates the effectiveness of using multimodal physiological signals for fatigue detection while incorporating uncertainty quantification to enhance the reliability of predictions.Physiological signals,including Electrocardiogram(ECG),Galvanic Skin Response(GSR),and Electroencephalogram(EEG),were transformed into image representations and analyzed using pretrained deep neu-ral networks.The extracted features were classified through a feedforward neural network,and prediction reliability was assessed using uncertainty quantification techniques such as Monte Carlo Dropout(MCD),model ensembles,and combined approaches.Evaluation metrics included standard measures(sensitivity,specificity,precision,and accuracy)along with uncertainty-aware metrics such as uncertainty sensitivity and uncertainty precision.Across all evaluations,ECG-based models consistently demonstrated strong performance.The findings indicate that combining multimodal physi-ological signals,Transfer Learning(TL),and uncertainty quantification can significantly improve both the accuracy and trustworthiness of fatigue detection systems.This approach supports the development of more reliable driver assistance technologies aimed at preventing fatigue-related accidents. 展开更多
关键词 Fatigue detection Multimodal physiological signals deep transfer learning Uncertainty-aware learning Driver monitoring
在线阅读 下载PDF
Energy Optimization for Autonomous Mobile Robot Path Planning Based on Deep Reinforcement Learning
14
作者 Longfei Gao Weidong Wang Dieyun Ke 《Computers, Materials & Continua》 2026年第1期984-998,共15页
At present,energy consumption is one of the main bottlenecks in autonomous mobile robot development.To address the challenge of high energy consumption in path planning for autonomous mobile robots navigating unknown ... At present,energy consumption is one of the main bottlenecks in autonomous mobile robot development.To address the challenge of high energy consumption in path planning for autonomous mobile robots navigating unknown and complex environments,this paper proposes an Attention-Enhanced Dueling Deep Q-Network(ADDueling DQN),which integrates a multi-head attention mechanism and a prioritized experience replay strategy into a Dueling-DQN reinforcement learning framework.A multi-objective reward function,centered on energy efficiency,is designed to comprehensively consider path length,terrain slope,motion smoothness,and obstacle avoidance,enabling optimal low-energy trajectory generation in 3D space from the source.The incorporation of a multihead attention mechanism allows the model to dynamically focus on energy-critical state features—such as slope gradients and obstacle density—thereby significantly improving its ability to recognize and avoid energy-intensive paths.Additionally,the prioritized experience replay mechanism accelerates learning from key decision-making experiences,suppressing inefficient exploration and guiding the policy toward low-energy solutions more rapidly.The effectiveness of the proposed path planning algorithm is validated through simulation experiments conducted in multiple off-road scenarios.Results demonstrate that AD-Dueling DQN consistently achieves the lowest average energy consumption across all tested environments.Moreover,the proposed method exhibits faster convergence and greater training stability compared to baseline algorithms,highlighting its global optimization capability under energy-aware objectives in complex terrains.This study offers an efficient and scalable intelligent control strategy for the development of energy-conscious autonomous navigation systems. 展开更多
关键词 Autonomous mobile robot deep reinforcement learning energy optimization multi-attention mechanism prioritized experience replay dueling deep Q-Network
在线阅读 下载PDF
An explainable deep learning approach to enhance the prediction of shield tunnel deviation
15
作者 Jiajie Zhen Fengwen Lai +4 位作者 Ming Huang Junjie Zheng Jim S.Shiau Ping Wang Jinhuo Zheng 《Journal of Rock Mechanics and Geotechnical Engineering》 2026年第1期566-579,共14页
Although machine learning models have achieved high enough accuracy in predicting shield position deviations,their“black box”nature makes the prediction mechanisms and decision-making processes opaque,leading to wea... Although machine learning models have achieved high enough accuracy in predicting shield position deviations,their“black box”nature makes the prediction mechanisms and decision-making processes opaque,leading to weaker explanations and practicability.This study introduces a novel explainable deep learning framework comprising the Informer model with enhanced attention mechanisms(EAMInfor)and deep learning important features(DeepLIFT),aimed at improving the prediction accuracy of shield position deviations and providing interpretability for predictive results.The EAMInfor model attempts to integrate channel attention,spatial attention,and simple attention modules to improve the Informer model's performance.The framework is tested with the four different geological conditions datasets generated from the Xiamen metro line 3,China.Results show that the EAMInfor model outperforms the traditional Informer and comparison models.The analysis with the DeepLIFT method indicates that the push thrust of push cylinder and the earth chamber pressure are the most significant features,while the stroke length of the push cylinder demonstrated lower importance.Furthermore,the variation trends in the significance of data points within input sequences exhibit substantial differences between single and composite strata.This framework not only improves predictive accuracy but also strengthens the credibility and reliability of the results. 展开更多
关键词 Shield tunnel position deviation Machine learning Explainable AI deep learning important features
在线阅读 下载PDF
Nondestructive detection of key phenotypes for the canopy of the watermelon plug seedlings based on deep learning
16
作者 Lei Li Zhilong Bie +4 位作者 Yi Zhang Yuan Huang Chengli Peng Binbin Han Shengyong Xu 《Horticultural Plant Journal》 2026年第1期149-160,共12页
Nondestructive measurement technology of phenotype can provide substantial phenotypic data support for applications such as seedling breeding,management,and quality testing.The current method of measuring seedling phe... Nondestructive measurement technology of phenotype can provide substantial phenotypic data support for applications such as seedling breeding,management,and quality testing.The current method of measuring seedling phenotypes mainly relies on manual measurement which is inefficient,subjective and destroys samples.Therefore,the paper proposes a nondestructive measurement method for the canopy phenotype of the watermelon plug seedlings based on deep learning.The Azure Kinect was used to shoot canopy color images,depth images,and RGB-D images of the watermelon plug seedlings.The Mask-RCNN network was used to classify,segment,and count the canopy leaves of the watermelon plug seedlings.To reduce the error of leaf area measurement caused by mutual occlusion of leaves,the leaves were repaired by CycleGAN,and the depth images were restored by image processing.Then,the Delaunay triangulation was adopted to measure the leaf area in the leaf point cloud.The YOLOX target detection network was used to identify the growing point position of each seedling on the plug tray.Then the depth differences between the growing point and the upper surface of the plug tray were calculated to obtain plant height.The experiment results show that the nondestructive measurement algorithm proposed in this paper achieves good measurement performance for the watermelon plug seedlings from the 1 true-leaf to 3 true-leaf stages.The average relative error of measurement is 2.33%for the number of true leaves,4.59%for the number of cotyledons,8.37%for the leaf area,and 3.27%for the plant height.The experiment results demonstrate that the proposed algorithm in this paper provides an effective solution for the nondestructive measurement of the canopy phenotype of the plug seedlings. 展开更多
关键词 Watermelon seedlings Azure Kinect CANOPY Phenotype detection deep learning
在线阅读 下载PDF
Inverse design of 3D integrated high-efficiency grating couplers using deep learning
17
作者 Yu Wang Yue Wang +4 位作者 Guohui Yang Kuang Zhang Xing Yang Chunhui Wang Yu Zhang 《Chinese Physics B》 2026年第2期363-373,共11页
In recent years,the use of deep learning to replace traditional numerical methods for electromagnetic propagation has shown tremendous potential in the rapid design of photonic devices.However,most research on deep le... In recent years,the use of deep learning to replace traditional numerical methods for electromagnetic propagation has shown tremendous potential in the rapid design of photonic devices.However,most research on deep learning has focused on single-layer grating couplers,and the accuracy of multi-layer grating couplers has not yet reached a high level.This paper proposes and demonstrates a novel deep learning network-assisted strategy for inverse design.The network model is based on a multi-layer perceptron(MLP)and incorporates convolutional neural networks(CNNs)and transformers.Through the stacking of multiple layers,it achieves a high-precision design for both multi-layer and single-layer raster couplers with various functionalities.The deep learning network exhibits exceptionally high predictive accuracy,with an average absolute error across the full wavelength range of 1300–1700 nm being only 0.17%,and an even lower predictive absolute error below 0.09%at the specific wavelength of 1550 nm.By combining the deep learning network with the genetic algorithm,we can efficiently design grating couplers that perform different functions.Simulation results indicate that the designed single-wavelength grating couplers achieve coupling efficiencies exceeding 80%at central wavelengths of 1550 nm and 1310 nm.The performance of designed dual-wavelength and broadband grating couplers also reaches high industry standards.Furthermore,the network structure and inverse design method are highly scalable and can be applied not only to multi-layer grating couplers but also directly to the prediction and design of single-layer grating couplers,providing a new perspective for the innovative development of photonic devices. 展开更多
关键词 deep learning inverse design grating couplers photonic devices
原文传递
Composite Deep-Learning Model for 90-Day mRS Prediction in Post-Stroke Patients
18
作者 Shihan Dong Zhengwei Yao +2 位作者 Yuhang Chuai Ran Li Handong Zhang 《Journal of Clinical and Nursing Research》 2026年第1期301-307,共7页
To counteract small sample size,severe class imbalance and high feature redundancy in 90-day mRS prediction after stroke,this study proposes a four-stage pipeline-“ADASYN re-sampling→clinical+statistical feature scr... To counteract small sample size,severe class imbalance and high feature redundancy in 90-day mRS prediction after stroke,this study proposes a four-stage pipeline-“ADASYN re-sampling→clinical+statistical feature screening→dimensionality reduction→5-fold cross-validation”-and benchmark composite deep-learning architectures.ADASYN first balances the minority classes in the original feature space.Next,a tri-level filter(clinical domain knowledge,variance threshold,mutual information)removes clinically meaningless or redundant variables,after which PCA compresses the remaining features while preserving critical neurological signatures(e.g.,brain-herniation history).Four hybrid CNN-RNN models are trained and compared under strict 5-fold cross-validation;the optimal ensemble yields stable,clinically interpretable probabilities that can support individualized rehabilitation planning. 展开更多
关键词 STROKE 90-day mRS Composite deep learning ADASYN 5-fold cross-validation
在线阅读 下载PDF
Advances in deep learning for bacterial image segmentation in optical microscopy
19
作者 Zhijun Tan Yang Ding +6 位作者 Huibin Ma Jintao Li Danrou Zheng Hua Bai Weini Xin Lin Li Bo Peng 《Journal of Innovative Optical Health Sciences》 2026年第1期30-44,共15页
Microscopy imaging is fundamental in analyzing bacterial morphology and dynamics,offering critical insights into bacterial physiology and pathogenicity.Image segmentation techniques enable quantitative analysis of bac... Microscopy imaging is fundamental in analyzing bacterial morphology and dynamics,offering critical insights into bacterial physiology and pathogenicity.Image segmentation techniques enable quantitative analysis of bacterial structures,facilitating precise measurement of morphological variations and population behaviors at single-cell resolution.This paper reviews advancements in bacterial image segmentation,emphasizing the shift from traditional thresholding and watershed methods to deep learning-driven approaches.Convolutional neural networks(CNNs),U-Net architectures,and three-dimensional(3D)frameworks excel at segmenting dense biofilms and resolving antibiotic-induced morphological changes.These methods combine automated feature extraction with physics-informed postprocessing.Despite progress,challenges persist in computational efficiency,cross-species generalizability,and integration with multimodal experimental workflows.Future progress will depend on improving model robustness across species and imaging modalities,integrating multimodal data for phenotype-function mapping,and developing standard pipelines that link computational tools with clinical diagnostics.These innovations will expand microbial phenotyping beyond structural analysis,enabling deeper insights into bacterial physiology and ecological interactions. 展开更多
关键词 Bacterial image deep learning optical microscopy image segmentation artificial intelligence
原文传递
Prediction of Regional Surface Wave Parameters in the Qinhuangdao Sea Using a Deep Learning Model with Limited Observational Data
20
作者 WANG Lei FANG Kezhao +2 位作者 ZHOU Long GONG Lixin HUO Yongwei 《Journal of Ocean University of China》 2026年第1期74-90,共17页
Waves are important physical phenomena in an ocean,and their accurate prediction is essential for ocean engineering,maritime traffic,and marine early warning systems.This study focuses on the Qinhuangdao Sea area loca... Waves are important physical phenomena in an ocean,and their accurate prediction is essential for ocean engineering,maritime traffic,and marine early warning systems.This study focuses on the Qinhuangdao Sea area located in the Bohai Sea,China.Herein,we use on-site wind data to correct the reanalysis wind data obtained from the European Centre for Medium-Range Weather Forecasts(ECMWF),improving the accuracy of boundary conditions.Then,we use the Simulating WAves Nearshore(SWAN)model to simulate the regional wave field over time.A regional wave-parameter prediction model is then developed using a limited number of sampled data(covering only 2 years,2020–2021);the model is based on the Whale Optimization Algorithm(WOA),convolutional neural networks(CNNs),and long short-term memory(LSTM)neural networks.WOA is used to optimize the CNN and LSTM framework;in this framework,CNN extracts spatial features,and the LSTM network captures temporal features,enabling accurate short and long-term predictions of wave height,period,and direction.The experimental results showed that despite the small sample size,the model achieves a goodness of fit of 0.9957 for wave height prediction,0.9973 for period,and 0.9749 for wave direction in short-term forecasting.As the prediction step size increases,the accuracy of the model decreases.When the prediction step size reaches 9 h,the root mean square error for the prediction of wave height,period,and direction increases to 0.2060 m,0.4582 s,and32.5358°,respectively.The reliability and applicability of the model are further validated by the experimental results.Our findings highlighted the potential of the developed model in operational wave forecasting,even with a limited number of sampled data. 展开更多
关键词 regional wave prediction deep learning WOA-CNN-LSTM numerical simulation Bohai Sea
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部