期刊文献+
共找到260,018篇文章
< 1 2 250 >
每页显示 20 50 100
Mitigating Attribute Inference in Split Learning via Channel Pruning and Adversarial Training
1
作者 Afnan Alhindi Saad Al-Ahmadi Mohamed Maher Ben Ismail 《Computers, Materials & Continua》 2026年第3期1465-1489,共25页
Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subn... Split Learning(SL)has been promoted as a promising collaborative machine learning technique designed to address data privacy and resource efficiency.Specifically,neural networks are divided into client and server subnetworks in order to mitigate the exposure of sensitive data and reduce the overhead on client devices,thereby making SL particularly suitable for resource-constrained devices.Although SL prevents the direct transmission of raw data,it does not alleviate entirely the risk of privacy breaches.In fact,the data intermediately transmitted to the server sub-model may include patterns or information that could reveal sensitive data.Moreover,achieving a balance between model utility and data privacy has emerged as a challenging problem.In this article,we propose a novel defense approach that combines:(i)Adversarial learning,and(ii)Network channel pruning.In particular,the proposed adversarial learning approach is specifically designed to reduce the risk of private data exposure while maintaining high performance for the utility task.On the other hand,the suggested channel pruning enables the model to adaptively adjust and reactivate pruned channels while conducting adversarial training.The integration of these two techniques reduces the informativeness of the intermediate data transmitted by the client sub-model,thereby enhancing its robustness against attribute inference attacks without adding significant computational overhead,making it wellsuited for IoT devices,mobile platforms,and Internet of Vehicles(IoV)scenarios.The proposed defense approach was evaluated using EfficientNet-B0,a widely adopted compact model,along with three benchmark datasets.The obtained results showcased its superior defense capability against attribute inference attacks compared to existing state-of-the-art methods.This research’s findings demonstrated the effectiveness of the proposed channel pruning-based adversarial training approach in achieving the intended compromise between utility and privacy within SL frameworks.In fact,the classification accuracy attained by the attackers witnessed a drastic decrease of 70%. 展开更多
关键词 Split learning privacy-preserving split learning distributed collaborative machine learning channel pruning adversarial learning resource-constrained devices
在线阅读 下载PDF
Robotic Cell Micromanipulation Skill Learning via Imitation-Enhanced Reinforcement Learning
2
作者 Youchao Zhang Fanghao Wang +4 位作者 Xiangyu Guo Yibin Ying Mingchuan Zhou Zhongliang Jiang Alois Knoll 《CAAI Transactions on Intelligence Technology》 2026年第1期123-136,共14页
Humans can learn complex and dexterous manipulation tasks by observing videos,imitating and exploring.Multiple endeffectors manipulation of free micron-sized deformable cells is one of the challenging tasks in robotic... Humans can learn complex and dexterous manipulation tasks by observing videos,imitating and exploring.Multiple endeffectors manipulation of free micron-sized deformable cells is one of the challenging tasks in robotic micromanipulation.We propose an imitation-enhanced reinforcement learning method inspired by the human learning process that enables robots to learn cell micromanipulation skills from videos.Firstly,for the microscopic robot micromanipulation videos,a multi-task observation(MTO)network is designed to identify the two end-effectors and the manipulated objects to obtain the spatiotemporal trajectories.The spatiotemporal constraints of the robot's actions are obtained by the task-parameterised hidden Markov model(THMM).To simultaneously address the safety and dexterity of robot micromanipulation,an imitation learning optimisation-based soft actor-critic(ILOSAC)algorithm is proposed in which the robot can perform skill learning by demonstration and exploration.The proposed method is capable of performing complex cell manipulation tasks in a realistic physical environment.Experiments indicated that compared with current methods and manual remote manipulation,the proposed framework achieved a shorter operation time and less deformation of cells,which is expected to facilitate the development of robot skill learning. 展开更多
关键词 deep learning intelligent robots intelligent systems ROBOTICS
在线阅读 下载PDF
Imitation Learning for Unmanned Aerial Vehicle Obstacle Avoidance Based on Visual Features with DAgger
3
作者 Yuqi Yang Mengyun Wang +1 位作者 Yifeng Niu Bo Wang 《Journal of Beijing Institute of Technology》 2026年第1期114-126,共13页
Unmanned aerial vehicles(UAVs)face the challenge of autonomous obstacle avoidance in complex,multi-obstacle environments.Behavior cloning offers a promising approach to rapidly acquire a learning policy from limited e... Unmanned aerial vehicles(UAVs)face the challenge of autonomous obstacle avoidance in complex,multi-obstacle environments.Behavior cloning offers a promising approach to rapidly acquire a learning policy from limited expert demonstrations.However,pure imitation learning inherently suffers from poor exploration and limited generalization,typically necessitating extensive datasets to train competent student policies.We utilize a cross-modal variational autoencoder(CM-VAE)to extract compact features from raw visual inputs and UAV states,which then feed into a policy network.We evaluated our approach in a simulated environment featuring a challenging circular trajectory with eight gate obstacles.The results demonstrate that the policy trained with pure behavior cloning consistently failed.In stark contrast,our DAgger-augmented behavior cloning method successfully traversed all gates without collision.Our findings confirm that DAgger effectively mitigates the shortcomings of behavior cloning,enabling the creation of reliable and sample-efficient navigation policies for UAVs. 展开更多
关键词 imitation learning unmanned aerial vehicle obstacle avoidance DAGGER
在线阅读 下载PDF
DPIL-Traj: Differential Privacy Trajectory Generation Framework with Imitation Learning
4
作者 Huaxiong Liao Xiangxuan Zhong +4 位作者 Xueqi Chen Yirui Huang Yuwei Lin Jing Zhang Bruce Gu 《Computers, Materials & Continua》 2026年第1期1530-1550,共21页
The generation of synthetic trajectories has become essential in various fields for analyzing complex movement patterns.However,the use of real-world trajectory data poses significant privacy risks,such as location re... The generation of synthetic trajectories has become essential in various fields for analyzing complex movement patterns.However,the use of real-world trajectory data poses significant privacy risks,such as location reidentification and correlation attacks.To address these challenges,privacy-preserving trajectory generation methods are critical for applications relying on sensitive location data.This paper introduces DPIL-Traj,an advanced framework designed to generate synthetic trajectories while achieving a superior balance between data utility and privacy preservation.Firstly,the framework incorporates Differential Privacy Clustering,which anonymizes trajectory data by applying differential privacy techniques that add noise,ensuring the protection of sensitive user information.Secondly,Imitation Learning is used to replicate decision-making behaviors observed in real-world trajectories.By learning from expert trajectories,this component generates synthetic data that closely mimics real-world decision-making processes while optimizing the quality of the generated trajectories.Finally,Markov-based Trajectory Generation is employed to capture and maintain the inherent temporal dynamics of movement patterns.Extensive experiments conducted on the GeoLife trajectory dataset show that DPIL-Traj improves utility performance by an average of 19.85%,and in terms of privacy performance by an average of 12.51%,compared to state-of-the-art approaches.Ablation studies further reveal that DP clustering effectively safeguards privacy,imitation learning enhances utility under noise,and the Markov module strengthens temporal coherence. 展开更多
关键词 PRIVACY-PRESERVING trajectory generation differential privacy imitation learning Markov chain
在线阅读 下载PDF
Enhanced Scene Recognition via Multi-Model Transfer Learning with Limited Labeled Data
5
作者 Samia Allaoua Chelloug Ahmed A.Abd El-Latif +1 位作者 Samah Al Shathri Mohamed Hammad 《Computers, Materials & Continua》 2026年第5期1191-1211,共21页
Scene recognition is a critical component of computer vision,powering applications from autonomous vehicles to surveillance systems.However,its development is often constrained by a heavy reliance on large,expensively... Scene recognition is a critical component of computer vision,powering applications from autonomous vehicles to surveillance systems.However,its development is often constrained by a heavy reliance on large,expensively annotated datasets.This research presents a novel,efficient approach that leveragesmulti-model transfer learning from pre-trained deep neural networks—specifically DenseNet201 and Visual Geometry Group(VGG)—to overcome this limitation.Ourmethod significantly reduces dependency on vast labeled data while achieving high accuracy.Evaluated on the Aerial Image Dataset(AID)dataset,the model attained a validation accuracy of 93.6%with a loss of 0.35,demonstrating robust performance with minimal training data.These results underscore the viability of our approach for real-time,data-efficient scene recognition,offering a practical and cost-effective advancement for the field. 展开更多
关键词 Scene recognition transfer learning pre-trained deep models DenseNet201 VGG
在线阅读 下载PDF
Prediction of Regional Surface Wave Parameters in the Qinhuangdao Sea Using a Deep Learning Model with Limited Observational Data
6
作者 WANG Lei FANG Kezhao +2 位作者 ZHOU Long GONG Lixin HUO Yongwei 《Journal of Ocean University of China》 2026年第1期74-90,共17页
Waves are important physical phenomena in an ocean,and their accurate prediction is essential for ocean engineering,maritime traffic,and marine early warning systems.This study focuses on the Qinhuangdao Sea area loca... Waves are important physical phenomena in an ocean,and their accurate prediction is essential for ocean engineering,maritime traffic,and marine early warning systems.This study focuses on the Qinhuangdao Sea area located in the Bohai Sea,China.Herein,we use on-site wind data to correct the reanalysis wind data obtained from the European Centre for Medium-Range Weather Forecasts(ECMWF),improving the accuracy of boundary conditions.Then,we use the Simulating WAves Nearshore(SWAN)model to simulate the regional wave field over time.A regional wave-parameter prediction model is then developed using a limited number of sampled data(covering only 2 years,2020–2021);the model is based on the Whale Optimization Algorithm(WOA),convolutional neural networks(CNNs),and long short-term memory(LSTM)neural networks.WOA is used to optimize the CNN and LSTM framework;in this framework,CNN extracts spatial features,and the LSTM network captures temporal features,enabling accurate short and long-term predictions of wave height,period,and direction.The experimental results showed that despite the small sample size,the model achieves a goodness of fit of 0.9957 for wave height prediction,0.9973 for period,and 0.9749 for wave direction in short-term forecasting.As the prediction step size increases,the accuracy of the model decreases.When the prediction step size reaches 9 h,the root mean square error for the prediction of wave height,period,and direction increases to 0.2060 m,0.4582 s,and32.5358°,respectively.The reliability and applicability of the model are further validated by the experimental results.Our findings highlighted the potential of the developed model in operational wave forecasting,even with a limited number of sampled data. 展开更多
关键词 regional wave prediction deep learning WOA-CNN-LSTM numerical simulation Bohai Sea
在线阅读 下载PDF
Machine Learning-Based GPS Spoofing Detection and Mitigation for UAVs
7
作者 Charlotte Olivia Namagembe Mohamad Ibrahim +1 位作者 Md Arafatur Rahman Prashant Pillai 《Computers, Materials & Continua》 2026年第2期1153-1172,共20页
The rapid proliferation of commercial unmanned aerial vehicles(UAVs)has revolutionized fields such as precision agriculture and disaster response.However,their heavy reliance on GPS navigation leaves them highly vulne... The rapid proliferation of commercial unmanned aerial vehicles(UAVs)has revolutionized fields such as precision agriculture and disaster response.However,their heavy reliance on GPS navigation leaves them highly vulnerable to spoofing attacks,with potentially severe consequences.To mitigate this threat,we present a machine learning-driven framework for real-time GPS spoofing detection,designed with a balance of detection accuracy and computational efficiency.Our work is distinguished by the creation of a comprehensive dataset of 10,000 instances that integrates both simulated and real-world data,enabling robust and generalizable model development.A comprehensive evaluation ofmultiple classification algorithms identifies XGBoost as the superior performer,achieving 93.07% accuracy alongside outstanding precision,recall,and F1-scores.Beyond standard classification metrics,our assessment encompasses ROC-AUC,detection latency,and false positive rate,providing a comprehensive assessment of performance.This work contributes to UAV security by providing a robust and reproducible solution for detecting GPS spoofing attacks,supported by a detailed methodology,a comprehensive evaluation including inference-time latency,and a publicly available dataset. 展开更多
关键词 Commercial unmanned aerial vehicles global positioning systems machine learning techniques spoofing attack
在线阅读 下载PDF
A review of dynamic monitoring methods for intermittent rivers:Integrating remote sensing and machine learning
8
作者 XIE Chaoshuai Lv Aifeng 《Journal of Geographical Sciences》 2026年第3期763-796,共34页
Intermittent rivers and ephemeral streams(IRES),also known as non-perennial river segments(NPRs),have garnered attention due to their significant roles in watershed hydrology and ecosystem services,especially in the c... Intermittent rivers and ephemeral streams(IRES),also known as non-perennial river segments(NPRs),have garnered attention due to their significant roles in watershed hydrology and ecosystem services,especially in the context of climate change and escalating human activities.Recent advances in machine learning(ML)techniques have significantly improved the analysis of dynamic changes in IRES.Various ML models,including random forest(RF),long short-term memory(LSTM),and U-Net,demonstrate clear advantages in processing complex hydrological data,enhancing the efficiency and accuracy of IRES extraction from remote sensing data.Furthermore,hybrid ML approaches enhance predictive performance in complex hydrological scenarios by integrating multiple algorithms.However,ML methods still face challenges,including high data dependence,computational complexity,and scalability issues with models.This review proposes an IRES monitoring framework that combines satellite data with ML algorithms,integrating remote sensing technologies such as optical imaging and synthetic aperture radar,and evaluates the advantages and limitations of different ML methods.It further highlights the potential of integrating multiple ML techniques and high-resolution remote sensing data to monitor IRES dynamics,conduct ecological assessments,and support sustainable water management,offering a scientific foundation for addressing environmental and anthropogenic pressures. 展开更多
关键词 machine learning intermittent rivers and ephemeral streams remote sensing framework algorithm selection
原文传递
FRF-BiLSTM:Recognising and Mitigating DDoS Attacks through a Secure Decentralized Feature Optimized Federated Learning Approach
9
作者 Sushruta Mishra Sunil Kumar Mohapatra +2 位作者 Kshira Sagar Sahoo Anand Nayyar Tae-Kyung Kim 《Computers, Materials & Continua》 2026年第3期1118-1138,共21页
With an increase in internet-connected devices and a dependency on online services,the threat of Distributed Denial of Service(DDoS)attacks has become a significant concern in cybersecurity.The proposed system follows... With an increase in internet-connected devices and a dependency on online services,the threat of Distributed Denial of Service(DDoS)attacks has become a significant concern in cybersecurity.The proposed system follows a multi-step process,beginning with the collection of datasets from different edge devices and network nodes.To verify its effectiveness,experiments were conducted using the CICDoS2017,NSL-KDD,and CICIDS benchmark datasets alongside other existing models.Recursive feature elimination(RFE)with random forest is used to select features from the CICDDoS2019 dataset,on which a BiLSTM model is trained on local nodes.Local models are trained until convergence or stability criteria are met while simultaneously sharing the updates globally for collaborative learning.A centralised server evaluates real-time traffic using the global BiLSTM model,which triggers alerts for potential DDoS attacks.Furthermore,blockchain technology is employed to secure model updates and to provide an immutable audit trail,thereby ensuring trust and accountability among network nodes.This research introduces a novel decentralized method called Federated Random Forest Bidirectional Long Short-Term Memory(FRF-BiLSTM)for detecting DDoS attacks,utilizing the advanced Bidirectional Long Short-Term Memory Networks(BiLSTMs)to analyze sequences in both forward and backward directions.The outcome shows the proposed model achieves a mean accuracy of 97.1%with an average training delay of 88.7 s and testing delay of 21.4 s.The model demonstrates scalability and the best detection performance in large-scale attack scenarios. 展开更多
关键词 Bi-directional long short-term memory network distributed denial of service(DDoS) CYBERSECURITY federated learning random forest
在线阅读 下载PDF
A deep transfer learning model for the deformation of braced excavations with limited monitoring data 被引量:2
10
作者 Yuanqin Tao Shaoxiang Zeng +3 位作者 Tiantian Ying Honglei Sun Sunjuexu Pan Yuanqiang Cai 《Journal of Rock Mechanics and Geotechnical Engineering》 2025年第3期1555-1568,共14页
The current deep learning models for braced excavation cannot predict deformation from the beginning of excavation due to the need for a substantial corpus of sufficient historical data for training purposes.To addres... The current deep learning models for braced excavation cannot predict deformation from the beginning of excavation due to the need for a substantial corpus of sufficient historical data for training purposes.To address this issue,this study proposes a transfer learning model based on a sequence-to-sequence twodimensional(2D)convolutional long short-term memory neural network(S2SCL2D).The model can use the existing data from other adjacent similar excavations to achieve wall deflection prediction once a limited amount of monitoring data from the target excavation has been recorded.In the absence of adjacent excavation data,numerical simulation data from the target project can be employed instead.A weight update strategy is proposed to improve the prediction accuracy by integrating the stochastic gradient masking with an early stopping mechanism.To illustrate the proposed methodology,an excavation project in Hangzhou,China is adopted.The proposed deep transfer learning model,which uses either adjacent excavation data or numerical simulation data as the source domain,shows a significant improvement in performance when compared to the non-transfer learning model.Using the simulation data from the target project even leads to better prediction performance than using the actual monitoring data from other adjacent excavations.The results demonstrate that the proposed model can reasonably predict the deformation with limited data from the target project. 展开更多
关键词 Braced excavation Wall deflections Transfer learning Deep learning Finite element simulation
在线阅读 下载PDF
基于Q-learning的专家权重优化与多级共识反馈决策
11
作者 杜秀丽 程伟龙 +2 位作者 高星 潘成胜 吕亚娜 《计算机应用研究》 北大核心 2026年第2期420-426,共7页
针对动态复杂多属性决策环境下大规模异构专家群体共识达成效率低、权重分配不精准的问题,提出一种基于Q-learning的权重优化与多级共识反馈方法,旨在提升共识水平与决策质量。该方法通过将专家权重动态调整建模为马尔可夫决策过程,利用... 针对动态复杂多属性决策环境下大规模异构专家群体共识达成效率低、权重分配不精准的问题,提出一种基于Q-learning的权重优化与多级共识反馈方法,旨在提升共识水平与决策质量。该方法通过将专家权重动态调整建模为马尔可夫决策过程,利用Q-learning实现权重自适应优化,并设计涵盖属性、方案、专家与群体四个层级的多级共识反馈机制,从而精准识别并协调不同来源的分歧。实验结果表明,该方法能够显著降低共识达成所需迭代次数,提升权重分配与专家专业度的匹配精度,并获得更可靠的方案排序结果,验证了其在大规模异构专家群体中的鲁棒性与计算效率。研究表明,所提方法为复杂多属性群体决策问题提供了有效的共识建模与决策支持工具。 展开更多
关键词 群体决策 Q-learning 多层共识反馈 动态权重调整
在线阅读 下载PDF
CMBA-FL: Communication-mitigated and blockchain-assisted federated learning for traffic flow predictions 被引量:1
12
作者 Kaiyin Zhu Mingming Lu +2 位作者 Haifeng Li Neal NXiong Wenyong He 《Digital Communications and Networks》 2025年第3期724-733,共10页
As an effective strategy to address urban traffic congestion,traffic flow prediction has gained attention from Federated-Learning(FL)researchers due FL’s ability to preserving data privacy.However,existing methods fa... As an effective strategy to address urban traffic congestion,traffic flow prediction has gained attention from Federated-Learning(FL)researchers due FL’s ability to preserving data privacy.However,existing methods face challenges:some are too simplistic to capture complex traffic patterns effectively,and others are overly complex,leading to excessive communication overhead between cloud and edge devices.Moreover,the problem of single point failure limits their robustness and reliability in real-world applications.To tackle these challenges,this paper proposes a new method,CMBA-FL,a Communication-Mitigated and Blockchain-Assisted Federated Learning model.First,CMBA-FL improves the client model’s ability to capture temporal traffic patterns by employing the Encoder-Decoder framework for each edge device.Second,to reduce the communication overhead during federated learning,we introduce a verification method based on parameter update consistency,avoiding unnecessary parameter updates.Third,to mitigate the risk of a single point of failure,we integrate consensus mechanisms from blockchain technology.To validate the effectiveness of CMBA-FL,we assess its performance on two widely used traffic datasets.Our experimental results show that CMBA-FL reduces prediction error by 11.46%,significantly lowers communication overhead,and improves security. 展开更多
关键词 Blockchain Communication mitigating Federated learning Traffic flow prediction
在线阅读 下载PDF
Machine learning-based investigation of uplift resistance in special-shaped shield tunnels using numerical finite element modeling 被引量:1
13
作者 ZHANG Wengang YE Wenyu +2 位作者 SUN Weixin LIU Zhicheng LI Zhengchuan 《土木与环境工程学报(中英文)》 北大核心 2026年第1期1-13,共13页
The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combi... The uplift resistance of the soil overlying shield tunnels significantly impacts their anti-floating stability.However,research on uplift resistance concerning special-shaped shield tunnels is limited.This study combines numerical simulation with machine learning techniques to explore this issue.It presents a summary of special-shaped tunnel geometries and introduces a shape coefficient.Through the finite element software,Plaxis3D,the study simulates six key parameters—shape coefficient,burial depth ratio,tunnel’s longest horizontal length,internal friction angle,cohesion,and soil submerged bulk density—that impact uplift resistance across different conditions.Employing XGBoost and ANN methods,the feature importance of each parameter was analyzed based on the numerical simulation results.The findings demonstrate that a tunnel shape more closely resembling a circle leads to reduced uplift resistance in the overlying soil,whereas other parameters exhibit the contrary effects.Furthermore,the study reveals a diminishing trend in the feature importance of buried depth ratio,internal friction angle,tunnel longest horizontal length,cohesion,soil submerged bulk density,and shape coefficient in influencing uplift resistance. 展开更多
关键词 special-shaped tunnel shield tunnel uplift resistance numerical simulation machine learning
在线阅读 下载PDF
Machine Learning-Assisted Mitigation of Interlayer Heat Accumulation and Optimization of Forming Quality for Laser Powder Bed Fusions 被引量:1
14
作者 Hairui Gao Yanbing Liu +7 位作者 Zhenwu Zhang Jikang Li Juntao Shen Honghao Xiang Wei Li Chao Cai Weisheng Xia Qingsong Wei 《Additive Manufacturing Frontiers》 2025年第2期68-80,共13页
Interlayer heat accumulation(IHA)is major challenge in the laser powder bed fusion(LPBF)process,as it exacerbates the instability of melt pools,and compromises the quality of the as-built samples.Infrared radiation mo... Interlayer heat accumulation(IHA)is major challenge in the laser powder bed fusion(LPBF)process,as it exacerbates the instability of melt pools,and compromises the quality of the as-built samples.Infrared radiation monitoring is an effective method for exploring IHA.Based on the defined sequence features of interlayer infrared radiation intensity(IIRI),this study established a gated recurrent unit(GRU)neural network model for predicting IIRI in formed samples using machine learning to mitigate the IHA.The model trained on 316 L alloys achieved precise prediction results when transferred to the DZ125 superalloy,effectively managing various emergencies in the LPBF process.The truncated pyramid components were fabricated through parameter optimization based on IIRI prediction results.Compared with the non-optimized components,the CT results demonstrated a significant reduction in internal voids,with the relative density increasing from 91.6% to 98.5%.Additionally,surface roughness(Ra)decreased from 32.58μm to 19.91μm,while residual stress on the top surface was reduced from 169.21 MPa to 102.37 MPa. 展开更多
关键词 Laser powder bed fusion Machine learning Gated recurrent unit(GRU) Infrared radiation Residual stress
在线阅读 下载PDF
An Integrated Framework of Grasp Detection and Imitation Learning for Space Robotics Applications 被引量:1
15
作者 Yuming Ning Tuanjie Li +3 位作者 Yulin Zhang Ziang Li Wenqian Du Yan Zhang 《Chinese Journal of Mechanical Engineering》 2025年第4期316-335,共20页
Robots are key to expanding the scope of space applications.The end-to-end training for robot vision-based detection and precision operations is challenging owing to constraints such as extreme environments and high c... Robots are key to expanding the scope of space applications.The end-to-end training for robot vision-based detection and precision operations is challenging owing to constraints such as extreme environments and high computational overhead.This study proposes a lightweight integrated framework for grasp detection and imitation learning,named GD-IL;it comprises a grasp detection algorithm based on manipulability and Gaussian mixture model(manipulability-GMM),and a grasp trajectory generation algorithm based on a two-stage robot imitation learning algorithm(TS-RIL).In the manipulability-GMM algorithm,we apply GMM clustering and ellipse regression to the object point cloud,propose two judgment criteria to generate multiple candidate grasp bounding boxes for the robot,and use manipulability as a metric for selecting the optimal grasp bounding box.The stages of the TS-RIL algorithm are grasp trajectory learning and robot pose optimization.In the first stage,the robot grasp trajectory is characterized using a second-order dynamic movement primitive model and Gaussian mixture regression(GMM).By adjusting the function form of the forcing term,the robot closely approximates the target-grasping trajectory.In the second stage,a robot pose optimization model is built based on the derived pose error formula and manipulability metric.This model allows the robot to adjust its configuration in real time while grasping,thereby effectively avoiding singularities.Finally,an algorithm verification platform is developed based on a Robot Operating System and a series of comparative experiments are conducted in real-world scenarios.The experimental results demonstrate that GD-IL significantly improves the effectiveness and robustness of grasp detection and trajectory imitation learning,outperforming existing state-of-the-art methods in execution efficiency,manipulability,and success rate. 展开更多
关键词 Grasp detection Robot imitation learning MANIPULABILITY Dynamic movement primitives Gaussian mixture model and Gaussian mixture regression Pose optimization
在线阅读 下载PDF
Optimization Scheduling of Hydrogen-Coupled Electro-Heat-Gas Integrated Energy System Based on Generative Adversarial Imitation Learning
16
作者 Baiyue Song Chenxi Zhang +1 位作者 Wei Zhang Leiyu Wan 《Energy Engineering》 2025年第12期4919-4945,共27页
Hydrogen energy is a crucial support for China’s low-carbon energy transition.With the large-scale integration of renewable energy,the combination of hydrogen and integrated energy systems has become one of the most ... Hydrogen energy is a crucial support for China’s low-carbon energy transition.With the large-scale integration of renewable energy,the combination of hydrogen and integrated energy systems has become one of the most promising directions of development.This paper proposes an optimized schedulingmodel for a hydrogen-coupled electro-heat-gas integrated energy system(HCEHG-IES)using generative adversarial imitation learning(GAIL).The model aims to enhance renewable-energy absorption,reduce carbon emissions,and improve grid-regulation flexibility.First,the optimal scheduling problem of HCEHG-IES under uncertainty is modeled as a Markov decision process(MDP).To overcome the limitations of conventional deep reinforcement learning algorithms—including long optimization time,slow convergence,and subjective reward design—this study augments the PPO algorithm by incorporating a discriminator network and expert data.The newly developed algorithm,termed GAIL,enables the agent to perform imitation learning from expert data.Based on this model,dynamic scheduling decisions are made in continuous state and action spaces,generating optimal energy-allocation and management schemes.Simulation results indicate that,compared with traditional reinforcement-learning algorithms,the proposed algorithmoffers better economic performance.Guided by expert data,the agent avoids blind optimization,shortens the offline training time,and improves convergence performance.In the online phase,the algorithm enables flexible energy utilization,thereby promoting renewable-energy absorption and reducing carbon emissions. 展开更多
关键词 Hydrogen energy optimization dispatch generative adversarial imitation learning proximal policy optimization imitation learning renewable energy
在线阅读 下载PDF
Quantifying Global Black Carbon Aging Responses to Emission Reductions Using a Machine Learning-based Climate Model 被引量:1
17
作者 Wenxiang SHEN Minghuai WANG +5 位作者 Junchang WANG Yawen LIU Xinyi DONG Xinyue SHAO Man YUE Yaman LIU 《Advances in Atmospheric Sciences》 2026年第2期361-372,I0004-I0009,共18页
Countries around the world have been making efforts to reduce pollutant emissions. However, the response of global black carbon(BC) aging to emission changes remains unclear. Using the Community Atmosphere Model versi... Countries around the world have been making efforts to reduce pollutant emissions. However, the response of global black carbon(BC) aging to emission changes remains unclear. Using the Community Atmosphere Model version 6 with a machine-learning-integrated four-mode version of the Modal Aerosol Module, we quantify global BC aging responses to emission reductions for 2011–2018 and for 2050 and 2100 under carbon neutrality. During 2011–18, global trends in BC aging degree(mass ratio of coatings to BC, R_(BC)) exhibited marked regional disparities, with a significant increase in China(5.4% yr^(-1)), which contrasts with minimal changes in the USA, Europe, and India. The divergence is attributed to opposing trends in secondary organic aerosol(SOA) and sulfate coatings, driven by regional changes in the emission ratios of corresponding coating precursors to BC(volatile organic compounds-VOCs/BC and SO_(2)/BC). Projections under carbon neutrality reveal that R_(BC) will increase globally by 47%(118%) in 2050(2100), with strong convergent increases expected across major source regions. The R_(BC) increase, primarily driven by enhanced SOA coatings due to sharper BC reductions relative to VOCs, will enhance the global BC mass absorption cross-section(MAC) by 11%(17%) in 2050(2100).Consequently, although the global BC burden will decline sharply by 60%(76%), the enhanced MAC partially offsets the magnitude of the decline in the BC direct radiative effect, resulting in the moderation of global BC DRE decreases to 88%(92%) of the BC burden reductions in 2050(2100). This study highlights the globally enhanced BC aging and light absorption capacity under carbon neutrality, thereby partly offsetting the impact of BC direct emission reductions on future changes in BC radiative effects globally. 展开更多
关键词 black carbon aging trend emission reduction carbon neutrality machine learning
在线阅读 下载PDF
PowerVLM:基于Federated Learning与模型剪枝的电力视觉语言大模型
18
作者 欧阳旭东 雒鹏鑫 +3 位作者 何绍洋 崔艺林 张中超 闫云凤 《全球能源互联网》 北大核心 2026年第1期101-111,共11页
智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learnin... 智能电网的快速发展衍生出多模态、多源异构的海量电力数据,给人工智能模型在复杂电力场景感知带来了挑战,同时行业数据的敏感性和隐私保护需求进一步限制了通用模型在电力领域的跨场景迁移能力。对此,提出了一种基于Federated Learning与模型剪枝的电力视觉语言大模型。提出了一种基于类别引导的电力视觉语言大模型PowerVLM,设计了类别引导增强模块,增强模型对电力图文数据的理解和问答能力;采用FL的强化学习训练策略,在满足数据隐私保护下,降低域间差异对模型性能的影响;最后,提出了一种基于信息决议的模型剪枝算法,可实现低训练参数的模型高效微调。分别在变电巡检、输电任务、作业安监3种典型电力场景开展实验,结果表明,该方法在电力场景多模态问答任务中的METEOR、BLEU和CIDEr等各项指标均表现优异,为电力场景智能感知提供了新的技术思路和方法支撑。 展开更多
关键词 智能电网 人工智能 视觉语言大模型 Federated learning 模型剪枝
在线阅读 下载PDF
A DDPG-based Path Following Control Strategy for Autonomous Vehicles by Integrated Imitation Learning and Feedforward Exploration
19
作者 Qianjie Liu Peixiang Xiong +4 位作者 Qingyuan Zhu Wei Xiao Kejie Wang Guoliang Hu Gang Li 《Chinese Journal of Mechanical Engineering》 2025年第5期207-223,共17页
Autonomous driving technology is constantly developing to a higher level of complex scenes,and there is a growing demand for the utilization of end-to-end data-driven control.However,the end-to-end path tracking proce... Autonomous driving technology is constantly developing to a higher level of complex scenes,and there is a growing demand for the utilization of end-to-end data-driven control.However,the end-to-end path tracking process often encounters challenges in learning efficiency and generalization.To address this issue,this paper designs a deep deterministic policy gradient(DDPG)-based reinforcement learning strategy that integrates imitation learning and feedforward exploration in the path following process.In imitation learning,the path tracking control data generated by the model predictive control(MPC)method is used to train an end-to-end steering control model of a deep neural network.Another feedforward exploration behavior is predicted by road curvature and vehicle speed,and adds it and imitation learning to the DDPG reinforcement learning to obtain decision-making experience and action prediction behavior of the path tracking process.In the reinforcement learning process,imitation learning is used to update the pre-training parameters of the actor network,and a feedforward steering technique with random noise is adopted for strategy exploration.In the reward function,a hierarchical progressive reward form and a constrained objective reward function referring to MPC are designed,and the actor-critic network architecture is determined.Finally,the path tracking performance of the designed method is verified by comparing various training results,simulations,and HIL tests.The results show that the designed method can effectively utilize pre-training and feedforward prior experience to obtain optimal path tracking performance of an autonomous vehicle,and has better generalization ability than other methods.This study provides an efficient control scheme for improving the end-to-end control performance of autonomous vehicles. 展开更多
关键词 Autonomous vehicle Path following Feedforward exploration Reinforcement learning
在线阅读 下载PDF
Early identification of high-risk patients admitted to emergency departments using vital signs and machine learning
20
作者 Qingyuan Liu Yixin Zhang +10 位作者 Jian Sun Kaipeng Wang Yueguo Wang Yulan Wang Cailing Ren Yan Wang Jiashan Zhu Shusheng Zhou Mengping Zhang Yinglei Lai Kui Jin 《World Journal of Emergency Medicine》 2025年第2期113-120,共8页
BACKGROUND:Rapid and accurate identification of high-risk patients in the emergency departments(EDs)is crucial for optimizing resource allocation and improving patient outcomes.This study aimed to develop an early pre... BACKGROUND:Rapid and accurate identification of high-risk patients in the emergency departments(EDs)is crucial for optimizing resource allocation and improving patient outcomes.This study aimed to develop an early prediction model for identifying high-risk patients in EDs using initial vital sign measurements.METHODS:This retrospective cohort study analyzed initial vital signs from the Chinese Emergency Triage,Assessment,and Treatment(CETAT)database,which was collected between January 1^(st),2020,and June 25^(th),2023.The primary outcome was the identification of high-risk patients needing immediate treatment.Various machine learning methods,including a deep-learningbased multilayer perceptron(MLP)classifier were evaluated.Model performance was assessed using the area under the receiver operating characteristic curve(AUC-ROC).AUC-ROC values were reported for three scenarios:a default case,a scenario requiring sensitivity greater than 0.8(Scenario I),and a scenario requiring specificity greater than 0.8(Scenario II).SHAP values were calculated to determine the importance of each predictor within the MLP model.RESULTS:A total of 38,797 patients were analyzed,of whom 18.2%were identified as high-risk.Comparative analysis of the predictive models for high-risk patients showed AUC-ROC values ranging from 0.717 to 0.738,with the MLP model outperforming logistic regression(LR),Gaussian Naive Bayes(GNB),and the National Early Warning Score(NEWS).SHAP value analysis identified coma state,peripheral capillary oxygen saturation(SpO_(2)),and systolic blood pressure as the top three predictive factors in the MLP model,with coma state exerting the most contribution.CONCLUSION:Compared with other methods,the MLP model with initial vital signs demonstrated optimal prediction accuracy,highlighting its potential to enhance clinical decision-making in triage in the EDs. 展开更多
关键词 Machine learning TRIAGE Emergency medicine Decision support systems
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部