期刊文献+
共找到1,233篇文章
< 1 2 62 >
每页显示 20 50 100
Resource Allocation in V2X Networks:A Double Deep Q-Network Approach with Graph Neural Networks
1
作者 Zhengda Huan Jian Sun +3 位作者 Zeyu Chen Ziyi Zhang Xiao Sun Zenghui Xiao 《Computers, Materials & Continua》 2025年第9期5427-5443,共17页
With the advancement of Vehicle-to-Everything(V2X)technology,efficient resource allocation in dynamic vehicular networks has become a critical challenge for achieving optimal performance.Existing methods suffer from h... With the advancement of Vehicle-to-Everything(V2X)technology,efficient resource allocation in dynamic vehicular networks has become a critical challenge for achieving optimal performance.Existing methods suffer from high computational complexity and decision latency under high-density traffic and heterogeneous network conditions.To address these challenges,this study presents an innovative framework that combines Graph Neural Networks(GNNs)with a Double Deep Q-Network(DDQN),utilizing dynamic graph structures and reinforcement learning.An adaptive neighbor sampling mechanism is introduced to dynamically select the most relevant neighbors based on interference levels and network topology,thereby improving decision accuracy and efficiency.Meanwhile,the framework models communication links as nodes and interference relationships as edges,effectively capturing the direct impact of interference on resource allocation while reducing computational complexity and preserving critical interaction information.Employing an aggregation mechanism based on the Graph Attention Network(GAT),it dynamically adjusts the neighbor sampling scope and performs attention-weighted aggregation based on node importance,ensuring more efficient and adaptive resource management.This design ensures reliable Vehicle-to-Vehicle(V2V)communication while maintaining high Vehicle-to-Infrastructure(V2I)throughput.The framework retains the global feature learning capabilities of GNNs and supports distributed network deployment,allowing vehicles to extract low-dimensional graph embeddings from local observations for real-time resource decisions.Experimental results demonstrate that the proposed method significantly reduces computational overhead,mitigates latency,and improves resource utilization efficiency in vehicular networks under complex traffic scenarios.This research not only provides a novel solution to resource allocation challenges in V2X networks but also advances the application of DDQN in intelligent transportation systems,offering substantial theoretical significance and practical value. 展开更多
关键词 Resource allocation V2X double deep q-network graph neural network
在线阅读 下载PDF
Convolutional Neural Network-Based Deep Q-Network (CNN-DQN) Resource Management in Cloud Radio Access Network 被引量:3
2
作者 Amjad Iqbal Mau-Luen Tham Yoong Choon Chang 《China Communications》 SCIE CSCD 2022年第10期129-142,共14页
The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a promi... The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a prominent framework in the 5G mobile network to meet the above requirements by deploying low-cost and intelligent multiple distributed antennas known as remote radio heads (RRHs). However, achieving the optimal resource allocation (RA) in CRAN using the traditional approach is still challenging due to the complex structure. In this paper, we introduce the convolutional neural network-based deep Q-network (CNN-DQN) to balance the energy consumption and guarantee the user quality of service (QoS) demand in downlink CRAN. We first formulate the Markov decision process (MDP) for energy efficiency (EE) and build up a 3-layer CNN to capture the environment feature as an input state space. We then use DQN to turn on/off the RRHs dynamically based on the user QoS demand and energy consumption in the CRAN. Finally, we solve the RA problem based on the user constraint and transmit power to guarantee the user QoS demand and maximize the EE with a minimum number of active RRHs. In the end, we conduct the simulation to compare our proposed scheme with nature DQN and the traditional approach. 展开更多
关键词 energy efficiency(EE) markov decision process(MDP) convolutional neural network(CNN) cloud RAN deep q-network(DQN)
在线阅读 下载PDF
Multi-Agent Path Planning Method Based on Improved Deep Q-Network in Dynamic Environments 被引量:2
3
作者 LI Shuyi LI Minzhe JING Zhongliang 《Journal of Shanghai Jiaotong university(Science)》 EI 2024年第4期601-612,共12页
The multi-agent path planning problem presents significant challenges in dynamic environments,primarily due to the ever-changing positions of obstacles and the complex interactions between agents’actions.These factor... The multi-agent path planning problem presents significant challenges in dynamic environments,primarily due to the ever-changing positions of obstacles and the complex interactions between agents’actions.These factors contribute to a tendency for the solution to converge slowly,and in some cases,diverge altogether.In addressing this issue,this paper introduces a novel approach utilizing a double dueling deep Q-network(D3QN),tailored for dynamic multi-agent environments.A novel reward function based on multi-agent positional constraints is designed,and a training strategy based on incremental learning is performed to achieve collaborative path planning of multiple agents.Moreover,the greedy and Boltzmann probability selection policy is introduced for action selection and avoiding convergence to local extremum.To match radar and image sensors,a convolutional neural network-long short-term memory(CNN-LSTM)architecture is constructed to extract the feature of multi-source measurement as the input of the D3QN.The algorithm’s efficacy and reliability are validated in a simulated environment,utilizing robot operating system and Gazebo.The simulation results show that the proposed algorithm provides a real-time solution for path planning tasks in dynamic scenarios.In terms of the average success rate and accuracy,the proposed method is superior to other deep learning algorithms,and the convergence speed is also improved. 展开更多
关键词 MULTI-AGENT path planning deep reinforcement learning deep q-network
原文传递
Manufacturing Resource Scheduling Based on Deep Q-Network 被引量:1
4
作者 ZHANG Yufei Zou Yuanhao ZHAO Xiaodong 《Wuhan University Journal of Natural Sciences》 CAS CSCD 2022年第6期531-538,共8页
To optimize machine allocation and task dispatching in smart manufacturing factories, this paper proposes a manufacturing resource scheduling framework based on reinforcement learning(RL). The framework formulates the... To optimize machine allocation and task dispatching in smart manufacturing factories, this paper proposes a manufacturing resource scheduling framework based on reinforcement learning(RL). The framework formulates the entire scheduling process as a multi-stage sequential decision problem, and further obtains the scheduling order by the combination of deep convolutional neural network(CNN) and improved deep Q-network(DQN). Specifically, with respect to the representation of the Markov decision process(MDP), the feature matrix is considered as the state space and a set of heuristic dispatching rules are denoted as the action space. In addition, the deep CNN is employed to approximate the state-action values, and the double dueling deep Qnetwork with prioritized experience replay and noisy network(D3QPN2) is adopted to determine the appropriate action according to the current state. In the experiments, compared with the traditional heuristic method, the proposed method is able to learn high-quality scheduling policy and achieve shorter makespan on the standard public datasets. 展开更多
关键词 smart manufacturing job shop scheduling convolutional neural network deep q-network
原文传递
Reinforcement Learning with an Ensemble of Binary Action Deep Q-Networks 被引量:1
5
作者 A.M.Hafiz M.Hassaballah +2 位作者 Abdullah Alqahtani Shtwai Alsubai Mohamed Abdel Hameed 《Computer Systems Science & Engineering》 SCIE EI 2023年第9期2651-2666,共16页
With the advent of Reinforcement Learning(RL)and its continuous progress,state-of-the-art RL systems have come up for many challenging and real-world tasks.Given the scope of this area,various techniques are found in ... With the advent of Reinforcement Learning(RL)and its continuous progress,state-of-the-art RL systems have come up for many challenging and real-world tasks.Given the scope of this area,various techniques are found in the literature.One such notable technique,Multiple Deep Q-Network(DQN)based RL systems use multiple DQN-based-entities,which learn together and communicate with each other.The learning has to be distributed wisely among all entities in such a scheme and the inter-entity communication protocol has to be carefully designed.As more complex DQNs come to the fore,the overall complexity of these multi-entity systems has increased many folds leading to issues like difficulty in training,need for high resources,more training time,and difficulty in fine-tuning leading to performance issues.Taking a cue from the parallel processing found in the nature and its efficacy,we propose a lightweight ensemble based approach for solving the core RL tasks.It uses multiple binary action DQNs having shared state and reward.The benefits of the proposed approach are overall simplicity,faster convergence and better performance compared to conventional DQN based approaches.The approach can potentially be extended to any type of DQN by forming its ensemble.Conducting extensive experimentation,promising results are obtained using the proposed ensemble approach on OpenAI Gym tasks,and Atari 2600 games as compared to recent techniques.The proposed approach gives a stateof-the-art score of 500 on the Cartpole-v1 task,259.2 on the LunarLander-v2 task,and state-of-the-art results on four out of five Atari 2600 games. 展开更多
关键词 Deep q-networks ensemble learning reinforcement learning OpenAI Gym environments
在线阅读 下载PDF
Walking Stability Control Method for Biped Robot on Uneven Ground Based on Deep Q-Network
6
作者 Baoling Han Yuting Zhao Qingsheng Luo 《Journal of Beijing Institute of Technology》 EI CAS 2019年第3期598-605,共8页
A gait control method for a biped robot based on the deep Q-network (DQN) algorithm is proposed to enhance the stability of walking on uneven ground. This control strategy is an intelligent learning method of posture ... A gait control method for a biped robot based on the deep Q-network (DQN) algorithm is proposed to enhance the stability of walking on uneven ground. This control strategy is an intelligent learning method of posture adjustment. A robot is taken as an agent and trained to walk steadily on an uneven surface with obstacles, using a simple reward function based on forward progress. The reward-punishment (RP) mechanism of the DQN algorithm is established after obtaining the offline gait which was generated in advance foot trajectory planning. Instead of implementing a complex dynamic model, the proposed method enables the biped robot to learn to adjust its posture on the uneven ground and ensures walking stability. The performance and effectiveness of the proposed algorithm was validated in the V-REP simulation environment. The results demonstrate that the biped robot's lateral tile angle is less than 3° after implementing the proposed method and the walking stability is obviously improved. 展开更多
关键词 DEEP q-network (DQN) BIPED robot uneven ground WALKING STABILITY gait control
在线阅读 下载PDF
Multi-Agent Deep Q-Networks for Efficient Edge Federated Learning Communications in Software-Defined IoT
7
作者 Prohim Tam Sa Math +1 位作者 Ahyoung Lee Seokhoon Kim 《Computers, Materials & Continua》 SCIE EI 2022年第5期3319-3335,共17页
Federated learning(FL)activates distributed on-device computation techniques to model a better algorithm performance with the interaction of local model updates and global model distributions in aggregation averaging ... Federated learning(FL)activates distributed on-device computation techniques to model a better algorithm performance with the interaction of local model updates and global model distributions in aggregation averaging processes.However,in large-scale heterogeneous Internet of Things(IoT)cellular networks,massive multi-dimensional model update iterations and resource-constrained computation are challenging aspects to be tackled significantly.This paper introduces the system model of converging softwaredefined networking(SDN)and network functions virtualization(NFV)to enable device/resource abstractions and provide NFV-enabled edge FL(eFL)aggregation servers for advancing automation and controllability.Multi-agent deep Q-networks(MADQNs)target to enforce a self-learning softwarization,optimize resource allocation policies,and advocate computation offloading decisions.With gathered network conditions and resource states,the proposed agent aims to explore various actions for estimating expected longterm rewards in a particular state observation.In exploration phase,optimal actions for joint resource allocation and offloading decisions in different possible states are obtained by maximum Q-value selections.Action-based virtual network functions(VNF)forwarding graph(VNFFG)is orchestrated to map VNFs towards eFL aggregation server with sufficient communication and computation resources in NFV infrastructure(NFVI).The proposed scheme indicates deficient allocation actions,modifies the VNF backup instances,and reallocates the virtual resource for exploitation phase.Deep neural network(DNN)is used as a value function approximator,and epsilongreedy algorithm balances exploration and exploitation.The scheme primarily considers the criticalities of FL model services and congestion states to optimize long-term policy.Simulation results presented the outperformance of the proposed scheme over reference schemes in terms of Quality of Service(QoS)performance metrics,including packet drop ratio,packet drop counts,packet delivery ratio,delay,and throughput. 展开更多
关键词 Deep q-networks federated learning network functions virtualization quality of service software-defined networking
在线阅读 下载PDF
Transformer-Aided Deep Double Dueling Spatial-Temporal Q-Network for Spatial Crowdsourcing Analysis
8
作者 Yu Li Mingxiao Li +2 位作者 Dongyang Ou Junjie Guo Fangyuan Pan 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第4期893-909,共17页
With the rapid development ofmobile Internet,spatial crowdsourcing has becomemore andmore popular.Spatial crowdsourcing consists of many different types of applications,such as spatial crowd-sensing services.In terms ... With the rapid development ofmobile Internet,spatial crowdsourcing has becomemore andmore popular.Spatial crowdsourcing consists of many different types of applications,such as spatial crowd-sensing services.In terms of spatial crowd-sensing,it collects and analyzes traffic sensing data from clients like vehicles and traffic lights to construct intelligent traffic prediction models.Besides collecting sensing data,spatial crowdsourcing also includes spatial delivery services like DiDi and Uber.Appropriate task assignment and worker selection dominate the service quality for spatial crowdsourcing applications.Previous research conducted task assignments via traditional matching approaches or using simple network models.However,advanced mining methods are lacking to explore the relationship between workers,task publishers,and the spatio-temporal attributes in tasks.Therefore,in this paper,we propose a Deep Double Dueling Spatial-temporal Q Network(D3SQN)to adaptively learn the spatialtemporal relationship between task,task publishers,and workers in a dynamic environment to achieve optimal allocation.Specifically,D3SQNis revised through reinforcement learning by adding a spatial-temporal transformer that can estimate the expected state values and action advantages so as to improve the accuracy of task assignments.Extensive experiments are conducted over real data collected fromDiDi and ELM,and the simulation results verify the effectiveness of our proposed models. 展开更多
关键词 Historical behavior analysis spatial crowdsourcing deep double dueling q-networks
在线阅读 下载PDF
UAV Autonomous Navigation for Wireless Powered Data Collection with Onboard Deep Q-Network
9
作者 LI Yuting DING Yi +3 位作者 GAO Jiangchuan LIU Yusha HU Jie YANG Kun 《ZTE Communications》 2023年第2期80-87,共8页
In a rechargeable wireless sensor network,utilizing the unmanned aerial vehicle(UAV)as a mobile base station(BS)to charge sensors and collect data effectively prolongs the network’s lifetime.In this paper,we jointly ... In a rechargeable wireless sensor network,utilizing the unmanned aerial vehicle(UAV)as a mobile base station(BS)to charge sensors and collect data effectively prolongs the network’s lifetime.In this paper,we jointly optimize the UAV’s flight trajectory and the sensor selection and operation modes to maximize the average data traffic of all sensors within a wireless sensor network(WSN)during finite UAV’s flight time,while ensuring the energy required for each sensor by wireless power transfer(WPT).We consider a practical scenario,where the UAV has no prior knowledge of sensor locations.The UAV performs autonomous navigation based on the status information obtained within the coverage area,which is modeled as a Markov decision process(MDP).The deep Q-network(DQN)is employed to execute the navigation based on the UAV position,the battery level state,channel conditions and current data traffic of sensors within the UAV’s coverage area.Our simulation results demonstrate that the DQN algorithm significantly improves the network performance in terms of the average data traffic and trajectory design. 展开更多
关键词 unmanned aerial vehicle wireless power transfer deep q-network autonomous navigation
在线阅读 下载PDF
Intelligent and efficient fiber allocation strategy based on the dueling-double-deep Q-network
10
作者 Yong ZHANG Zhipeng YUAN +2 位作者 Jia DING Feng GUO Junyang JIN 《Frontiers of Engineering Management》 2025年第4期721-735,共15页
Fiber allocation in optical cable production is critical for optimizing production efficiency,product quality,and inventory management.However,factors like fiber length and storage time complicate this process,making ... Fiber allocation in optical cable production is critical for optimizing production efficiency,product quality,and inventory management.However,factors like fiber length and storage time complicate this process,making heuristic optimization algorithms inadequate.To tackle these challenges,this paper proposes a new framework:the dueling-double-deep Q-network with twin state-value and action-advantage functions (D3QNTF).First,dual action-advantage and state-value functions are used to prevent overestimation of action values.Second,a method for random initialization of feasible solutions improves sample quality early in the optimization.Finally,a strict penalty for errors is added to the reward mechanism,making the agent more sensitive to and better at avoiding illegal actions,which reduces decision errors.Experimental results show that the proposed method outperforms state-of-the-art algorithms,including greedy algorithms,genetic algorithms,deep Q-networks,double deep Q-networks,and standard dueling-double-deep Q-networks.The findings highlight the potential of the D3QNTF framework for fiber allocation in optical cable production. 展开更多
关键词 optical fiber allocation deep reinforcement learning dueling-double-deep q-network dual action-advantage and state-value functions feasible solutions
原文传递
基于视觉感知与深度Q网络的循环水养殖智能投喂方法
11
作者 黄海燕 崔丹 +4 位作者 叶佳鑫 周月明 陈猷鹏 杨军超 申渝 《南方水产科学》 北大核心 2026年第2期37-49,共13页
为实现循环水养殖系统(Recirculating aquaculture system,RAS)中投喂策略的精准化与自适应调控,解决因投喂策略静态化导致的饲料利用率低、生长调控粗放等问题,研究提出了一种融合视觉感知与深度Q网络(Deep Q-Network,DQN)的智能投喂... 为实现循环水养殖系统(Recirculating aquaculture system,RAS)中投喂策略的精准化与自适应调控,解决因投喂策略静态化导致的饲料利用率低、生长调控粗放等问题,研究提出了一种融合视觉感知与深度Q网络(Deep Q-Network,DQN)的智能投喂方法。该方法以淡水石斑鱼(Cichlasoma managuense)为实验对象,通过YOLOv8与DeepSORT实现鱼体运动速度的实时跟踪,并结合灰度共生矩阵提取的纹理特征,构建鱼群摄食强度量化指标;进而将摄食强度、水温、溶解氧及期望摄食强度共同作为状态输入,设计多目标奖励函数,采用深度Q网络训练决策模型,形成闭环控制系统。实验结果表明,目标检测平均精度均值(mAP@.5)达85.3%,决策模型在个体平均总投喂量仅378.4 g的条件下,使鱼群体质量增长率(Weight gain rate,WGR)升至54.38%,饲料转化率(Feed conversion ratio,FCR)降至1.09,性能显著优于传统投喂方式。本方法有效实现了摄食行为的实时感知与投喂策略的动态优化,为循环水养殖的精细化管理提供了可靠的技术路径。 展开更多
关键词 视觉感知 深度Q网络 智能投喂 循环水养殖
在线阅读 下载PDF
基于改进深度Q网络的智能网联汽车路径规划
12
作者 文家燕 王怡博 +1 位作者 辛华健 谢广明 《智能系统学报》 北大核心 2026年第1期226-235,共10页
针对非结构环境中的智能网联汽车路径规划问题,传统的深度Q网络(deep Q-network,DQN)算法存在规划效率低、收敛速度慢、泛化性差等问题,本文提出了一种结合注意力机制和经验分类的DQN规划方法。通过结合注意力机制设计经验回放池,通过... 针对非结构环境中的智能网联汽车路径规划问题,传统的深度Q网络(deep Q-network,DQN)算法存在规划效率低、收敛速度慢、泛化性差等问题,本文提出了一种结合注意力机制和经验分类的DQN规划方法。通过结合注意力机制设计经验回放池,通过动态权重分配解决多目标优化冲突,提升相似环境中的经验利用率,降低规划时间,加快收敛;构建非稀疏奖励约束,结合交通环境特性优化状态空间,以便适应多目标场景和实现多场景泛化。仿真表明,优化后的算法平均规划速度提升了28.6%,行进路程较优化前缩短了25.2%,且在不同场景下通过载入训练数据,首次规划成功的耗时缩短了32.8%。 展开更多
关键词 智能网联汽车 路径规划 非结构化环境 注意力机制 经验回放 避障 深度Q网络 深度强化学习
在线阅读 下载PDF
基于改进深度Q网络的异构无人机快速任务分配
13
作者 王月海 邱国帅 +3 位作者 邢娜 赵欣怡 王婕 韩曦 《工程科学学报》 北大核心 2026年第1期142-151,共10页
随着无人机技术的快速发展,多无人机系统在执行复杂任务时展现出巨大潜力,高效的任务分配策略对提升多无人机系统的整体性能至关重要.然而,传统方法如集中式优化、拍卖算法及鸽群算法等,在面对复杂环境干扰时往往难以生成有效的分配策略... 随着无人机技术的快速发展,多无人机系统在执行复杂任务时展现出巨大潜力,高效的任务分配策略对提升多无人机系统的整体性能至关重要.然而,传统方法如集中式优化、拍卖算法及鸽群算法等,在面对复杂环境干扰时往往难以生成有效的分配策略,为此,本文考虑了环境不确定性如不同风速和降雨量,重点研究了改进的强化学习算法在无人机任务分配中的应用,使多无人机系统能够迅速响应并实现资源的高效利用.首先,本文将无人机任务分配问题建模为马尔可夫决策过程,通过神经网络进行策略逼近用以任务分配中高效处理高维和复杂的状态空间,同时引入优先经验重放机制,有效降低了在线计算的负担.仿真结果表明,与其他强化学习方法相比,该算法具有较强的收敛性.在面对复杂环境时,其鲁棒性更为显著.此外,该算法在处理不同任务时仅需0.24 s即可完成一组适合的无人机分配,并能够快速生成大规模无人机集群的任务分配方案. 展开更多
关键词 无人机群 任务分配 强化学习 深度Q网络 马尔可夫决策过程
在线阅读 下载PDF
基于类加权迁移深度Q网络策略的分层数据存储框架
14
作者 冯刚 《计算机应用与软件》 北大核心 2026年第2期387-397,共11页
为了实现数据管理的高效性和适用能力,提出一种基于类加权迁移深度Q网络策略的分层数据存储框架。为了提升其在不同存储解决方案之间进行在线动态数据传输的能力,以及根据数据访问模式和可用性做出决策的能力,引入类加权迁移深度Q网络... 为了实现数据管理的高效性和适用能力,提出一种基于类加权迁移深度Q网络策略的分层数据存储框架。为了提升其在不同存储解决方案之间进行在线动态数据传输的能力,以及根据数据访问模式和可用性做出决策的能力,引入类加权迁移深度Q网络策略来解决分层存储系统中的数据迁移问题,同时忽略源异常值,有效激励了正知识的转移,提升域自适应的效果。最后设计了一个仿真软件和一个云框架进行试验测试,结果证明了提出方法的高效性和自适应能力。 展开更多
关键词 深度Q网络 动态迁移策略 分层数据存储 云框架
在线阅读 下载PDF
基于深度Q网络的分布式车辆路径规划策略研究
15
作者 文家燕 王怡博 景永年 《工业控制计算机》 2026年第2期94-96,共3页
针对智能网联汽车在非结构化环境中路径规划存在学习效率低、泛化能力不足、数据隐私安全性不足等问题,提出一种基于深度Q网络(Deep Q-Network,DQN)的分布式车辆路径规划方法。设计了动态权重分配算法,结合本地数据归一化技术,对多源车... 针对智能网联汽车在非结构化环境中路径规划存在学习效率低、泛化能力不足、数据隐私安全性不足等问题,提出一种基于深度Q网络(Deep Q-Network,DQN)的分布式车辆路径规划方法。设计了动态权重分配算法,结合本地数据归一化技术,对多源车辆模型训练参数进行整合,平衡模型性能的同时保护隐私安全。引入云端共享经验池机制,消除跨场景数据分布偏移,加速全局模型收敛。仿真实验表明,相较于传统DQN算法,提出的算法在环境中路径规划成功率提升显著,模型达到成功率90%时所需的训练回合数减少77.6%。 展开更多
关键词 深度Q网络 路径规划 联邦学习 归一化
在线阅读 下载PDF
Energy Optimization for Autonomous Mobile Robot Path Planning Based on Deep Reinforcement Learning
16
作者 Longfei Gao Weidong Wang Dieyun Ke 《Computers, Materials & Continua》 2026年第1期984-998,共15页
At present,energy consumption is one of the main bottlenecks in autonomous mobile robot development.To address the challenge of high energy consumption in path planning for autonomous mobile robots navigating unknown ... At present,energy consumption is one of the main bottlenecks in autonomous mobile robot development.To address the challenge of high energy consumption in path planning for autonomous mobile robots navigating unknown and complex environments,this paper proposes an Attention-Enhanced Dueling Deep Q-Network(ADDueling DQN),which integrates a multi-head attention mechanism and a prioritized experience replay strategy into a Dueling-DQN reinforcement learning framework.A multi-objective reward function,centered on energy efficiency,is designed to comprehensively consider path length,terrain slope,motion smoothness,and obstacle avoidance,enabling optimal low-energy trajectory generation in 3D space from the source.The incorporation of a multihead attention mechanism allows the model to dynamically focus on energy-critical state features—such as slope gradients and obstacle density—thereby significantly improving its ability to recognize and avoid energy-intensive paths.Additionally,the prioritized experience replay mechanism accelerates learning from key decision-making experiences,suppressing inefficient exploration and guiding the policy toward low-energy solutions more rapidly.The effectiveness of the proposed path planning algorithm is validated through simulation experiments conducted in multiple off-road scenarios.Results demonstrate that AD-Dueling DQN consistently achieves the lowest average energy consumption across all tested environments.Moreover,the proposed method exhibits faster convergence and greater training stability compared to baseline algorithms,highlighting its global optimization capability under energy-aware objectives in complex terrains.This study offers an efficient and scalable intelligent control strategy for the development of energy-conscious autonomous navigation systems. 展开更多
关键词 Autonomous mobile robot deep reinforcement learning energy optimization multi-attention mechanism prioritized experience replay dueling deep q-network
在线阅读 下载PDF
基于深度强化学习决策的雷达干扰抑制方法
17
作者 肖易寒 孟祥乾 陆钱融 《制导与引信》 2026年第1期22-31,共10页
针对目前雷达干扰抑制决策智能化程度低的问题,提出了一种基于双深度优先经验回放和可变贪婪算法改进的双重竞争深度Q网络(double dueling deep Q network,D3QN)决策的雷达干扰抑制方法。首先对雷达目标回波和干扰混合信号进行特征提取... 针对目前雷达干扰抑制决策智能化程度低的问题,提出了一种基于双深度优先经验回放和可变贪婪算法改进的双重竞争深度Q网络(double dueling deep Q network,D3QN)决策的雷达干扰抑制方法。首先对雷达目标回波和干扰混合信号进行特征提取;然后根据信号特征通过可变贪婪算法选择动作作用于干扰,并将动作前后的信号特征存储于双深度优先经验回放池后,经过学习决策出最优的干扰抑制策略;最后使用该策略抑制干扰后输出。实验结果表明,该方法有效改善了信号的脉压结果,显著提升了信号的信干噪比,相较于基于D3QN的传统干扰抑制方法,在策略准确率和收敛速度上分别提升了7.3%和8.7%。 展开更多
关键词 雷达干扰抑制 双重竞争深度Q网络 双深度优先经验回放 可变贪婪算法 脉冲压缩
在线阅读 下载PDF
DDQN-Based 3D Path Planning Algorithm for UAVs in Dynamic Dense Obstacle Environments
18
作者 Wenjie Zhang Meng Yu Yin Wang 《Journal of Beijing Institute of Technology》 2026年第1期84-96,共13页
Online three-dimensional(3D)path planning in dynamic environments is a fundamental problem for achieving autonomous navigation of unmanned aerial vehicles(UAVs).However,existing methods struggle to model traversable d... Online three-dimensional(3D)path planning in dynamic environments is a fundamental problem for achieving autonomous navigation of unmanned aerial vehicles(UAVs).However,existing methods struggle to model traversable dynamic gaps,resulting in conservative and suboptimal trajectories.To address these challenges,this paper proposes a hierarchical reinforcement learning(RL)framework that integrates global path guidance,local trajectory generation,predictive safety evaluation,and neural network-based decision-making.Specifically,the global planner provides long-term navigation guidance,and the local module then utilizes an improved 3D dynamic window approach(DWA)to generate dynamically feasible candidate trajectories.To enhance safety in dense dynamic scenarios,the algorithm introduces a predictive axis-aligned bounding box(AABB)strategy to model the future occupancy of obstacles,combined with convex hull verification for efficient trajectory safety assessment.Furthermore,a double deep Q-network(DDQN)is employed with structured feature encoding,enabling the neural network to reliably select the optimal trajectory from the candidate set,thereby improving robustness and generalization.Comparative experiments conducted in a high-fidelity simulation environment show that the algorithm outperforms existing algorithms,reducing the average number of collisions to 0.2 while shortening the average task completion time by approximately 15%,and achieving a success rate of 97%. 展开更多
关键词 unmanned aerial vehicle(UAV)three-dimensional(3D)path planning 3D dynamic window approach(DWA) predictive axis-aligned bounding box(AABB) double deep q-network(DDQN) autonomous navigation
在线阅读 下载PDF
基于网络药理学、分子对接与动物实验验证探讨花椒提取物外用干预特应性皮炎的作用机制
19
作者 黎量 李雨恒 +2 位作者 彭长恩 王蛟龙 吴纯洁 《中药药理与临床》 北大核心 2026年第1期92-101,共10页
目的:基于网络药理学、分子对接结合体内验证的方法探究花椒提取物外用对特应性皮炎(atopic dermatitis,AD)小鼠的作用机制。方法:利用UPLC-Q-Orbitrap HRMS技术对花椒提取物的化学成分进行快速识别和鉴定;通过Swiss Target Prediction... 目的:基于网络药理学、分子对接结合体内验证的方法探究花椒提取物外用对特应性皮炎(atopic dermatitis,AD)小鼠的作用机制。方法:利用UPLC-Q-Orbitrap HRMS技术对花椒提取物的化学成分进行快速识别和鉴定;通过Swiss Target Prediction、SEA平台预测花椒提取物成分的潜在靶点,通过GeneCards和Drugbank数据库收集AD相关靶点,运用String数据库构建关键靶点PPI网络图;基于Metascape数据库进行KEGG分析,利用Cytoscape软件构建药物-成分-靶点-疾病-通路网络图;采用Discovery Studio的CDOCKER模块进行分子对接。构建AD模型小鼠,给予外用花椒提取物0.02、0.03 g/kg进行药理实验验证,观察AD小鼠抓挠次数、计算耳厚度差和脾脏指数;ELISA法检测小鼠皮肤白介素-6(IL-6)、IL-13、IL-31、肿瘤坏死因子α(TNF-α)和血清免疫球蛋白E(IgE)含量;HE染色法观察皮肤组织病理;免疫组化法检测皮肤组织TRPV1和TRPV3的阳性表达。结果:鉴定出花椒提取物中22个化学成分,得到成分相关靶点194个,疾病相关靶点共1680个,交集靶点68个;通过PPI网络筛选获得核心靶点29个;KEGG通路富集分析发现,TRP通道的炎性介质调节通路可能是花椒提取物作用于AD的主要信号通路;药物-成分-靶点-疾病-通路网络图显示花椒提取物中羟基-ε-山椒素、羟基-α-山椒素、羟基-β-山椒素排名靠前,可能是治疗AD的重要成分。瞬时电位离子通道(TRP通道)中的TRPV1、TRPA1排名靠前,提示其是治疗疾病的重要靶点。分子对接表明,羟基-ε-山椒素、羟基-α-山椒素、羟基-β-山椒素与TRPV1、TRPV3有良好的结合亲和力。动物实验结果显示,与模型对照组比较,花椒提取物外用各剂量组AD小鼠皮损评分降低(P<0.01),AD小鼠抓挠次数、耳厚度差和脾脏指数降低(P<0.01),AD小鼠皮肤IL-6、IL-13、IL-31含量降低(P<0.01),花椒提取物0.03 g/kg组皮肤TNF-α含量和血清IgE含量明显降低(P<0.05),病理性增厚和皮肤炎性细胞浸润明显减轻,TRPV1和TRPV3的阳性表达显著下调(P<0.01)。结论:花椒提取物外用可以有效缓解AD的瘙痒和炎症,其作用机制可能是抑制TRPV1、TRPV3通道的活性或下调TRPV1、TRPV3的表达。 展开更多
关键词 花椒提取物 特应性皮炎 超高效液相色谱-四极杆-静电场轨道阱高分辨质谱 网络药理学 分子对接 瞬时受体电位通道
原文传递
LGDLight:线图注意力驱动的区域交通信号控制
20
作者 胡援浩 娄铮铮 +3 位作者 胡世哲 张朝阳 焦重阳 周清雷 《计算机工程与应用》 北大核心 2026年第1期379-390,共12页
针对深度强化学习在区域交通信号控制优化时,多智能体间状态传递过程中所呈现的复杂动态问题,采用线图结构来描述复杂交通网络,使用图注意力网络(graph attention network,GAT)建模路口间动态关系,并将其融入决斗式双深度Q网络(dueling ... 针对深度强化学习在区域交通信号控制优化时,多智能体间状态传递过程中所呈现的复杂动态问题,采用线图结构来描述复杂交通网络,使用图注意力网络(graph attention network,GAT)建模路口间动态关系,并将其融入决斗式双深度Q网络(dueling double deep Q network,D3QN)中,提出线图注意力Q网络交通信号控制方法(line graph deep Q network light,LGDLight)。LGDLight将道路作为节点,路网视为线图,利用图注意力网络来捕捉复杂道路节点间的动态信息交互,增强D3QN对复杂交通状态的感知能力。在仿真环境下的真实车流数据测试实验中,LGDLight相较于其他方法表现出更好的性能,尤其是在车流量较大及路网场景较为复杂情况下所表现出的性能更佳。进一步的实验分析不同奖励机制的影响,结果表明以排队长度作为奖励机制时,该方法能更有效地优化交通信号控制。 展开更多
关键词 交通信号控制 深度强化学习 决斗式双深度Q网络(D3QN) 线图 图注意力网络
在线阅读 下载PDF
上一页 1 2 62 下一页 到第
使用帮助 返回顶部