期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Timely Updates in MEC-Assisted Status Update Systems: Joint Task Generation and Computation Offloading Scheme 被引量:3
1
作者 Long Liu Xiaoqi Qin +1 位作者 Yunzheng Tao Zhi Zhang 《China Communications》 SCIE CSCD 2020年第8期168-186,共19页
Fresh status updates are vital to the efficient operation of network monitoring and real-time control applications. In this paper, we consider a mobile edge computing(MEC)-assisted status update system, where smart de... Fresh status updates are vital to the efficient operation of network monitoring and real-time control applications. In this paper, we consider a mobile edge computing(MEC)-assisted status update system, where smart devices extract valuable status updates from sensed data to achieve timely awareness of the surroundings by exploiting computational resources at the device and edge server. To quantify the freshness of status updates obtained by executing computation tasks, we employ the concept of age of information(Ao I) to characterize the timeliness of status updates. To cope with the limited energy at devices, we investigate a joint task generation and computation offloading scheme under a given energy budget for minimizing the age of obtained status updates. The age minimization problem is modeled as a constrained Markov decision process(CMDP). To obtain the optimal policy, we derive the structural properties of the optimal deterministic policy and propose a lightweight structure-based status update algorithm in the case of known channel statistics. Moreover, we consider a more realistic scenario without prior knowledge of channel statistics, and propose a Q-learning-based status update algorithm to make online decisions. Simulation results show that the performance of our proposed algorithms is competitive when compared with existing schemes. 展开更多
关键词 mobile edge computing age of information constrained markov decision process structural analysis Q-LEARNING
在线阅读 下载PDF
Multi-constraint reinforcement learning in complex robot environments
2
作者 Sheng HAN Hengrui ZHANG +2 位作者 Hao WU Youfang LIN Kai LV 《Frontiers of Computer Science》 2025年第8期105-107,共3页
1 Introduction Constrained Reinforcement Learning(CRL),modeled as a Constrained Markov Decision Process(CMDP)[1,2],is commonly used to address applications with security restrictions.Previous works[3]primarily focused... 1 Introduction Constrained Reinforcement Learning(CRL),modeled as a Constrained Markov Decision Process(CMDP)[1,2],is commonly used to address applications with security restrictions.Previous works[3]primarily focused on the single-constraint issue,overlooking the more common multi-constraint setting which involves extensive computations and combinatorial optimization of multiple Lagrange multipliers. 展开更多
关键词 constrained reinforcement learning combinatorial optimization multiple lagrange multipliers constrained markov decision process complex robot environments constrained reinforcement learning crl modeled constrained markov decision process cmdp multi constraint lagrange multipliers
原文传递
Partially Distributed Channel and Power Management Based on Reinforcement Learning
3
作者 Zhiwei Jiang Caiyong Hao +2 位作者 Yang Huang Qihui Wu Fuhui Zhou 《Journal of Communications and Information Networks》 CSCD 2020年第4期423-437,共15页
This paper studies a dynamic multi-user wireless network,where users have no knowledge of the arrival rate and size of data block and suffer from a constraint on long-term average power consumption.Considering such a ... This paper studies a dynamic multi-user wireless network,where users have no knowledge of the arrival rate and size of data block and suffer from a constraint on long-term average power consumption.Considering such a network,we address the problem of dynamically optimizing channel/power allocation,so as to minimize the long-term average data backlog.The design problem is shown to be a constrained Markov decision process.In order to solve the problem without knowledge on dynamics of the system,we introduce post-decision states and propose a resource allocation algorithm based on reinforcement learning.Since the channel/power allocation problem is coupled,the multiuser decision problem suffers from curses of dimensions(of state/action/outcome space).This makes centralized decision-making and optimization on channel/power allocation suffer from a long convergence time.As a countermeasure,a partially distributed resource allocation framework is proposed.The multiuser power allocation problem is decoupled into single-user decision problems,while channel allocation optimization is performed in a centralized manner.In order to further reduce computational complexity,we propose a low-complexity reinforcement learning method.Simulation results reveal that the proposed algorithm outperforms the state-of-the-art myopic optimizations in terms of energy efficiency and the backlog performance. 展开更多
关键词 constrained markov decision processes multi-user optimization reinforcement learning the Internet of things
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部