期刊文献+
共找到149篇文章
< 1 2 8 >
每页显示 20 50 100
Call for papers Journal of Control Theory and Applications Special issue on Approximate dynamic programming and reinforcement learning
1
《控制理论与应用(英文版)》 EI 2010年第2期257-257,共1页
Approximate dynamic programming (ADP) is a general and effective approach for solving optimal control and estimation problems by adapting to uncertain and nonconvex environments over time.
关键词 Call for papers Journal of control Theory and Applications Special issue on approximate dynamic programming and reinforcement learning
在线阅读 下载PDF
Approximate Dynamic Programming for Self-Learning Control 被引量:14
2
作者 DerongLiu 《自动化学报》 EI CSCD 北大核心 2005年第1期13-18,共6页
This paper introduces a self-learning control approach based on approximate dynamic programming. Dynamic programming was introduced by Bellman in the 1950's for solving optimal control problems of nonlinear dynami... This paper introduces a self-learning control approach based on approximate dynamic programming. Dynamic programming was introduced by Bellman in the 1950's for solving optimal control problems of nonlinear dynamical systems. Due to its high computational complexity, the applications of dynamic programming have been limited to simple and small problems. The key step in finding approximate solutions to dynamic programming is to estimate the performance index in dynamic programming. The optimal control signal can then be determined by minimizing (or maximizing) the performance index. Artificial neural networks are very efficient tools in representing the performance index in dynamic programming. This paper assumes the use of neural networks for estimating the performance index in dynamic programming and for generating optimal control signals, thus to achieve optimal control through self-learning. 展开更多
关键词 近似动态程序 自学习控制 神经网络 人工智能
在线阅读 下载PDF
Heuristic dynamic programming-based learning control for discrete-time disturbed multi-agent systems
3
作者 Yao Zhang Chaoxu Mu +1 位作者 Yong Zhang Yanghe Feng 《Control Theory and Technology》 EI CSCD 2021年第3期339-353,共15页
Owing to extensive applications in many fields,the synchronization problem has been widely investigated in multi-agent systems.The synchronization for multi-agent systems is a pivotal issue,which means that under the ... Owing to extensive applications in many fields,the synchronization problem has been widely investigated in multi-agent systems.The synchronization for multi-agent systems is a pivotal issue,which means that under the designed control policy,the output of systems or the state of each agent can be consistent with the leader.The purpose of this paper is to investigate a heuristic dynamic programming(HDP)-based learning tracking control for discrete-time multi-agent systems to achieve synchronization while considering disturbances in systems.Besides,due to the difficulty of solving the coupled Hamilton–Jacobi–Bellman equation analytically,an improved HDP learning control algorithm is proposed to realize the synchronization between the leader and all following agents,which is executed by an action-critic neural network.The action and critic neural network are utilized to learn the optimal control policy and cost function,respectively,by means of introducing an auxiliary action network.Finally,two numerical examples and a practical application of mobile robots are presented to demonstrate the control performance of the HDP-based learning control algorithm. 展开更多
关键词 Multi-agent systems Heuristic dynamic programming(HDP) learning control neural network SYNCHRONIZATION
原文传递
Data-based Fault Tolerant Control for Affine Nonlinear Systems Through Particle Swarm Optimized Neural Networks 被引量:17
4
作者 Haowei Lin Bo Zhao +1 位作者 Derong Liu Cesare Alippi 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2020年第4期954-964,共11页
In this paper, a data-based fault tolerant control(FTC) scheme is investigated for unknown continuous-time(CT)affine nonlinear systems with actuator faults. First, a neural network(NN) identifier based on particle swa... In this paper, a data-based fault tolerant control(FTC) scheme is investigated for unknown continuous-time(CT)affine nonlinear systems with actuator faults. First, a neural network(NN) identifier based on particle swarm optimization(PSO) is constructed to model the unknown system dynamics. By utilizing the estimated system states, the particle swarm optimized critic neural network(PSOCNN) is employed to solve the Hamilton-Jacobi-Bellman equation(HJBE) more efficiently.Then, a data-based FTC scheme, which consists of the NN identifier and the fault compensator, is proposed to achieve actuator fault tolerance. The stability of the closed-loop system under actuator faults is guaranteed by the Lyapunov stability theorem. Finally, simulations are provided to demonstrate the effectiveness of the developed method. 展开更多
关键词 Adaptive dynamic programming(ADP) critic neural network data-based fault tolerant control(FTC) particle swarm optimization(PSO)
在线阅读 下载PDF
Adaptive Multi-Step Evaluation Design With Stability Guarantee for Discrete-Time Optimal Learning Control 被引量:7
5
作者 Ding Wang Jiangyu Wang +2 位作者 Mingming Zhao Peng Xin Junfei Qiao 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2023年第9期1797-1809,共13页
This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge t... This paper is concerned with a novel integrated multi-step heuristic dynamic programming(MsHDP)algorithm for solving optimal control problems.It is shown that,initialized by the zero cost function,MsHDP can converge to the optimal solution of the Hamilton-Jacobi-Bellman(HJB)equation.Then,the stability of the system is analyzed using control policies generated by MsHDP.Also,a general stability criterion is designed to determine the admissibility of the current control policy.That is,the criterion is applicable not only to traditional value iteration and policy iteration but also to MsHDP.Further,based on the convergence and the stability criterion,the integrated MsHDP algorithm using immature control policies is developed to accelerate learning efficiency greatly.Besides,actor-critic is utilized to implement the integrated MsHDP scheme,where neural networks are used to evaluate and improve the iterative policy as the parameter architecture.Finally,two simulation examples are given to demonstrate that the learning effectiveness of the integrated MsHDP scheme surpasses those of other fixed or integrated methods. 展开更多
关键词 Adaptive critic artificial neural networks Hamilton-Jacobi-Bellman(HJB)equation multi-step heuristic dynamic programming multi-step reinforcement learning optimal control
在线阅读 下载PDF
Neural networks-based iterative learning control consensus for periodically time-varying multi-agent systems
6
作者 CHEN JiaXi LI JunMin +1 位作者 CHEN WeiSheng GAO WeiFeng 《Science China(Technological Sciences)》 SCIE EI CAS CSCD 2024年第2期464-474,共11页
In this paper,the problem of adaptive iterative learning based consensus control for periodically time-varying multi-agent systems is studied,in which the dynamics of each follower are driven by nonlinearly parameteri... In this paper,the problem of adaptive iterative learning based consensus control for periodically time-varying multi-agent systems is studied,in which the dynamics of each follower are driven by nonlinearly parameterized terms with periodic disturbances.Neural networks and Fourier base expansions are introduced to describe the periodically time-varying dynamic terms.On this basis,an adaptive learning parameter with a positively convergent series term is constructed,and a distributed control protocol based on local signals between agents is designed to ensure accurate consensus of the closed-loop systems.Furthermore,consensus algorithm is generalized to solve the formation control problem.Finally,simulation experiments are implemented through MATLAB to demonstrate the effectiveness of the method used. 展开更多
关键词 multi-agent systems adaptive iterative learning control nonlinearly parameterized dynamics Fourier series expansion neural networks
原文传递
Online Learning Control for Harmonics Reduction Based on Current Controlled Voltage Source Power Inverters 被引量:3
7
作者 Naresh Malla Ujjwol Tamrakar +2 位作者 Dipesh Shrestha Zhen Ni Reinaldo Tonkoski 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2017年第3期447-457,共11页
Nonlinear loads in the power distribution system cause non-sinusoidal currents and voltages with harmonic components.Shunt active filters(SAF) with current controlled voltage source inverters(CCVSI) are usually used t... Nonlinear loads in the power distribution system cause non-sinusoidal currents and voltages with harmonic components.Shunt active filters(SAF) with current controlled voltage source inverters(CCVSI) are usually used to obtain balanced and sinusoidal source currents by injecting compensation currents.However,CCVSI with traditional controllers have a limited transient and steady state performance.In this paper,we propose an adaptive dynamic programming(ADP) controller with online learning capability to improve transient response and harmonics.The proposed controller works alongside existing proportional integral(PI) controllers to efficiently track the reference currents in the d-q domain.It can generate adaptive control actions to compensate the PI controller.The proposed system was simulated under different nonlinear(three-phase full wave rectifier) load conditions.The performance of the proposed approach was compared with the traditional approach.We have also included the simulation results without connecting the traditional PI control based power inverter for reference comparison.The online learning based ADP controller not only reduced average total harmonic distortion by 18.41%,but also outperformed traditional PI controllers during transients. 展开更多
关键词 Adaptive dynamic programming(ADP) current controlled voltage source power inverter(CCVSI) online learning based controller neural networks shunt active filter(SAF) total harmonic distortion(THD)
在线阅读 下载PDF
Recent Progress in Reinforcement Learning and Adaptive Dynamic Programming for Advanced Control Applications 被引量:11
8
作者 Ding Wang Ning Gao +2 位作者 Derong Liu Jinna Li Frank L.Lewis 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2024年第1期18-36,共19页
Reinforcement learning(RL) has roots in dynamic programming and it is called adaptive/approximate dynamic programming(ADP) within the control community. This paper reviews recent developments in ADP along with RL and ... Reinforcement learning(RL) has roots in dynamic programming and it is called adaptive/approximate dynamic programming(ADP) within the control community. This paper reviews recent developments in ADP along with RL and its applications to various advanced control fields. First, the background of the development of ADP is described, emphasizing the significance of regulation and tracking control problems. Some effective offline and online algorithms for ADP/adaptive critic control are displayed, where the main results towards discrete-time systems and continuous-time systems are surveyed, respectively.Then, the research progress on adaptive critic control based on the event-triggered framework and under uncertain environment is discussed, respectively, where event-based design, robust stabilization, and game design are reviewed. Moreover, the extensions of ADP for addressing control problems under complex environment attract enormous attention. The ADP architecture is revisited under the perspective of data-driven and RL frameworks,showing how they promote ADP formulation significantly.Finally, several typical control applications with respect to RL and ADP are summarized, particularly in the fields of wastewater treatment processes and power systems, followed by some general prospects for future research. Overall, the comprehensive survey on ADP and RL for advanced control applications has d emonstrated its remarkable potential within the artificial intelligence era. In addition, it also plays a vital role in promoting environmental protection and industrial intelligence. 展开更多
关键词 Adaptive dynamic programming(ADP) advanced control complex environment data-driven control event-triggered design intelligent control neural networks nonlinear systems optimal control reinforcement learning(RL)
在线阅读 下载PDF
Adaptable and Dynamic Access Control Decision-Enforcement Approach Based on Multilayer Hybrid Deep Learning Techniques in BYOD Environment
9
作者 Aljuaid Turkea Ayedh M Ainuddin Wahid Abdul Wahab Mohd Yamani Idna Idris 《Computers, Materials & Continua》 SCIE EI 2024年第9期4663-4686,共24页
Organizations are adopting the Bring Your Own Device(BYOD)concept to enhance productivity and reduce expenses.However,this trend introduces security challenges,such as unauthorized access.Traditional access control sy... Organizations are adopting the Bring Your Own Device(BYOD)concept to enhance productivity and reduce expenses.However,this trend introduces security challenges,such as unauthorized access.Traditional access control systems,such as Attribute-Based Access Control(ABAC)and Role-Based Access Control(RBAC),are limited in their ability to enforce access decisions due to the variability and dynamism of attributes related to users and resources.This paper proposes a method for enforcing access decisions that is adaptable and dynamic,based on multilayer hybrid deep learning techniques,particularly the Tabular Deep Neural Network Tabular DNN method.This technique transforms all input attributes in an access request into a binary classification(allow or deny)using multiple layers,ensuring accurate and efficient access decision-making.The proposed solution was evaluated using the Kaggle Amazon access control policy dataset and demonstrated its effectiveness by achieving a 94%accuracy rate.Additionally,the proposed solution enhances the implementation of access decisions based on a variety of resource and user attributes while ensuring privacy through indirect communication with the Policy Administration Point(PAP).This solution significantly improves the flexibility of access control systems,making themmore dynamic and adaptable to the evolving needs ofmodern organizations.Furthermore,it offers a scalable approach to manage the complexities associated with the BYOD environment,providing a robust framework for secure and efficient access management. 展开更多
关键词 BYOD security access control access control decision-enforcement deep learning neural network techniques TabularDNN MULTILAYER dynamic adaptable FLEXIBILITY bottlenecks performance policy conflict
在线阅读 下载PDF
Neural-network-based stochastic linear quadratic optimal tracking control scheme for unknown discrete-time systems using adaptive dynamic programming 被引量:2
10
作者 Xin Chen Fang Wang 《Control Theory and Technology》 EI CSCD 2021年第3期315-327,共13页
In this paper,a stochastic linear quadratic optimal tracking scheme is proposed for unknown linear discrete-time(DT)systems based on adaptive dynamic programming(ADP)algorithm.First,an augmented system composed of the... In this paper,a stochastic linear quadratic optimal tracking scheme is proposed for unknown linear discrete-time(DT)systems based on adaptive dynamic programming(ADP)algorithm.First,an augmented system composed of the original system and the command generator is constructed and then an augmented stochastic algebraic equation is derived based on the augmented system.Next,to obtain the optimal control strategy,the stochastic case is converted into the deterministic one by system transformation,and then an ADP algorithm is proposed with convergence analysis.For the purpose of realizing the ADP algorithm,three back propagation neural networks including model network,critic network and action network are devised to guarantee unknown system model,optimal value function and optimal control strategy,respectively.Finally,the obtained optimal control strategy is applied to the original stochastic system,and two simulations are provided to demonstrate the effectiveness of the proposed algorithm. 展开更多
关键词 Stochastic system Optimal tracking control Adaptive dynamic programming neural networks
原文传递
Parallel Control for Optimal Tracking via Adaptive Dynamic Programming 被引量:25
11
作者 Jingwei Lu Qinglai Wei Fei-Yue Wang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2020年第6期1662-1674,共13页
This paper studies the problem of optimal parallel tracking control for continuous-time general nonlinear systems.Unlike existing optimal state feedback control,the control input of the optimal parallel control is int... This paper studies the problem of optimal parallel tracking control for continuous-time general nonlinear systems.Unlike existing optimal state feedback control,the control input of the optimal parallel control is introduced into the feedback system.However,due to the introduction of control input into the feedback system,the optimal state feedback control methods can not be applied directly.To address this problem,an augmented system and an augmented performance index function are proposed firstly.Thus,the general nonlinear system is transformed into an affine nonlinear system.The difference between the optimal parallel control and the optimal state feedback control is analyzed theoretically.It is proven that the optimal parallel control with the augmented performance index function can be seen as the suboptimal state feedback control with the traditional performance index function.Moreover,an adaptive dynamic programming(ADP)technique is utilized to implement the optimal parallel tracking control using a critic neural network(NN)to approximate the value function online.The stability analysis of the closed-loop system is performed using the Lyapunov theory,and the tracking error and NN weights errors are uniformly ultimately bounded(UUB).Also,the optimal parallel controller guarantees the continuity of the control input under the circumstance that there are finite jump discontinuities in the reference signals.Finally,the effectiveness of the developed optimal parallel control method is verified in two cases. 展开更多
关键词 Adaptive dynamic programming(ADP) nonlinear optimal control parallel controller parallel control theory parallel system tracking control neural network(NN)
在线阅读 下载PDF
An Optimal Control Scheme for a Class of Discrete-time Nonlinear Systems with Time Delays Using Adaptive Dynamic Programming 被引量:17
12
作者 WEI Qing-Lai ZHANG Hua-Guang +1 位作者 LIU De-Rong ZHAO Yan 《自动化学报》 EI CSCD 北大核心 2010年第1期121-129,共9页
关键词 非线性系统 最优控制 控制变量 动态规划
在线阅读 下载PDF
An ADP-based robust control scheme for nonaffine nonlinear systems with uncertainties and input constraints
13
作者 Shijie Luo Kun Zhang Wenchao Xue 《Chinese Physics B》 2025年第6期251-260,共10页
The paper develops a robust control approach for nonaffine nonlinear continuous systems with input constraints and unknown uncertainties. Firstly, this paper constructs an affine augmented system(AAS) within a pre-com... The paper develops a robust control approach for nonaffine nonlinear continuous systems with input constraints and unknown uncertainties. Firstly, this paper constructs an affine augmented system(AAS) within a pre-compensation technique for converting the original nonaffine dynamics into affine dynamics. Secondly, the paper derives a stability criterion linking the original nonaffine system and the auxiliary system, demonstrating that the obtained optimal policies from the auxiliary system can achieve the robust controller of the nonaffine system. Thirdly, an online adaptive dynamic programming(ADP) algorithm is designed for approximating the optimal solution of the Hamilton–Jacobi–Bellman(HJB) equation.Moreover, the gradient descent approach and projection approach are employed for updating the actor-critic neural network(NN) weights, with the algorithm's convergence being proven. Then, the uniformly ultimately bounded stability of state is guaranteed. Finally, in simulation, some examples are offered for validating the effectiveness of this presented approach. 展开更多
关键词 adaptive dynamic programming robust control nonaffine nonlinear system neural network
原文传递
Event-Triggered Robust Parallel Optimal Consensus Control for Multiagent Systems
14
作者 Qinglai Wei Shanshan Jiao +1 位作者 Qi Dong Fei-Yue Wang 《IEEE/CAA Journal of Automatica Sinica》 2025年第1期40-53,共14页
This paper highlights the utilization of parallel control and adaptive dynamic programming(ADP) for event-triggered robust parallel optimal consensus control(ETRPOC) of uncertain nonlinear continuous-time multiagent s... This paper highlights the utilization of parallel control and adaptive dynamic programming(ADP) for event-triggered robust parallel optimal consensus control(ETRPOC) of uncertain nonlinear continuous-time multiagent systems(MASs).First, the parallel control system, which consists of a virtual control variable and a specific auxiliary variable obtained from the coupled Hamiltonian, allows general systems to be transformed into affine systems. Of interest is the fact that the parallel control technique's introduction provides an unprecedented perspective on eliminating the negative effects of disturbance. Then, an eventtriggered mechanism is adopted to save communication resources while ensuring the system's stability. The coupled HamiltonJacobi(HJ) equation's solution is approximated using a critic neural network(NN), whose weights are updated in response to events. Furthermore, theoretical analysis reveals that the weight estimation error is uniformly ultimately bounded(UUB). Finally,numerical simulations demonstrate the effectiveness of the developed ETRPOC method. 展开更多
关键词 Adaptive dynamic programming(ADP) critic neural network(NN) event-triggered control optimal consensus control robust control
在线阅读 下载PDF
PDP:Parallel Dynamic Programming 被引量:15
15
作者 Fei-Yue Wang Jie Zhang +2 位作者 Qinglai Wei Xinhu Zheng Li Li 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2017年第1期1-5,共5页
Deep reinforcement learning is a focus research area in artificial intelligence.The principle of optimality in dynamic programming is a key to the success of reinforcement learning methods.The principle of adaptive dy... Deep reinforcement learning is a focus research area in artificial intelligence.The principle of optimality in dynamic programming is a key to the success of reinforcement learning methods.The principle of adaptive dynamic programming ADP is first presented instead of direct dynamic programming DP,and the inherent relationship between ADP and deep reinforcement learning is developed.Next,analytics intelligence,as the necessary requirement,for the real reinforcement learning,is discussed.Finally,the principle of the parallel dynamic programming,which integrates dynamic programming and analytics intelligence,is presented as the future computational intelligence.©2014 Chinese Association of Automation. 展开更多
关键词 Parallel dynamic programming dynamic programming Adaptive dynamic programming Reinforcement learning Deep learning neural networks Artificial intelligence
在线阅读 下载PDF
Neural-Network-Based Control for Discrete-Time Nonlinear Systems with Input Saturation Under Stochastic Communication Protocol 被引量:10
16
作者 Xueli Wang Derui Ding +1 位作者 Hongli Dong Xian-Ming Zhang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2021年第4期766-778,共13页
In this paper,an adaptive dynamic programming(ADP)strategy is investigated for discrete-time nonlinear systems with unknown nonlinear dynamics subject to input saturation.To save the communication resources between th... In this paper,an adaptive dynamic programming(ADP)strategy is investigated for discrete-time nonlinear systems with unknown nonlinear dynamics subject to input saturation.To save the communication resources between the controller and the actuators,stochastic communication protocols(SCPs)are adopted to schedule the control signal,and therefore the closed-loop system is essentially a protocol-induced switching system.A neural network(NN)-based identifier with a robust term is exploited for approximating the unknown nonlinear system,and a set of switch-based updating rules with an additional tunable parameter of NN weights are developed with the help of the gradient descent.By virtue of a novel Lyapunov function,a sufficient condition is proposed to achieve the stability of both system identification errors and the update dynamics of NN weights.Then,a value iterative ADP algorithm in an offline way is proposed to solve the optimal control of protocol-induced switching systems with saturation constraints,and the convergence is profoundly discussed in light of mathematical induction.Furthermore,an actor-critic NN scheme is developed to approximate the control law and the proposed performance index function in the framework of ADP,and the stability of the closed-loop system is analyzed in view of the Lyapunov theory.Finally,the numerical simulation results are presented to demonstrate the effectiveness of the proposed control scheme. 展开更多
关键词 Adaptive dynamic programming(ADP) constrained inputs neural network(NN) stochastic communication protocols(SCPs) suboptimal control
在线阅读 下载PDF
基于模糊神经网络在线自学习的多智能体一致性控制 被引量:1
17
作者 张宪霞 唐胜杰 俞寅生 《自动化学报》 北大核心 2025年第3期590-603,共14页
针对多智能体系统分布式一致性控制问题,提出一种新的融合动态模糊神经网络(Dynamic fuzzy neural network,DFNN)和自适应动态规划(Adaptive dynamic programming,ADP)算法的无模型自适应控制方法.类似于强化学习中执行者-评论家结构,D... 针对多智能体系统分布式一致性控制问题,提出一种新的融合动态模糊神经网络(Dynamic fuzzy neural network,DFNN)和自适应动态规划(Adaptive dynamic programming,ADP)算法的无模型自适应控制方法.类似于强化学习中执行者-评论家结构,DFNN和神经网络(Neural network,NN)分别逼近控制策略和性能指标.每个智能体的DFNN执行者从零规则开始,通过在线学习,与其局部邻域的智能体交互而生成和合并规则.最终,每个智能体都有一个独特的DFNN控制器,具有不同的结构和参数,实现了最优的分布式同步控制律.仿真结果表明,本文提出的在线算法在非线性多智能体系统分布式一致性控制中优于传统基于NN的ADP算法. 展开更多
关键词 多智能体系统 自适应动态规划 动态模糊神经网络 分布式一致性控制 在线学习
在线阅读 下载PDF
A Novel Distributed Optimal Adaptive Control Algorithm for Nonlinear Multi-Agent Differential Graphical Games 被引量:7
18
作者 Majid Mazouchi Mohammad Bagher Naghibi-Sistani Seyed Kamal Hosseini Sani 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2018年第1期331-341,共11页
In this paper, an online optimal distributed learning algorithm is proposed to solve leader-synchronization problem of nonlinear multi-agent differential graphical games. Each player approximates its optimal control p... In this paper, an online optimal distributed learning algorithm is proposed to solve leader-synchronization problem of nonlinear multi-agent differential graphical games. Each player approximates its optimal control policy using a single-network approximate dynamic programming(ADP) where only one critic neural network(NN) is employed instead of typical actorcritic structure composed of two NNs. The proposed distributed weight tuning laws for critic NNs guarantee stability in the sense of uniform ultimate boundedness(UUB) and convergence of control policies to the Nash equilibrium. In this paper, by introducing novel distributed local operators in weight tuning laws, there is no more requirement for initial stabilizing control policies. Furthermore, the overall closed-loop system stability is guaranteed by Lyapunov stability analysis. Finally, Simulation results show the effectiveness of the proposed algorithm. 展开更多
关键词 approximate dynamic programming(ADP) distributed control neural networks(NNs) nonlinear differentia graphical games optimal control
在线阅读 下载PDF
Discounted Iterative Adaptive Critic Designs With Novel Stability Analysis for Tracking Control 被引量:9
19
作者 Mingming Ha Ding Wang Derong Liu 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2022年第7期1262-1272,共11页
The core task of tracking control is to make the controlled plant track a desired trajectory.The traditional performance index used in previous studies cannot eliminate completely the tracking error as the number of t... The core task of tracking control is to make the controlled plant track a desired trajectory.The traditional performance index used in previous studies cannot eliminate completely the tracking error as the number of time steps increases.In this paper,a new cost function is introduced to develop the value-iteration-based adaptive critic framework to solve the tracking control problem.Unlike the regulator problem,the iterative value function of tracking control problem cannot be regarded as a Lyapunov function.A novel stability analysis method is developed to guarantee that the tracking error converges to zero.The discounted iterative scheme under the new cost function for the special case of linear systems is elaborated.Finally,the tracking performance of the present scheme is demonstrated by numerical results and compared with those of the traditional approaches. 展开更多
关键词 Adaptive critic design adaptive dynamic programming(ADP) approximate dynamic programming discrete-time nonlinear systems reinforcement learning stability analysis tracking control value iteration(VI)
在线阅读 下载PDF
基于深度优化算法的风光储多能互补电力系统优化调度策略 被引量:1
20
作者 杨银国 谢平平 +3 位作者 刘洋 陆秋瑜 徐展鹏 黄泽杰 《电网与清洁能源》 北大核心 2025年第7期122-131,共10页
风光新能源固有的间歇性和波动性给大规模电力系统发电资源的调度带来了难题,风光储多能互补是应对风光新能源大规模并网的可行途径之一。为制定考虑风光储多能互补的电力系统的优化调度方案,首先,考虑总运行成本最小、新能源弃电量最... 风光新能源固有的间歇性和波动性给大规模电力系统发电资源的调度带来了难题,风光储多能互补是应对风光新能源大规模并网的可行途径之一。为制定考虑风光储多能互补的电力系统的优化调度方案,首先,考虑总运行成本最小、新能源弃电量最小的目标,构建了风光储多能互补优化调度模型。然后,面对庞大的新能源规模和逐步完善的电力系统网架结构所带来的优化调度模型求解困难的问题,基于马尔科夫决策过程和近似动态规划理论,将涉及多时段联合求解的优化模型解耦为所有时段单独求解的子问题。在此基础上,采用深度神经网络对解耦后的子问题进行逐时段的求解,提出了一种基于近似动态规划和深度神经网络的深度优化算法。最后,通过在仿真软件和实际大规模电力系统上进行算例测试,验证了所提方法可行性与有效性。 展开更多
关键词 多能互补 风光新能源 输电网 近似动态规划 深度神经网络
在线阅读 下载PDF
上一页 1 2 8 下一页 到第
使用帮助 返回顶部