期刊文献+
共找到6,810篇文章
< 1 2 250 >
每页显示 20 50 100
Dual-Mode Data-Driven Iterative Learning Control:Applications in Precision Manufacturing and Intelligent Transportation Systems
1
作者 Lei Wang Menghan Wei +3 位作者 Ziwei Huangfu Shunjie Zhu Xuejian Ge Zhengquan Li 《Computers, Materials & Continua》 2026年第2期153-184,共32页
Iterative Learning Control(ILC)provides an effective framework for optimizing repetitive tasks,making it particularly suitable for high-precision applications in both precision manufacturing and intelligent transporta... Iterative Learning Control(ILC)provides an effective framework for optimizing repetitive tasks,making it particularly suitable for high-precision applications in both precision manufacturing and intelligent transportation systems(ITS).This paper presents a systematic review of ILC's developmental progress,current methodologies,and practical implementations across these two critical domains.The review first analyzes the key technical challenges encountered when integrating ILC into precision manufacturing workflows.Through case studies,it evaluates demonstrated improvements in positioning accuracy,surface finish quality,and production throughput.Furthermore,the study examines ILC’s applications in ITS,with particular focus on vehicular motion control applications including autonomous vehicle trajectory tracking,platoon coordination,and traffic signal timing optimization,where its data-driven characteristics enhance adaptability to dynamic environments.Finally,the paper proposes targeted future research directions that are essential for fully realizing ILC’s potential in advancing these interconnected yet distinct fields. 展开更多
关键词 Iterative learning control systematic review precisionmanufacturing intelligent transportation systems
在线阅读 下载PDF
Consensus learning based coordinated formation control of multiple UAVs
2
作者 Yong TANG Yingxin SHOU +1 位作者 Bin XU Zhenbao LIU 《Chinese Journal of Aeronautics》 2026年第2期402-413,共12页
This paper presents a hierarchical formation control strategy to address the challenges of multiple Unmanned Aerial Vehicles(UAVs)formation control within a cooperative consensus framework.The proposed strategy incorp... This paper presents a hierarchical formation control strategy to address the challenges of multiple Unmanned Aerial Vehicles(UAVs)formation control within a cooperative consensus framework.The proposed strategy incorporates a reference command generation layer,which derives UAV attitude commands based on formation requirements,and a tracking control layer to ensure accurate execution.Collaborative variables,including trajectory position and flight speed,are defined using a three-dimensional track particle and autopilot model,enabling the development of a consensus-based formation control law.Desired attitude angles are computed through altitudehold and coordinated-turn strategies.A sliding surface is designed based on reference models derived from flight quality metrics,while an adaptive controller compensates for aerodynamic model uncertainties.To enhance learning capabilities,a prediction error mechanism based on a series-parallel estimation model is introduced,enabling collaborative learning and the sharing of network weight estimation parameters within the multi-agent system.This facilitates the design of a distributed composite learning law.Lyapunov stability analysis confirms the local exponential stability of the tracking error.The simulations of a twelve-UAV formation,along with comparative analysis of two algorithms,demonstrate the system’s capability for formation maintenance and high-precision tracking control. 展开更多
关键词 Collaborative consistency Distributed composite learning Multiple unmanned aerial vehicles system Serial-parallel estimation model Sliding mode adaptive controller
原文传递
Adaptive Reinforcement Learning with Multi-Modal Perception for Autonomous Formation Control and Exploration in Large-Scale Multi-UAV Swarms
3
作者 Ziyuan Ma Huajun Gong Xinhua Wang 《Journal of Beijing Institute of Technology》 2026年第1期63-83,共21页
To address the challenge of achieving decentralized,scalable,and adaptive control for large-scale multiple unmanned aerial vehicle(multi-UAV)swarms in dynamic urban environments with obstacles and wind perturbations,w... To address the challenge of achieving decentralized,scalable,and adaptive control for large-scale multiple unmanned aerial vehicle(multi-UAV)swarms in dynamic urban environments with obstacles and wind perturbations,we proposed a hybrid framework integrating adaptive reinforcement learning(RL),multi-modal perception fusion,and enhanced pigeon flock optimization(PFO)with curiosity-driven exploration to enable robust autonomous and formation control.The framework leverages meta-learning to optimize RL policies for real-time adaptation,fuses sensor data for precise state estimation,and enhances PFO with learned leader-follower dynamics and exploration rewards to maintain cohesive formations and explore uncertain areas.For swarms of 10–30 UAVs,it achieves 34%faster convergence,61%reduced stability root mean square error(RMSE),88%fewer collisions and 85.6%–92.3%success rates in target detection and encirclement,outperforming standard multi-agent RL,pure PFO,and single-modality RL.Three-dimensional trajectory visualizations confirm cohesive formations,collision-free maneuvers,and efficient exploration in urban search-and-rescue scenarios.Innovations include meta-RL for rapid adaptation,multi-modal fusion for robust perception,and curiosity-driven PFO for scalable,decentralized control,advancing real-world multi-UAV swarm autonomy and coordination. 展开更多
关键词 multiple unmanned aerial vehicle(multi-UAV)swarm autonomous control reinforcement learning(RL) multi-modal perception pigeon flock optimization(PFO)
在线阅读 下载PDF
Context-Aware Relational Learning for Cooperative UAV Formation
4
作者 Zhuxun Li Haoxian Jiang Rui Zhou 《Journal of Beijing Institute of Technology》 2026年第1期44-52,共9页
Robust cooperative unmanned aerial vehicle(UAV)formation in complex 3D environments is hampered by reward sparsity and inefficient collaboration.To address this,we propose context-aware relational agent learning(CORAL... Robust cooperative unmanned aerial vehicle(UAV)formation in complex 3D environments is hampered by reward sparsity and inefficient collaboration.To address this,we propose context-aware relational agent learning(CORAL),a novel multi-agent deep reinforcement learning framework.CORAL synergistically integrates two modules:(1)a novelty-based intrinsic reward module to drive efficient exploration and(2)an explicit relational learning module that allows agents to predict peer intentions and enhance coordination.Built on a multi-agent Actor-Critic architecture,CORAL enables agents to balance self-interest with group objectives.Comprehensive evaluations in a high-fidelity simulation show that our method significantly outperforms state-of-theart baselines like multi-agent deep deterministic policy gradient(MADDPG)and monotonic value function factorisation for deep multi-agent reinforcement learning(QMIX)in path planning efficiency,collision avoidance,and scalability. 展开更多
关键词 multi-agent reinforcement learning UAV swarm cooperative formation control path planning context-aware exploration relational learning
在线阅读 下载PDF
Fault-observer-based iterative learning model predictive controller for trajectory tracking of hypersonic vehicles 被引量:2
5
作者 CUI Peng GAO Changsheng AN Ruoming 《Journal of Systems Engineering and Electronics》 2025年第3期803-813,共11页
This work proposes the application of an iterative learning model predictive control(ILMPC)approach based on an adaptive fault observer(FOBILMPC)for fault-tolerant control and trajectory tracking in air-breathing hype... This work proposes the application of an iterative learning model predictive control(ILMPC)approach based on an adaptive fault observer(FOBILMPC)for fault-tolerant control and trajectory tracking in air-breathing hypersonic vehicles.In order to increase the control amount,this online control legislation makes use of model predictive control(MPC)that is based on the concept of iterative learning control(ILC).By using offline data to decrease the linearized model’s faults,the strategy may effectively increase the robustness of the control system and guarantee that disturbances can be suppressed.An adaptive fault observer is created based on the suggested ILMPC approach in order to enhance overall fault tolerance by estimating and compensating for actuator disturbance and fault degree.During the derivation process,a linearized model of longitudinal dynamics is established.The suggested ILMPC approach is likely to be used in the design of hypersonic vehicle control systems since numerical simulations have demonstrated that it can decrease tracking error and speed up convergence when compared to the offline controller. 展开更多
关键词 hypersonic vehicle actuator fault tracking control iterative learning control(ILC) model predictive control(MPC) fault observer
在线阅读 下载PDF
Multiplayer Pareto optimal control with H_(∞)constraint for nonlinear stochastic system via online synchronous reinforcement learning 被引量:1
6
作者 Li Wang Xiushan Jiang +1 位作者 Dongya Zhao Bor-Sen Chen 《Journal of Automation and Intelligence》 2025年第3期207-216,共10页
This paper investigates a multiplayer Pareto game for affine nonlinear stochastic systems disturbed by both external and the internal multiplicative noises.The Pareto cooperative optimal strategies with the H_(∞) con... This paper investigates a multiplayer Pareto game for affine nonlinear stochastic systems disturbed by both external and the internal multiplicative noises.The Pareto cooperative optimal strategies with the H_(∞) constraint are resolved by integrating H_(2)/H_(∞) theory with Pareto game theory.First,a nonlinear stochastic bounded real lemma(SBRL)is derived,explicitly accounting for non-zero initial conditions.Through the analysis of four cross-coupled Hamilton-Jacobi equations(HJEs),we establish necessary and sufficient conditions for the existence of Pareto optimal strategies with the H_(∞) constraint.Secondly,to address the complexity of solving these nonlinear partial differential HJEs,we propose a neural network(NN)framework with synchronous tuning rules for the actor,critic,and disturbance components,based on a reinforcement learning(RL)approach.The designed tuning rules ensure convergence of the actor-critic-disturbance components to the desired values,enabling the realization of robust Pareto control strategies.The convergence of the proposed algorithm is rigorously analyzed using a constructed Lyapunov function for the NN weight errors.Finally,a numerical simulation example is provided to demonstrate the effectiveness of the proposed methods and main results. 展开更多
关键词 Pareto control Nonlinear stochastic system Hamilton-Jacobi equations H_(∞)control Reinforcement learning
在线阅读 下载PDF
Geologic hazard susceptibility assessment based on statistical optimization and machine learning:A case study of the Loess Plateau,Shaanxi Province,northwestern China
7
作者 Hao Cheng Zhen-kai Zhang +5 位作者 Zeng-lin Hong Wen-long Zhang Hong-quan Teng Shuai Yang Zi-yao Wang Yu-xuan Dong 《China Geology》 2026年第1期136-151,共16页
This study developed a modeling methodology for statistical optimization-based geologic hazard susceptibility assessment,aiming to enhance the comprehensive performance and classification accuracy of the assessment mo... This study developed a modeling methodology for statistical optimization-based geologic hazard susceptibility assessment,aiming to enhance the comprehensive performance and classification accuracy of the assessment models.First,the cumulative probability method revealed that a low probability(15%)of geologic hazards between any two geologic hazard points occurred outside a buffer zone with a radius of 2297 m(i.e.,the distance threshold).The training dataset was established,consisting of negative samples(non-hazard points)randomly generated based on the distance threshold,positive samples(i.e.,historical hazards),and 13 conditioning factors.Then,models were built using five machine learning algorithms,namely random forest(RF),gradient boosting decision tree(GBDT),naive Bayes(NB),logistic regression(LR),and support vector machine(SVM).The comprehensive performance of the models was assessed using the area under the receiver operating characteristic curve(AUC)and overall accuracy(OA)as indicators,revealing that RF exhibited the best performance,with OA and AUC values of 2.7127 and 0.981,respectively.Furthermore,the machine learning models constructed by considering the distance threshold outperformed those built using the unoptimized dataset.The characteristic factors were ranked using the mutual information method,with their scores decreasing in the order of rainfall(0.1616),altitude(0.06),normalized difference vegetation index(NDVI;0.04),and distance from roads(0.03).Finally,the geologic hazard susceptibility classification was assessed using the natural breaks method combined with a clustering algorithm.The results indicate that the clustering algorithm exhibited higher classification accuracy than the natural breaks method.The findings of this study demonstrate that the proposed model optimization scheme can provide a scientific basis for the prevention and control of geologic hazards. 展开更多
关键词 COLLAPSE LANDSLIDE Debris flow Geologic hazard susceptibility assessment Machine learning RF-GBDT-NB-LR-SVM Cumulative probability Cluster analysis Loess Plateau Geologic hazard prevention and control Geological survey engineering
在线阅读 下载PDF
Machine Learning-Based Online Monitoring and Closed-Loop Controlling for 3D Printing of Continuous Fiber-Reinforced Composites 被引量:1
8
作者 Xinyun Chi Jiacheng Xue +6 位作者 Lei Jia Jiaqi Yao Huihui Miao Lingling Wu Tengfei Liu Xiaoyong Tian Dichen Li 《Additive Manufacturing Frontiers》 2025年第2期90-96,共7页
Ensuring the consistent mechanical performance of three-dimensional(3D)-printed continuous fiber-reinforced composites is a significant challenge in additive manufacturing.The current reliance on manual monitoring exa... Ensuring the consistent mechanical performance of three-dimensional(3D)-printed continuous fiber-reinforced composites is a significant challenge in additive manufacturing.The current reliance on manual monitoring exacerbates this challenge by rendering the process vulnerable to environmental changes and unexpected factors,resulting in defects and inconsistent product quality,particularly in unmanned long-term operations or printing in extreme environments.To address these issues,we developed a process monitoring and closed-loop feedback control strategy for the 3D printing process.Real-time printing image data were captured and analyzed using a well-trained neural network model,and a real-time control module-enabled closed-loop feedback control of the flow rate was developed.The neural network model,which was based on image processing and artificial intelligence,enabled the recognition of flow rate values with an accuracy of 94.70%.The experimental results showed significant improvements in both the surface performance and mechanical properties of printed composites,with three to six times improvement in tensile strength and elastic modulus,demonstrating the effectiveness of the strategy.This study provides a generalized process monitoring and feedback control method for the 3D printing of continuous fiber-reinforced composites,and offers a potential solution for remote online monitoring and closed-loop adjustment in unmanned or extreme space environments. 展开更多
关键词 Continuous fiber-reinforced composites 3D printing Computer vision Machine learning Defect detection Feedback control
在线阅读 下载PDF
Graded density impactor design via machine learning and numerical simulation:Achieve controllable stress and strain rate 被引量:1
9
作者 Yahui Huang Ruizhi Zhang +6 位作者 Shuaixiong Liu Jian Peng Yong Liu Han Chen Jian Zhang Guoqiang Luo Qiang Shen 《Defence Technology(防务技术)》 2025年第9期262-273,共12页
The graded density impactor(GDI)dynamic loading technique is crucial for acquiring the dynamic physical property parameters of materials used in weapons.The accuracy and timeliness of GDI structural design are key to ... The graded density impactor(GDI)dynamic loading technique is crucial for acquiring the dynamic physical property parameters of materials used in weapons.The accuracy and timeliness of GDI structural design are key to achieving controllable stress-strain rate loading.In this study,we have,for the first time,combined one-dimensional fluid computational software with machine learning methods.We first elucidated the mechanisms by which GDI structures control stress and strain rates.Subsequently,we constructed a machine learning model to create a structure-property response surface.The results show that altering the loading velocity and interlayer thickness has a pronounced regulatory effect on stress and strain rates.In contrast,the impedance distribution index and target thickness have less significant effects on stress regulation,although there is a matching relationship between target thickness and interlayer thickness.Compared with traditional design methods,the machine learning approach offers a10^(4)—10^(5)times increase in efficiency and the potential to achieve a global optimum,holding promise for guiding the design of GDI. 展开更多
关键词 Machine learning Numerical simulation Graded density impactor controllable stress-strain rate loading Response surface methodology
在线阅读 下载PDF
Continuous Learning and Adaptation of Neural Control for Proprioceptive Feedback Integration in a Quadruped Robot
10
作者 Yanbin Zhang Yang Li Zhendong Dai 《Journal of Bionic Engineering》 2025年第5期2367-2382,共16页
Autonomous legged robots,capable of navigating uneven terrain,can perform a diverse array of tasks.However,designing locomotion controllers remains challenging.In particular,designing a controller based on durable and... Autonomous legged robots,capable of navigating uneven terrain,can perform a diverse array of tasks.However,designing locomotion controllers remains challenging.In particular,designing a controller based on durable and reliable proprioceptive sensors,is essential for achieving adaptability.Presently,the controller must either be manually designed for specific robots and tasks,or developed using machine-learning techniques,which require extensive training time and result in complex controllers.Inspired by animal locomotion,we propose a simple yet comprehensive closed-loop modular framework that utilizes minimal proprioceptive feedback(i.e.,the Coxa-Femur(CF)joint angle),enabling a quadruped robot to efficiently navigate unpredictable and uneven terrains,including the step and slope.The framework comprises a basic neural control network capable of rapidly learning optimized motor patterns,and a straightforward module for sensory feedback sharing and integration.In a series of experiments,we show that integrating sensory feedback into the base neural control network aids the robot in continually learning robust motor patterns on flat,step,and slope terrain,compared with the open-loop base framework.Sharing sensory feedback information across the four legs enables a quadruped robot to proactively navigate unpredictable steps with minimal interaction.Furthermore,the controller remains functional even in the absence of sensor signals.This control configuration was successfully transferred to a physical robot without any modifications. 展开更多
关键词 Bioinspired robot learning Continual learning Optimization and optimal control Sensor-based control
在线阅读 下载PDF
Beyond Performance of Learning Control Subject to Uncertainties and Noise: A Frequency-Domain Approach Applied to Wafer Stages
11
作者 Fazhi Song Ning Cui +4 位作者 Shuaiqi Chen Kai Zhang Yang Liu Xinkai Chen Jiubin Tan 《IEEE/CAA Journal of Automatica Sinica》 2025年第1期198-214,共17页
The increasingly stringent performance requirement in integrated circuit manufacturing, characterized by smaller feature sizes and higher productivity, necessitates the wafer stage executing a extreme motion with the ... The increasingly stringent performance requirement in integrated circuit manufacturing, characterized by smaller feature sizes and higher productivity, necessitates the wafer stage executing a extreme motion with the accuracy in terms of nanometers. This demanding requirement witnesses a widespread application of iterative learning control(ILC), given the repetitive nature of wafer scanning. ILC enables substantial performance improvement by using past measurement data in combination with the system model knowledge. However, challenges arise in cases where the data is contaminated by the stochastic noise, or when the system model exhibits significant uncertainties, constraining the achievable performance. In response to this issue, an extended state observer(ESO) based adaptive ILC approach is proposed in the frequency domain.Despite being model-based, it utilizes only a rough system model and then compensates for the resulting model uncertainties using an ESO, thereby achieving high robustness against uncertainties with minimal modeling effort. Additionally, an adaptive learning law is developed to mitigate the limited performance in the presence of stochastic noise, yielding high convergence accuracy yet without compromising convergence speed. Simulation and experimental comparisons with existing model-based and data-driven inversion-based ILC validate the effectiveness as well as the superiority of the proposed method. 展开更多
关键词 Extended state observer learning control model uncertainties motion control stochastic noise
在线阅读 下载PDF
A Survey on Reinforcement Learning for Optimal Decision-Making and Control of Intelligent Vehicles
12
作者 Yixing Lan Xin Xu +3 位作者 Jiahang Liu Xinglong Zhang Yang Lu Long Cheng 《CAAI Transactions on Intelligence Technology》 2025年第6期1593-1615,共23页
Reinforcement learning(RL)has been widely studied as an efficient class of machine learning methods for adaptive optimal control under uncertainties.In recent years,the applications of RL in optimised decision-making ... Reinforcement learning(RL)has been widely studied as an efficient class of machine learning methods for adaptive optimal control under uncertainties.In recent years,the applications of RL in optimised decision-making and motion control of intelligent vehicles have received increasing attention.Due to the complex and dynamic operating environments of intelligent vehicles,it is necessary to improve the learning efficiency and generalisation ability of RL-based decision and control algorithms under different conditions.This survey systematically examines the theoretical foundations,algorithmic advancements and practical challenges of applying RL to intelligent vehicle systems operating in complex and dynamic environments.The major algorithm frameworks of RL are first introduced,and the recent advances in RL-based decision-making and control of intelligent vehicles are overviewed.In addition to self-learning decision and control approaches using state measurements,the developments of DRL methods for end-to-end driving control of intelligent vehicles are summarised.The open problems and directions for further research works are also discussed. 展开更多
关键词 adaptive dynamic programming intelligent vehicles learning control optimal decision-making reinforcement learning
在线阅读 下载PDF
The Emergency Control Method for Multi-Scenario Sub-Synchronous Oscillation in Wind Power Grid Integration Systems Based on Transfer Learning
13
作者 Qing Zhu Denghui Guo +3 位作者 Rui Ruan Zhidong Chai Chaoqun Wang Zhiwen Guan 《Energy Engineering》 2025年第8期3133-3154,共22页
This study presents an emergency control method for sub-synchronous oscillations in wind power gridconnected systems based on transfer learning,addressing the issue of insufficient generalization ability of traditiona... This study presents an emergency control method for sub-synchronous oscillations in wind power gridconnected systems based on transfer learning,addressing the issue of insufficient generalization ability of traditional methods in complex real-world scenarios.By combining deep reinforcement learning with a transfer learning framework,cross-scenario knowledge transfer is achieved,significantly enhancing the adaptability of the control strategy.First,a sub-synchronous oscillation emergency control model for the wind power grid integration system is constructed under fixed scenarios based on deep reinforcement learning.A reward evaluation system based on the active power oscillation pattern of the system is proposed,introducing penalty functions for the number of machine-shedding rounds and the number of machines shed.This avoids the economic losses and grid security risks caused by the excessive one-time shedding of wind turbines.Furthermore,transfer learning is introduced into model training to enhance the model’s generalization capability in dealing with complex scenarios of actual wind power grid integration systems.By introducing the Maximum Mean Discrepancy(MMD)algorithm to calculate the distribution differences between source data and target data,the online decision-making reliability of the emergency control model is improved.Finally,the effectiveness of the proposed emergency control method for multi-scenario sub-synchronous oscillation in wind power grid integration systems based on transfer learning is analyzed using the New England 39-bus system. 展开更多
关键词 Synchronous phasor data sub-synchronous oscillation emergency control deep reinforcement learning transfer learning
在线阅读 下载PDF
基于深度Q-learning算法的智能电网管控模型研究
14
作者 王筠 李志鹏 +2 位作者 项旭 张军堂 石雷波 《自动化技术与应用》 2026年第2期54-57,142,共5页
设计基于深度Q-learning算法的智能电网管控模型,将可验证声明(verifiable credential, VC)和分布式数字身份(decentralized identity, DID)作为应用程序身份凭证与软件定义网络(software-defined networking, SDN)控制器,结合动态信任... 设计基于深度Q-learning算法的智能电网管控模型,将可验证声明(verifiable credential, VC)和分布式数字身份(decentralized identity, DID)作为应用程序身份凭证与软件定义网络(software-defined networking, SDN)控制器,结合动态信任评估算法与基于属性的访问控制策略,构建基于区块链的智能电网分布式SDN管控模型。在资源分配、网络拓扑动态变化以及安全威胁不断演变的情况下,实施基于区块链的分布式SDN网络的优化。实验测试结果表明,设计方法在通过深度Q-learning优化模型后累积奖励明显大幅增加,在多种安全性能方面表现出色,能够清除恶意域,确保网络环境的安全。 展开更多
关键词 SDN控制器 分布式SDN网络 深度Q-learning算法 区块链 智能电网管控模型
在线阅读 下载PDF
Reinforcement Learning in Mechatronic Systems: A Case Study on DC Motor Control
15
作者 Alexander Nüßgen Alexander Lerch +5 位作者 René Degen Marcus Irmer Martin de Fries Fabian Richter Cecilia Boström Margot Ruschitzka 《Circuits and Systems》 2025年第1期1-24,共24页
The integration of artificial intelligence into the development and production of mechatronic products offers a substantial opportunity to enhance efficiency, adaptability, and system performance. This paper examines ... The integration of artificial intelligence into the development and production of mechatronic products offers a substantial opportunity to enhance efficiency, adaptability, and system performance. This paper examines the utilization of reinforcement learning as a control strategy, with a particular focus on its deployment in pivotal stages of the product development lifecycle, specifically between system architecture and system integration and verification. A controller based on reinforcement learning was developed and evaluated in comparison to traditional proportional-integral controllers in dynamic and fault-prone environments. The results illustrate the superior adaptability, stability, and optimization potential of the reinforcement learning approach, particularly in addressing dynamic disturbances and ensuring robust performance. The study illustrates how reinforcement learning can facilitate the transition from conceptual design to implementation by automating optimization processes, enabling interface automation, and enhancing system-level testing. Based on the aforementioned findings, this paper presents future directions for research, which include the integration of domain-specific knowledge into the reinforcement learning process and the validation of this process in real-world environments. The results underscore the potential of artificial intelligence-driven methodologies to revolutionize the design and deployment of intelligent mechatronic systems. 展开更多
关键词 Artificial Intelligence in Product Development Mechatronic Systems Reinforcement learning for control System Integration and Verification Adaptive Optimization Processes Knowledge-Based Engineering
在线阅读 下载PDF
Reinforcement learning based optimized backstepping control for hypersonic vehicles with disturbance observer
16
作者 Haoyu CHENG Xin LIU +2 位作者 Xiaoxi LIANG Xiaoyan ZHANG Shaoyi LI 《Chinese Journal of Aeronautics》 2025年第11期413-437,共25页
This paper introduces an optimized backstepping control method for Flexible Airbreathing Hypersonic Vehicles(FAHVs).The approach incorporates nonlinear disturbance observation and reinforcement learning to address com... This paper introduces an optimized backstepping control method for Flexible Airbreathing Hypersonic Vehicles(FAHVs).The approach incorporates nonlinear disturbance observation and reinforcement learning to address complex control challenges.The Minimal Learning Parameter(MLP)technique is applied to manage unknown nonlinear dynamics,significantly reducing the computational load usually associated with Neural Network(NN)weight updates.To improve the control system robustness,an MLP-based nonlinear disturbance observer is designed,which estimates lumped disturbances,including flexibility effects,model uncertainties,and external disruptions within the FAHVs.In parallel,the control strategy integrates reinforcement learning using an MLP-based actor-critic framework within the backstepping design to achieve both optimality and robustness.The actor performs control actions,while the critic assesses the optimal performance index function.To minimize this index function,an adaptive gradient descent method constructs both the actor and critic.Lyapunov analysis is employed to demonstrate that all signals in the closed-loop system are semiglobally uniformly ultimately bounded.Simulation results confirm that the proposed control strategy delivers high control performance,marked by improved accuracy and reduced energy consumption. 展开更多
关键词 Hypersonic vehicles Minimal learning parameter Nonlinear disturbance observer Optimized backstepping control Reinforcement learning
原文传递
Machine learning assisted quality control in metal additive manufacturing:a review
17
作者 Zeqi Hu Changlin Huang +3 位作者 Lechun Xie Lin Hua Yujie Yuan Laichang Zhang 《Advanced Powder Materials》 2025年第6期138-182,共45页
Additive manufacturing(AM)promotes the production of metallic parts with significant design flexibility,yet its use in critical applications is hindered by challenges in ensuring consistent quality and performance.Pro... Additive manufacturing(AM)promotes the production of metallic parts with significant design flexibility,yet its use in critical applications is hindered by challenges in ensuring consistent quality and performance.Process variability often leads to defects,insufficient geometric accuracy and inadequate material properties,which are difficult to effectively manage due to limitations of traditional quality control methods in modeling highdimensional nonlinear relationships and enabling adaptive control.Machine learning(ML)offers a transformative approach to model intricate process-structure-property relationships by leveraging the rich data environment of AM.The study presents a comprehensive examination of ML-driven quality assurance implementations in metallic AM.First,it uniquely examines the innovative exploration of ML in predicting and understanding the fundamental multi-physics fields that influence the quality of a fabricated component,including temperature fields,fluid dynamics and stress/strain evolution.Subsequently,the application of ML in optimizing key quality attributes,including defect detection and mitigation(porosity,cracks,etc.),geometric fidelity enhancement(dimensional accuracy,surface roughness,etc.)and material property tailoring(mechanical strength,fatigue life,corrosion resistance,etc.),are discussed in detail.Finally,the development of ML-driven real-time closed-loop control systems for intelligent quality assurance,the strategies for addressing the data scarcity and cross-scenario transferability in metal AM are discussed.This article provides a novel perspective on the profound potential of ML technology for metal AM quality control applications,highlights the challenges faced during research,and outlines future development directions. 展开更多
关键词 Metal additive manufacturing(Metal AM) Machine learning(ML) Quality control Multi-physics optimization Closed-loop control
在线阅读 下载PDF
A Robust Large-Scale Multiagent Deep Reinforcement Learning Method for Coordinated Automatic Generation Control of Integrated Energy Systems in a Performance-Based Frequency Regulation Market
18
作者 Jiawen Li Tao Zhou 《IEEE/CAA Journal of Automatica Sinica》 2025年第7期1475-1488,共14页
To enhance the frequency stability and lower the regulation mileage payment of a multiarea integrated energy system(IES)that supports the power Internet of Things(IoT),this paper proposes a data-driven cooperative met... To enhance the frequency stability and lower the regulation mileage payment of a multiarea integrated energy system(IES)that supports the power Internet of Things(IoT),this paper proposes a data-driven cooperative method for automatic generation control(AGC).The method consists of adaptive fractional-order proportional-integral(FOPI)controllers and a novel efficient integration exploration multiagent twin delayed deep deterministic policy gradient(EIE-MATD3)algorithm.The FOPI controllers are designed for each area based on the performancebased frequency regulation market mechanism.The EIE-MATD3 algorithm is used to tune the coefficients of the FOPI controllers in real time using centralized training and decentralized execution.The algorithm incorporates imitation learning and efficient integration exploration to obtain a more robust coordinated control strategy.An experiment on the four-area China Southern Grid(CSG)real-time digital system shows that the proposed method can improve the control performance and reduce the regulation mileage payment of each area in the IES. 展开更多
关键词 Adaptive fractional-order PI controller automatic generation control deep reinforcement learning performance-based frequency regulation market
在线阅读 下载PDF
Stable and continuous vertical jumping control of hydraulic legged robots through reinforcement learning
19
作者 Junhui ZHANG Pengyuan JI +5 位作者 Lizhou FANG Jinyuan LIU Dandan WANG Jikun AI Huaizhi ZONG Bing XU 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 2025年第12期1163-1178,共16页
Hydraulic legged robots have potential for high-dynamic motion due to their large power-to-weight ratios. However, it is challenging to ensure both stability and continuity in the motion of such robots. In this study,... Hydraulic legged robots have potential for high-dynamic motion due to their large power-to-weight ratios. However, it is challenging to ensure both stability and continuity in the motion of such robots. In this study, we propose a jumping motion control framework based on deep reinforcement learning that enables hydraulic limb leg units to perform stable and continuous jumping motions. First, to accurately represent the performance of a physical prototype, a quasi-realistic model incorporating physical feasibility constraints is constructed. This model is informed by analysis of the relevant fluid dynamics, and incorporates a trajectory generator and a motion tracking controller. To achieve stable and continuous jumping performance, a deep reinforcement learning algorithm is developed, which jointly optimizes the trajectory generator and the motion tracking controller. Through validation on the physical prototype, we demonstrate that the proposed method reduces the maximum deviation and the average deviation by over 47% and 60%, respectively, and improves landing compliance by up to 7.7% compared to a baseline optimization algorithm, the non-dominated sorting genetic algorithm (NSGA-II). The proposed control framework may serve as a reference for high-dynamic motion control of legged robots and multi-objective optimization across several decision variables. 展开更多
关键词 Legged robot Deep reinforcement learning Quasi-realistic modelling Hydraulic system Jumping control
原文传递
Learning-based tracking control of AUV:Mixed policy improvement and game-based disturbance rejection
20
作者 Jun Ye Hongbo Gao +4 位作者 Manjiang Hu Yougang Bian Qingjia Cui Xiaohui Qin Rongjun Ding 《CAAI Transactions on Intelligence Technology》 2025年第2期510-528,共19页
A mixed adaptive dynamic programming(ADP)scheme based on zero-sum game theory is developed to address optimal control problems of autonomous underwater vehicle(AUV)systems subject to disturbances and safe constraints.... A mixed adaptive dynamic programming(ADP)scheme based on zero-sum game theory is developed to address optimal control problems of autonomous underwater vehicle(AUV)systems subject to disturbances and safe constraints.By combining prior dynamic knowledge and actual sampled data,the proposed approach effectively mitigates the defect caused by the inaccurate dynamic model and significantly improves the training speed of the ADP algorithm.Initially,the dataset is enriched with sufficient reference data collected based on a nominal model without considering modelling bias.Also,the control object interacts with the real environment and continuously gathers adequate sampled data in the dataset.To comprehensively leverage the advantages of model-based and model-free methods during training,an adaptive tuning factor is introduced based on the dataset that possesses model-referenced information and conforms to the distribution of the real-world environment,which balances the influence of model-based control law and data-driven policy gradient on the direction of policy improvement.As a result,the proposed approach accelerates the learning speed compared to data-driven methods,concurrently also enhancing the tracking performance in comparison to model-based control methods.Moreover,the optimal control problem under disturbances is formulated as a zero-sum game,and the actor-critic-disturbance framework is introduced to approximate the optimal control input,cost function,and disturbance policy,respectively.Furthermore,the convergence property of the proposed algorithm based on the value iteration method is analysed.Finally,an example of AUV path following based on the improved line-of-sight guidance is presented to demonstrate the effectiveness of the proposed method. 展开更多
关键词 adaptive dynamic programming autonomous underwater vehicle game theory optimal control reinforcement learning
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部