Q-learning is a classical reinforcement learning method with broad applicability.It can respond effectively to environmental changes and provide flexible strategies,making it suitable for solving robot path-planning p...Q-learning is a classical reinforcement learning method with broad applicability.It can respond effectively to environmental changes and provide flexible strategies,making it suitable for solving robot path-planning problems.However,Q-learning faces challenges in search and update efficiency.To address these issues,we propose an improved Q-learning(IQL)algorithm.We use an enhanced Ant Colony Optimization(ACO)algorithmto optimizeQtable initialization.We also introduce the UCH mechanism to refine the reward function and overcome the exploration dilemma.The IQL algorithm is extensively tested in three grid environments of different scales.The results validate the accuracy of themethod and demonstrate superior path-planning performance compared to traditional approaches.The algorithm reduces the number of trials required for convergence,improves learning efficiency,and enables faster adaptation to environmental changes.It also enhances stability and accuracy by reducing the standard deviation of trials to zero.On grid maps of different sizes,IQL achieves higher expected returns.Compared with the original Q-learning algorithm,IQL improves performance by 12.95%,18.28%,and 7.98% on 10*10,20*20,and 30*30 maps,respectively.The proposed algorithm has promising applications in robotics,path planning,intelligent transportation,aerospace,and game development.展开更多
基金Financial supports from the National Natural Science Foundation of China(GrantNo.52374123&51974144)Project of Liaoning Provincial Department of Education(GrantNo.LJKZ0340)Liaoning Revitalization Talents Program(Grant No.XLYC2211085)are greatly acknowledged.
文摘Q-learning is a classical reinforcement learning method with broad applicability.It can respond effectively to environmental changes and provide flexible strategies,making it suitable for solving robot path-planning problems.However,Q-learning faces challenges in search and update efficiency.To address these issues,we propose an improved Q-learning(IQL)algorithm.We use an enhanced Ant Colony Optimization(ACO)algorithmto optimizeQtable initialization.We also introduce the UCH mechanism to refine the reward function and overcome the exploration dilemma.The IQL algorithm is extensively tested in three grid environments of different scales.The results validate the accuracy of themethod and demonstrate superior path-planning performance compared to traditional approaches.The algorithm reduces the number of trials required for convergence,improves learning efficiency,and enables faster adaptation to environmental changes.It also enhances stability and accuracy by reducing the standard deviation of trials to zero.On grid maps of different sizes,IQL achieves higher expected returns.Compared with the original Q-learning algorithm,IQL improves performance by 12.95%,18.28%,and 7.98% on 10*10,20*20,and 30*30 maps,respectively.The proposed algorithm has promising applications in robotics,path planning,intelligent transportation,aerospace,and game development.