An Improved Q-Learning Algorithm for Path Planning in Maze Environments

N/ACitations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The path planning is the problem of finding the optimal paths in a given environment, which has become an important way to test the intelligent learning algorithms. In AI-based path planning, the earliest and more in-depth issue is Intelligent Obstacle Avoidance, that is, an agent needs to successfully avoid all obstacles or traps in an unknown environment. Compared with other learning methods, RL (Reinforcement Learning) has inherent advantages in path planning. Unlike most machine learning methods, RL is an unsupervised active learning method. It can not only effectively achieve intelligent obstacle avoidance, but also find the optimal path from unfamiliar environment such as maze through many experiments. Q-learning algorithm is recognized as one of the most typical RL algorithms. Its advantages are simple and practical, but it also has the significant disadvantage of slow convergence speed. This paper gives a called ɛ-Q-Learning algorithm, which is an improvement to the traditional Q-Learning algorithm by using Dynamic Search Factor technology. Experiments show that compared with the existing Q-Learning algorithms, ɛ-Q-Learning can find out a better optimal paths with lower costs of searching.

Cite

CITATION STYLE

APA

Gu, S., & Mao, G. (2021). An Improved Q-Learning Algorithm for Path Planning in Maze Environments. In Advances in Intelligent Systems and Computing (Vol. 1251 AISC, pp. 547–557). Springer. https://doi.org/10.1007/978-3-030-55187-2_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free