Abstract
In this paper, we investigate which exploration method yields the best performance in the game Bomberman. In Bomberman the controlled agent has to kill opponents by placing bombs. The agent is represented by a multi-layer perceptron that learns to play the game with the use of Q-learning. We introduce two novel exploration strategies: Error-Driven-ε and Interval-Q, which base their explorative behavior on the temporal-difference error of Q-learning. The learning capabilities of these exploration strategies are compared to five existing methods: Random-Walk, Greedy, ε-Greedy, Diminishing ε-Greedy, and Max-Boltzmann. The results show that the methods that combine exploration with exploitation perform much better than the Random-Walk and Greedy strategies, which only select exploration or exploitation actions. Furthermore, the results show that Max-Boltzmann exploration performs the best in overall from the different techniques. The Error-Driven-ε exploration strategy also performs very well, but suffers from an unstable learning behavior.
Author supplied keywords
Cite
CITATION STYLE
Kormelink, J. G., Drugan, M. M., & Wiering, M. A. (2018). Exploration methods for connectionist Q-learning in bomberman. In ICAART 2018 - Proceedings of the 10th International Conference on Agents and Artificial Intelligence (Vol. 2, pp. 355–362). SciTePress. https://doi.org/10.5220/0006556403550362
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.