Online learning of robot soccer free kick plans using a bandit approach

7Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

This paper presents an online learning approach for teams of autonomous soccer robots to select free kick plans. In robot soccer, free kicks present an opportunity to execute plans with relatively controllable initial conditions. However, the effectiveness of each plan is highly dependent on the adversary, and there are few free kicks during each game, making it necessary to learn online from sparse observations. To achieve learning, we first greatly reduce the planning space by framing the problem as a contextual multi-armed bandit problem, in which the actions are a set of pre-computed plans, and the state is the position of the free kick on the field. During execution, we model the reward function for different free kicks using Gaussian Processes, and perform online learning using the Upper Confidence Bound algorithm. Results from a physics-based simulation reveal that the robots are capable of adapting to various different realistic opponents to maximize their expected reward during free kicks.

Cite

CITATION STYLE

APA

Mendoza, J. P., Simmons, R., & Veloso, M. (2016). Online learning of robot soccer free kick plans using a bandit approach. In Proceedings International Conference on Automated Planning and Scheduling, ICAPS (Vol. 2016-January, pp. 504–508). AAAI press. https://doi.org/10.1609/icaps.v26i1.13795

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free