Beating-Time Gestures Imitation Learning for Humanoid Robots

  • Amelynck D
  • Maes P
  • Martens J
  • et al.
N/ACitations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

Beating-time gestures are movement patterns of the hand swaying along with music, thereby indicating accented musical pulses. The spatiotemporal configuration of these patterns makes it difficult to analyse and model them. In this paper we present an innovative modelling approach that is based upon imitation learning or Programming by Demonstration (PbD). Our approach-based on Dirichlet Process Mixture Models, Hidden Markov Models, Dynamic Time Warping, and non-uniform cubic spline regression-is particularly innovative as it handles spatial and temporal variability by the generation of a generalised trajectory from a set of periodically repeated movements. Although not within the scope of our study, our procedures may be implemented for the sake of controlling movement behaviour of robots and avatar animations in response to music.

Cite

CITATION STYLE

APA

Amelynck, D., Maes, P.-J., Martens, J.-P., & Leman, M. (2017). Beating-Time Gestures Imitation Learning for Humanoid Robots. EAI Endorsed Transactions on Creative Technologies, 4(13), 153335. https://doi.org/10.4108/eai.8-11-2017.153335

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free