The increasing amounts of media becoming available in converged digital broadcast and mobile broadband networks will require intelligent interfaces capable of personalizing the selection of content. Aiming to capture the mood in the content, we construct a semantic space based on tags, frequently used to describe emotions associated with music in the last.fm social network. Implementing latent semantic analysis (LSA), we model the affective context of songs based on their lyrics, and apply a similar approach to extract moods from BBC synopsis descriptions of TV episodes using TV-Anytime atmosphere terms. Based on our early results, we propose that LSA could be implemented as machinelearning method to extract emotional context and model affective user preferences.
CITATION STYLE
Petersen, M. K., & Butkus, A. (2008). Extracting Moods from Songs and BBC Programs Based on Emotional Context. International Journal of Digital Multimedia Broadcasting, 2008, 1–12. https://doi.org/10.1155/2008/289837
Mendeley helps you to discover research relevant for your work.