VideoIC: A Video Interactive Comments Dataset and Multimodal Multitask Learning for Comments Generation

18Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Live video interactive commenting, a.k.a. danmaku, is an emerging social feature on online video sites, which involves rich multimodal information interaction among viewers. In order to support various related research, we build a large scale video interactive comments dataset called VideoIC, which consists of 4951 videos spanning 557 hours and 5 million comments. Videos are collected from popular categories on the 'Bilibili' video streaming website. Comparing to other existing danmaku datasets, our VideoIC contains richer and denser comments information, with 1077 comments per video on average. High comment density and diverse video types make VideoIC a challenging corpus for various research such as automatic video comments generation. We also propose a novel model based on multimodal multitask learning for comment generation (MML-CG), which integrates multiple modalities to achieve effective comment generation and temporal relation prediction. A multitask loss function is designed to train both tasks jointly in the end-to-end manner. We conduct extensive experiments on both VideoIC and Livebot datasets. The results prove the effectiveness of our model and reveal some features of danmaku.

Cite

CITATION STYLE

APA

Wang, W., Chen, J., & Jin, Q. (2020). VideoIC: A Video Interactive Comments Dataset and Multimodal Multitask Learning for Comments Generation. In MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia (pp. 2599–2607). Association for Computing Machinery, Inc. https://doi.org/10.1145/3394171.3413890

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free