Video Object Segmentation with Language Referring Expressions

52Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Most state-of-the-art semi-supervised video object segmentation methods rely on a pixel-accurate mask of a target object provided for the first frame of a video. However, obtaining a detailed segmentation mask is expensive and time-consuming. In this work we explore an alternative way of identifying a target object, namely by employing language referring expressions. Besides being a more practical and natural way of pointing out a target object, using language specifications can help to avoid drift as well as make the system more robust to complex dynamics and appearance variations. Leveraging recent advances of language grounding models designed for images, we propose an approach to extend them to video data, ensuring temporally coherent predictions. To evaluate our approach we augment the popular video object segmentation benchmarks, $${\text {DAVIS}}_{{16}}$$ and $${\text {DAVIS}}_{{17}}$$ with language descriptions of target objects. We show that our language-supervised approach performs on par with the methods which have access to a pixel-level mask of the target object on $${\text {DAVIS}}_{{16}}$$ and is competitive to methods using scribbles on the challenging $${\text {DAVIS}}_{{17}}$$ dataset.

Cite

CITATION STYLE

APA

Khoreva, A., Rohrbach, A., & Schiele, B. (2019). Video Object Segmentation with Language Referring Expressions. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11364 LNCS, pp. 123–141). Springer Verlag. https://doi.org/10.1007/978-3-030-20870-7_8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free