Co-attention CNNs for unsupervised object co-segmentation

70Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Object co-segmentation aims to segment the common objects in images. This paper presents a CNN-based method that is unsupervised and end-to-end trainable to better solve this task. Our method is unsupervised in the sense that it does not require any training data in the form of object masks but merely a set of images jointly covering objects of a specific class. Our method comprises two collaborative CNN modules, a feature extractor and a co-attention map generator. The former module extracts the features of the estimated objects and backgrounds, and is derived based on the proposed co-attention loss, which minimizes inter-image object discrepancy while maximizing intra-image figure-ground separation. The latter module is learned to generate co-attention maps by which the estimated figure-ground segmentation can better fit the former module. Besides the co-attention loss, the mask loss is developed to retain the whole objects and remove noises. Experiments show that our method achieves superior results, even outperforming the state-of-the-art, supervised methods.

Cite

CITATION STYLE

APA

Hsu, K. J., Lin, Y. Y., & Chuang, Y. Y. (2018). Co-attention CNNs for unsupervised object co-segmentation. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2018-July, pp. 748–756). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2018/104

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free