Abstract
Object co-segmentation aims to segment the common objects in images. This paper presents a CNN-based method that is unsupervised and end-to-end trainable to better solve this task. Our method is unsupervised in the sense that it does not require any training data in the form of object masks but merely a set of images jointly covering objects of a specific class. Our method comprises two collaborative CNN modules, a feature extractor and a co-attention map generator. The former module extracts the features of the estimated objects and backgrounds, and is derived based on the proposed co-attention loss, which minimizes inter-image object discrepancy while maximizing intra-image figure-ground separation. The latter module is learned to generate co-attention maps by which the estimated figure-ground segmentation can better fit the former module. Besides the co-attention loss, the mask loss is developed to retain the whole objects and remove noises. Experiments show that our method achieves superior results, even outperforming the state-of-the-art, supervised methods.
Cite
CITATION STYLE
Hsu, K. J., Lin, Y. Y., & Chuang, Y. Y. (2018). Co-attention CNNs for unsupervised object co-segmentation. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2018-July, pp. 748–756). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2018/104
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.