Co-attention CNNs for Unsupervised Object Co-segmentation
Co-attention CNNs for Unsupervised Object Co-segmentation
Kuang-Jui Hsu, Yen-Yu Lin, Yung-Yu Chuang
Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 748-756.
https://doi.org/10.24963/ijcai.2018/104
Object co-segmentation aims to segment the common objects in images. This paper presents a CNN-based method that is unsupervised and end-to-end trainable to better solve this task. Our method is unsupervised in the sense that it does not require any training data in the form of object masks but merely a set of images jointly covering objects of a specific class. Our method comprises two collaborative CNN modules, a feature extractor and a co-attention map generator. The former module extracts the features of the estimated objects and backgrounds, and is derived based on the proposed co-attention loss which minimizes inter-image object discrepancy while maximizing intra-image figure-ground separation. The latter module is learned to generated co-attention maps by which the estimated figure-ground segmentation can better fit the former module. Besides, the co-attention loss, the mask loss is developed to retain the whole objects and remove noises. Experiments show that our method achieves superior results, even outperforming the state-of-the-art, supervised methods.
Keywords:
Machine Learning: Unsupervised Learning
Machine Learning: Deep Learning
Computer Vision: Computer Vision