SOTAVerified

InfoSeg: Unsupervised Semantic Image Segmentation with Mutual Information Maximization

2021-10-07Unverified0· sign in to hype

Robert Harb, Patrick Knöbelreiter

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

We propose a novel method for unsupervised semantic image segmentation based on mutual information maximization between local and global high-level image features. The core idea of our work is to leverage recent progress in self-supervised image representation learning. Representation learning methods compute a single high-level feature capturing an entire image. In contrast, we compute multiple high-level features, each capturing image segments of one particular semantic class. To this end, we propose a novel two-step learning procedure comprising a segmentation and a mutual information maximization step. In the first step, we segment images based on local and global features. In the second step, we maximize the mutual information between local features and high-level features of their respective class. For training, we provide solely unlabeled images and start from random network initialization. For quantitative and qualitative evaluation, we use established benchmarks, and COCO-Persons, whereby we introduce the latter in this paper as a challenging novel benchmark. InfoSeg significantly outperforms the current state-of-the-art, e.g., we achieve a relative increase of 26% in the Pixel Accuracy metric on the COCO-Stuff dataset.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
COCO-PersonsInfoSegPixel Accuracy69.6Unverified
COCO-Stuff-15InfoSegPixel Accuracy38.8Unverified
COCO-Stuff-3InfoSegPixel Accuracy73.8Unverified
Potsdam-3InfoSegPixel Accuracy71.6Unverified

Reproductions