On Regularized Losses for Weakly-supervised CNN Segmentation
This paper proposes and experimentally compares different losses integrating MRF/CRF regularization terms.
September 8, 2018
European Conference on Computer Vision (ECCV) 2018
Authors
Meng Tang (University of Waterloo)
Federico Perazzi (Adobe Research)
Aziz Djelouah (Disney Research)
Ismail Ben Ayed (ETS Montreal)
Christopher Schroers (Disney Research)
Yuri Boykov (University of Waterloo)
Minimization of regularized losses is a principled approach to weak supervision well-established in deep learning, in general. However, it is largely overlooked in semantic segmentation currently dominated by methods mimicking full supervision via “fake” fully-labeled training masks (proposals) generated from available partial input. To obtain such full masks the typical methods explicitly use standard regularization techniques for “shallow” segmentation, e.g. graph cuts or dense CRFs. In contrast, we integrate such standard regularizers directly into the loss functions over partial input. This approach simplifies weakly-supervised training by avoiding extra MRF/CRF inference steps or layers explicitly generating full masks, while improving both the quality and efficiency of training. This paper proposes and experimentally compares different losses integrating MRF/CRF regularization terms. We juxtapose our regularized losses with earlier proposal-generation methods using explicit regularization steps or layers. Our approach achieves state-of-the-art accuracy in semantic segmentation with near full-supervision quality.