MODNet: Real-Time Trimap-Free Portrait Matting via Objective Decomposition
Zhanghan Ke, Jiayu Sun, Kaican Li, Qiong Yan, Rynson W. H. Lau
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/ZHKKKe/MODNetOfficialIn paperpytorch★ 4,278
- github.com/ZHKKKe/PPMOfficialIn papernone★ 102
- github.com/PaddlePaddle/PaddleSegpaddle★ 9,319
- github.com/royshil/obs-backgroundremovalpaddle★ 4,202
- github.com/DeepranjanG/Image_Background_Removalpytorch★ 14
- github.com/Mind23-2/MindCode-101/tree/main/MODNetmindspore★ 0
- github.com/2023-MindSpore-4/Code-5/tree/main/MODNetmindspore★ 0
- github.com/MS-Mind/MS-Code-08/tree/main/MODNetmindspore★ 0
- github.com/code-implementation1/Code5/tree/main/MODNetmindspore★ 0
Abstract
Existing portrait matting methods either require auxiliary inputs that are costly to obtain or involve multiple stages that are computationally expensive, making them less suitable for real-time applications. In this work, we present a light-weight matting objective decomposition network (MODNet) for portrait matting in real-time with a single input image. The key idea behind our efficient design is by optimizing a series of sub-objectives simultaneously via explicit constraints. In addition, MODNet includes two novel techniques for improving model efficiency and robustness. First, an Efficient Atrous Spatial Pyramid Pooling (e-ASPP) module is introduced to fuse multi-scale features for semantic estimation. Second, a self-supervised sub-objectives consistency (SOC) strategy is proposed to adapt MODNet to real-world data to address the domain shift problem common to trimap-free methods. MODNet is easy to be trained in an end-to-end manner. It is much faster than contemporaneous methods and runs at 67 frames per second on a 1080Ti GPU. Experiments show that MODNet outperforms prior trimap-free methods by a large margin on both Adobe Matting Dataset and a carefully designed photographic portrait matting (PPM-100) benchmark proposed by us. Further, MODNet achieves remarkable results on daily photos and videos. Our code and models are available at https://github.com/ZHKKKe/MODNet, and the PPM-100 benchmark is released at https://github.com/ZHKKKe/PPM.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| AMD | MODNet+ | MAD | 0.81 | — | Unverified |
| PPM-100 | MODNet+ (Our) | MAD | 0.97 | — | Unverified |