TY - GEN
T1 - I-Matting
T2 - 2024 IEEE International Conference on Multimedia and Expo, ICME 2024
AU - Liu, Zichuan
AU - Wang, Ke
AU - Wu, Mingyuan
AU - Yu, Lantao
AU - Nahrstedt, Klara
AU - Lu, Xin
N1 - Mingyuan was partially funded by the National Science Foundation under grant contracts NSF 1835834, NSF 1900875, NSF 2106592. Any results and opinions do not represent views of National Science Foundation.
PY - 2024
Y1 - 2024
N2 - Image matting has become an essential functionality of image capturing and editing tools. While trimap and scribble-based techniques have shown notable success in these applications, generating high-quality alpha mattes without trimap inputs remains challenging. Existing trimap-free methods divide the task into coarse semantic mask prediction and detailed matte prediction, and an optimization is formulated by balancing these two tasks. However, emphasizing the optimization of the coarse mask leads to inaccurate matte, and emphasizing the optimization of the detailed matte leads to degraded semantic integrity or background artifacts. In this paper, we propose an improved trimap-free training strategy (I-Matting) that effectively ensures semantic integrity, removes background artifacts, and improves local details. First, we introduce two discriminators to distinguish the matting outputs versus the ground truths, which boosts the semantic without hurting the matte prediction. Second, a novel patch-rank module is proposed to improve the matting accuracy by leveraging high-resolution inputs, without hurting the semantic integrity. Meanwhile, the accuracy gain produced by I-Matting is not at the expense of any additional cost in the inference. Extensive experiments show that our method significantly outperforms existing approaches.
AB - Image matting has become an essential functionality of image capturing and editing tools. While trimap and scribble-based techniques have shown notable success in these applications, generating high-quality alpha mattes without trimap inputs remains challenging. Existing trimap-free methods divide the task into coarse semantic mask prediction and detailed matte prediction, and an optimization is formulated by balancing these two tasks. However, emphasizing the optimization of the coarse mask leads to inaccurate matte, and emphasizing the optimization of the detailed matte leads to degraded semantic integrity or background artifacts. In this paper, we propose an improved trimap-free training strategy (I-Matting) that effectively ensures semantic integrity, removes background artifacts, and improves local details. First, we introduce two discriminators to distinguish the matting outputs versus the ground truths, which boosts the semantic without hurting the matte prediction. Second, a novel patch-rank module is proposed to improve the matting accuracy by leveraging high-resolution inputs, without hurting the semantic integrity. Meanwhile, the accuracy gain produced by I-Matting is not at the expense of any additional cost in the inference. Extensive experiments show that our method significantly outperforms existing approaches.
UR - http://www.scopus.com/inward/record.url?scp=85206563821&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85206563821&partnerID=8YFLogxK
U2 - 10.1109/ICME57554.2024.10687689
DO - 10.1109/ICME57554.2024.10687689
M3 - Conference contribution
AN - SCOPUS:85206563821
T3 - Proceedings - IEEE International Conference on Multimedia and Expo
BT - 2024 IEEE International Conference on Multimedia and Expo, ICME 2024
PB - IEEE Computer Society
Y2 - 15 July 2024 through 19 July 2024
ER -