@inproceedings{d8de495115f14af18add88a471527034,
title = "Generative Modeling for Multi-task Visual Learning",
abstract = "Generative modeling has recently shown great promise in computer vision, but it has mostly focused on synthesizing visually realistic images. In this paper, motivated by multi-task learning of shareable feature representations, we consider a novel problem of learning a shared generative model that is useful across various visual perception tasks. Correspondingly, we propose a general multi-task oriented generative modeling (MGM) framework, by coupling a discriminative multi-task network with a generative network. While it is challenging to synthesize both RGB images and pixel-level annotations in multi-task scenarios, our framework enables us to use synthesized images paired with only weak annotations (i.e., image-level scene labels) to facilitate multiple visual tasks. Experimental evaluation on challenging multi-task benchmarks, including NYUv2 and Taskonomy, demonstrates that our MGM framework improves the performance of all the tasks by large margins, consistently outperforming state-of-the-art multi-task approaches in different sample-size regimes.",
author = "Zhipeng Bao and Martial Hebert and Yu-Xiong Wang",
year = "2022",
month = may,
day = "1",
language = "English (US)",
volume = "162",
series = "Proceedings of Machine Learning Research",
publisher = "PMLR",
pages = "1537--1554",
editor = "Kamalika Chaudhuri and Stefanie Jegelka and Le Song and Csaba Szepesvari and Gang Niu and Sivan Sabato",
booktitle = "Proceedings of the 39th International Conference on Machine Learning",
}