Recently there is an increasing interest in scene generation within the research community. However, scene layouts are largely being modeled in deterministic fashion, ignoring any plausible visual variations given the same textual description as input. We propose LayoutVAE, a variational autoencoder based framework for generating stochastic scene layouts. LayoutVAE is a versatile modeling framework that allows for generating full image layouts given a label set, or per label layouts for an existing image given a new label. In addition, it is also capable of detecting unusual layouts, potentially providing a way to evaluate layout generation problem. Extensive experiments on MNIST-Layouts and challenging COCO 2017 Panoptic dataset verifies the effectiveness of our proposed framework.
Bibtex
@inproceedings{JyothiDHSM19,
title = {LayoutVAE: Stochastic Scene Layout Generation from a Label Set},
author = {Akash Abdu Jyothi and Thibaut Durand and Jiawei He and Leonid Sigal and Greg Mori},
booktitle = {International Conference on Computer Vision (ICCV)},
year = 2019,
}
Related Research
-
What you need to know about our top 5 research papers at CVPR 2022
What you need to know about our top 5 research papers at CVPR 2022
D. Silverberg, and M. Zhai.
Research
-
Adversarial Robustness with AdverTorch Black-box Estimators
Adversarial Robustness with AdverTorch Black-box Estimators
M. Safa, and G. Castiglione.
Computer Vision; Natural Language Processing; Responsible AI
Research
-
Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation
Why Exposure Bias Matters: An Imitation Learning Perspective of Error Accumulation in Language Generation
K. Arora, L. El Asri, H. Bahuleyan, and J. Chi Kit Cheung. Association for Computational Linguistics (ACL)
Publications