Generating Annotated High-Fidelity Images Containing Multiple Coherent Objects

Authors
Publication date 2021
Book title 2021 IEEE International Conference on Image Processing
Book subtitle proceedings : 19-22 September 2021, Anchorage, Alaska, USA
ISBN
  • 9781665431026
ISBN (electronic)
  • 9781665441155
Series ICIP
Event IEEE International Conference on Image Processing 19-22 Sept. 2021
Pages (from-to) 834-838
Publisher Piscataway, NJ: IEEE
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Recent developments related to generative models have enabled the generation of diverse and high-fidelity images. In particular, layout-to-image generation models have gained significant attention due to their capability to generate realistic and complex images containing distinct objects. These models are generally conditioned on either semantic layouts or textual descriptions. However, unlike natural images, providing auxiliary information can be extremely hard in domains such as biomedical imaging and remote sensing. In this work, we propose a multi-object generation framework1 that can synthesize images with multiple objects without explicitly requiring their contextual information during the generation process. Based on a vector-quantized variational autoencoder (VQ-VAE) backbone, our model learns to preserve spatial coherency within an image as well as semantic coherency through the use of powerful autoregressive priors. An advantage of our approach is that the generated samples are accompanied by object-level annotations. The efficacy of our approach is demonstrated through application on medical imaging datasets, where we show that augmenting the training set with the samples generated by our approach improves the performance of existing models.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/ICIP42928.2021.9506406
Other links https://www.proceedings.com/64071.html
Permalink to this page
Back