File Download
  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Learning to predict layout-to-image conditional convolutions for semantic image synthesis

TitleLearning to predict layout-to-image conditional convolutions for semantic image synthesis
Authors
Issue Date2019
Citation
Advances in Neural Information Processing Systems, 2019, v. 32 How to Cite?
AbstractSemantic image synthesis aims at generating photorealistic images from semantic layouts. Previous approaches with conditional generative adversarial networks (GAN) show state-of-the-art performance on this task, which either feed the semantic label maps as inputs to the generator, or use them to modulate the activations in normalization layers via affine transformations. We argue that convolutional kernels in the generator should be aware of the distinct semantic labels at different locations when generating images. In order to better exploit the semantic layout for the image generator, we propose to predict convolutional kernels conditioned on the semantic label map to generate the intermediate feature maps from the noise maps and eventually generate the images. Moreover, we propose a feature pyramid semantics-embedding discriminator, which is more effective in enhancing fine details and semantic alignments between the generated images and the input semantic layouts than previous multi-scale discriminators. We achieve state-of-the-art results on both quantitative metrics and subjective evaluation on various semantic segmentation datasets, demonstrating the effectiveness of our approach.
Persistent Identifierhttp://hdl.handle.net/10722/316550
ISSN
2020 SCImago Journal Rankings: 1.399
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorLiu, Xihui-
dc.contributor.authorShao, Jing-
dc.contributor.authorYin, Guojun-
dc.contributor.authorWang, Xiaogang-
dc.contributor.authorLi, Hongsheng-
dc.date.accessioned2022-09-14T11:40:44Z-
dc.date.available2022-09-14T11:40:44Z-
dc.date.issued2019-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2019, v. 32-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/316550-
dc.description.abstractSemantic image synthesis aims at generating photorealistic images from semantic layouts. Previous approaches with conditional generative adversarial networks (GAN) show state-of-the-art performance on this task, which either feed the semantic label maps as inputs to the generator, or use them to modulate the activations in normalization layers via affine transformations. We argue that convolutional kernels in the generator should be aware of the distinct semantic labels at different locations when generating images. In order to better exploit the semantic layout for the image generator, we propose to predict convolutional kernels conditioned on the semantic label map to generate the intermediate feature maps from the noise maps and eventually generate the images. Moreover, we propose a feature pyramid semantics-embedding discriminator, which is more effective in enhancing fine details and semantic alignments between the generated images and the input semantic layouts than previous multi-scale discriminators. We achieve state-of-the-art results on both quantitative metrics and subjective evaluation on various semantic segmentation datasets, demonstrating the effectiveness of our approach.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleLearning to predict layout-to-image conditional convolutions for semantic image synthesis-
dc.typeConference_Paper-
dc.description.naturelink_to_OA_fulltext-
dc.identifier.scopuseid_2-s2.0-85088122806-
dc.identifier.volume32-
dc.identifier.isiWOS:000534424300052-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats