File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Michelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation

TitleMichelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation
Authors
Issue Date2023
Citation
Advances in Neural Information Processing Systems, 2023, v. 36 How to Cite?
AbstractWe present a novel alignment-before-generation approach to tackle the challenging task of generating general 3D shapes based on 2D images or texts. Directly learning a conditional generative model from images or texts to 3D shapes is prone to produce inconsistent results with the conditions because 3D shapes have an additional dimension whose distribution significantly differs from that of 2D images and texts. To bridge the domain gap among the three modalities and facilitate multi-modal-conditioned 3D shape generation, we explore representing 3D shapes in a shape-image-text-aligned space. Our framework comprises two models: a Shape-Image-Text-Aligned Variational Auto-Encoder (SITA-VAE) and a conditional Aligned Shape Latent Diffusion Model (ASLDM). The former model encodes the 3D shapes into the shape latent space aligned to the image and text and reconstructs the fine-grained 3D neural fields corresponding to given shape embeddings via the transformer-based decoder. The latter model learns a probabilistic mapping function from the image or text space to the latent shape space. Extensive experiments demonstrate that our proposed approach can generate higher-quality and more diverse 3D shapes that better semantically conform to the visual or textural conditional inputs, validating the effectiveness of the shape-image-text-aligned space for cross-modality 3D shape generation.
Persistent Identifierhttp://hdl.handle.net/10722/345378
ISSN
2020 SCImago Journal Rankings: 1.399

 

DC FieldValueLanguage
dc.contributor.authorZhao, Zibo-
dc.contributor.authorLiu, Wen-
dc.contributor.authorChen, Xin-
dc.contributor.authorZeng, Xianfang-
dc.contributor.authorWang, Rui-
dc.contributor.authorCheng, Pei-
dc.contributor.authorFu, Bin-
dc.contributor.authorChen, Tao-
dc.contributor.authorYu, Gang-
dc.contributor.authorGao, Shenghua-
dc.date.accessioned2024-08-15T09:26:58Z-
dc.date.available2024-08-15T09:26:58Z-
dc.date.issued2023-
dc.identifier.citationAdvances in Neural Information Processing Systems, 2023, v. 36-
dc.identifier.issn1049-5258-
dc.identifier.urihttp://hdl.handle.net/10722/345378-
dc.description.abstractWe present a novel alignment-before-generation approach to tackle the challenging task of generating general 3D shapes based on 2D images or texts. Directly learning a conditional generative model from images or texts to 3D shapes is prone to produce inconsistent results with the conditions because 3D shapes have an additional dimension whose distribution significantly differs from that of 2D images and texts. To bridge the domain gap among the three modalities and facilitate multi-modal-conditioned 3D shape generation, we explore representing 3D shapes in a shape-image-text-aligned space. Our framework comprises two models: a Shape-Image-Text-Aligned Variational Auto-Encoder (SITA-VAE) and a conditional Aligned Shape Latent Diffusion Model (ASLDM). The former model encodes the 3D shapes into the shape latent space aligned to the image and text and reconstructs the fine-grained 3D neural fields corresponding to given shape embeddings via the transformer-based decoder. The latter model learns a probabilistic mapping function from the image or text space to the latent shape space. Extensive experiments demonstrate that our proposed approach can generate higher-quality and more diverse 3D shapes that better semantically conform to the visual or textural conditional inputs, validating the effectiveness of the shape-image-text-aligned space for cross-modality 3D shape generation.-
dc.languageeng-
dc.relation.ispartofAdvances in Neural Information Processing Systems-
dc.titleMichelangelo: Conditional 3D Shape Generation based on Shape-Image-Text Aligned Latent Representation-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85188850274-
dc.identifier.volume36-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats