File Download

Conference Paper: SeSDF: Self-evolved Signed Distance Field for Implicit 3D Clothed Human Reconstruction

TitleSeSDF: Self-evolved Signed Distance Field for Implicit 3D Clothed Human Reconstruction
Authors
Issue Date18-Jun-2023
Abstract

We address the problem of clothed human reconstruction from a single image or uncalibrated multi-view images. Existing methods struggle with reconstructing detailed geometry of a clothed human and often require a calibrated setting for multi-view reconstruction. We propose a flexible framework which, by leveraging the parametric SMPL-X model, can take an arbitrary number of input images to reconstruct a clothed human model under an uncalibrated setting. At the core of our framework is our novel self-evolved signed distance field (SeSDF) module which allows the framework to learn to deform the signed distance field (SDF) derived from the fitted SMPL-X model, such that detailed geometry reflecting the actual clothed human can be encoded for better reconstruction. Besides, we propose a simple method for self-calibration of multi-view images via the fitted SMPLX parameters. This lifts the requirement of tedious manual calibration and largely increases the flexibility of our method. Further, we introduce an effective occlusion-aware feature fusion strategy to account for the most useful features to reconstruct the human model. We thoroughly evaluate our framework on public benchmarks, demonstrating significant superiority over the state-of-the-arts both qualitatively and quantitatively.


Persistent Identifierhttp://hdl.handle.net/10722/337254

 

DC FieldValueLanguage
dc.contributor.authorCao, Yukang-
dc.contributor.authorHan, Kai-
dc.contributor.authorWong, Kwan-Yee K-
dc.date.accessioned2024-03-11T10:19:16Z-
dc.date.available2024-03-11T10:19:16Z-
dc.date.issued2023-06-18-
dc.identifier.urihttp://hdl.handle.net/10722/337254-
dc.description.abstract<p>We address the problem of clothed human reconstruction from a single image or uncalibrated multi-view images. Existing methods struggle with reconstructing detailed geometry of a clothed human and often require a calibrated setting for multi-view reconstruction. We propose a flexible framework which, by leveraging the parametric SMPL-X model, can take an arbitrary number of input images to reconstruct a clothed human model under an uncalibrated setting. At the core of our framework is our novel self-evolved signed distance field (SeSDF) module which allows the framework to learn to deform the signed distance field (SDF) derived from the fitted SMPL-X model, such that detailed geometry reflecting the actual clothed human can be encoded for better reconstruction. Besides, we propose a simple method for self-calibration of multi-view images via the fitted SMPLX parameters. This lifts the requirement of tedious manual calibration and largely increases the flexibility of our method. Further, we introduce an effective occlusion-aware feature fusion strategy to account for the most useful features to reconstruct the human model. We thoroughly evaluate our framework on public benchmarks, demonstrating significant superiority over the state-of-the-arts both qualitatively and quantitatively.<br></p>-
dc.languageeng-
dc.relation.ispartofIEEE/CVF Conference on Computer Vision and Pattern Recognition (18/06/2023-22/06/2023, Vancouver)-
dc.titleSeSDF: Self-evolved Signed Distance Field for Implicit 3D Clothed Human Reconstruction-
dc.typeConference_Paper-
dc.description.naturepublished_or_final_version-
dc.identifier.spage4647-
dc.identifier.epage4657-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats