File Download
Supplementary

postgraduate thesis: Neural LiDAR fields for resimulation and scene reconstruction

TitleNeural LiDAR fields for resimulation and scene reconstruction
Authors
Advisors
Advisor(s):Zhang, FLam, J
Issue Date2025
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Hou, C. [侯超]. (2025). Neural LiDAR fields for resimulation and scene reconstruction. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractLiDAR (Light Detection and Ranging) is a remote sensing method that uses laser pulses to capture precise distance measurements. By recording the time it takes for light to bounce off surfaces and return to the sensor, LiDAR generates dense 3D point clouds of the surrounding environment. Its high accuracy and efficiency make it a key technology in robotics applications such as navigation, obstacle avoidance, and localization. The rapid development of autonomous driving and related robotics technologies has led to an increasing demand for large-scale annotated datasets. However, the distribution of real-world data often follows a long-tail pattern, making it challenging to collect sufficient samples of rare corner cases. Simulators offer a promising solution by enabling the reconstruction and resimulation of such cases, thereby enriching the available data. Compared to traditional simulators that rely on manually created 3D assets, data-driven simulators are gaining attention for their ability to bridge the sim-to-real gap and eliminate the need for labor-intensive, handcrafted design. In this thesis, we leverage the realistic rendering capabilities of Neural Radiance Fields (NeRFs) to focus on LiDAR resimulation and scene reconstruction. We first present a novel LiDAR Fields framework for space-time view synthesis of dynamic urban scenes. We observe that existing methods relying on U-Net architectures to simulate ray-drop characteristics struggle to effectively capture this pattern. By leveraging LiDAR positional embeddings, we are able to render realistic LiDAR patterns in range images. However, due to the inherent sparsity and occlusions in Li- DAR data—particularly around dynamic objects—accurate and complete reconstruction remains a significant challenge. To address this, we estimate dense and consistent scene flow through a self-supervised loss, enabling improved modeling of dynamic motion and structure. Next, we develop a NeRF training pipeline that ensures consistent reconstruction across different sensing modalities while preserving their unique characteristics. Our design leverages the complementary nature of camera and LiDAR inputs to achieve more robust reconstruction, while also mitigating cross-modal noise that degrades the performance of each individual modality.
DegreeMaster of Philosophy
SubjectOptical radar
Computer vision
Robotics
Dept/ProgramMechanical Engineering
Persistent Identifierhttp://hdl.handle.net/10722/367427

 

DC FieldValueLanguage
dc.contributor.advisorZhang, F-
dc.contributor.advisorLam, J-
dc.contributor.authorHou, Chao-
dc.contributor.author侯超-
dc.date.accessioned2025-12-11T06:41:55Z-
dc.date.available2025-12-11T06:41:55Z-
dc.date.issued2025-
dc.identifier.citationHou, C. [侯超]. (2025). Neural LiDAR fields for resimulation and scene reconstruction. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/367427-
dc.description.abstractLiDAR (Light Detection and Ranging) is a remote sensing method that uses laser pulses to capture precise distance measurements. By recording the time it takes for light to bounce off surfaces and return to the sensor, LiDAR generates dense 3D point clouds of the surrounding environment. Its high accuracy and efficiency make it a key technology in robotics applications such as navigation, obstacle avoidance, and localization. The rapid development of autonomous driving and related robotics technologies has led to an increasing demand for large-scale annotated datasets. However, the distribution of real-world data often follows a long-tail pattern, making it challenging to collect sufficient samples of rare corner cases. Simulators offer a promising solution by enabling the reconstruction and resimulation of such cases, thereby enriching the available data. Compared to traditional simulators that rely on manually created 3D assets, data-driven simulators are gaining attention for their ability to bridge the sim-to-real gap and eliminate the need for labor-intensive, handcrafted design. In this thesis, we leverage the realistic rendering capabilities of Neural Radiance Fields (NeRFs) to focus on LiDAR resimulation and scene reconstruction. We first present a novel LiDAR Fields framework for space-time view synthesis of dynamic urban scenes. We observe that existing methods relying on U-Net architectures to simulate ray-drop characteristics struggle to effectively capture this pattern. By leveraging LiDAR positional embeddings, we are able to render realistic LiDAR patterns in range images. However, due to the inherent sparsity and occlusions in Li- DAR data—particularly around dynamic objects—accurate and complete reconstruction remains a significant challenge. To address this, we estimate dense and consistent scene flow through a self-supervised loss, enabling improved modeling of dynamic motion and structure. Next, we develop a NeRF training pipeline that ensures consistent reconstruction across different sensing modalities while preserving their unique characteristics. Our design leverages the complementary nature of camera and LiDAR inputs to achieve more robust reconstruction, while also mitigating cross-modal noise that degrades the performance of each individual modality.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshOptical radar-
dc.subject.lcshComputer vision-
dc.subject.lcshRobotics-
dc.titleNeural LiDAR fields for resimulation and scene reconstruction-
dc.typePG_Thesis-
dc.description.thesisnameMaster of Philosophy-
dc.description.thesislevelMaster-
dc.description.thesisdisciplineMechanical Engineering-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2025-
dc.identifier.mmsid991045147148103414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats