File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: GRNet: Geometric relation network for 3D object detection from point clouds

TitleGRNet: Geometric relation network for 3D object detection from point clouds
Authors
Keywords3D object detection
Deep learning
Geometric relation
Indoor mapping
Point cloud
RGB-D
Issue Date2020
Citation
ISPRS Journal of Photogrammetry and Remote Sensing, 2020, v. 165, p. 43-53 How to Cite?
AbstractRapid detection of 3D objects in indoor environments is essential for indoor mapping and modeling, robotic perception and localization, and building reconstruction. 3D point clouds acquired by a low-cost RGB-D camera have become one of the most commonly used data sources for 3D indoor mapping. However, due to the sparse surface, empty object center, and various scales of point cloud objects, 3D bounding boxes are challenging to be estimated and located accurately. To address this, geometric shape, topological structure, and object relation are commonly employed to extract box reasoning information. In this paper, we describe the geometric feature among object points as an intra-object feature and the relation feature between different objects as an inter-object feature. Based on these two features, we propose an end-to-end point cloud geometric relation network focusing on 3D object detection, which is termed as geometric relation network (GRNet). GRNet first extracts intra-object and inter-object features for each representative point using our proposed backbone network. Then, a centralization module with a scalable loss function is proposed to centralize each representative object point to its center. Next, proposal points are sampled from these shifted points, following a proposal feature pooling operation. Finally, an object-relation learning module is applied to predict bounding box parameters. Such parameters are the additive sum of prediction results from the relation-based inter-object feature and the aggregated intra-object feature. Our model achieves state-of-the-art 3D detection results with 59.1% mAP@0.25 and 39.1% mAP@0.5 on ScanNetV2 dataset, 58.4% mAP@0.25 and 34.9% mAP@0.5 on SUN RGB-D dataset.
Persistent Identifierhttp://hdl.handle.net/10722/352991
ISSN
2023 Impact Factor: 10.6
2023 SCImago Journal Rankings: 3.760
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorLi, Ying-
dc.contributor.authorMa, Lingfei-
dc.contributor.authorTan, Weikai-
dc.contributor.authorSun, Chen-
dc.contributor.authorCao, Dongpu-
dc.contributor.authorLi, Jonathan-
dc.date.accessioned2025-01-13T03:01:30Z-
dc.date.available2025-01-13T03:01:30Z-
dc.date.issued2020-
dc.identifier.citationISPRS Journal of Photogrammetry and Remote Sensing, 2020, v. 165, p. 43-53-
dc.identifier.issn0924-2716-
dc.identifier.urihttp://hdl.handle.net/10722/352991-
dc.description.abstractRapid detection of 3D objects in indoor environments is essential for indoor mapping and modeling, robotic perception and localization, and building reconstruction. 3D point clouds acquired by a low-cost RGB-D camera have become one of the most commonly used data sources for 3D indoor mapping. However, due to the sparse surface, empty object center, and various scales of point cloud objects, 3D bounding boxes are challenging to be estimated and located accurately. To address this, geometric shape, topological structure, and object relation are commonly employed to extract box reasoning information. In this paper, we describe the geometric feature among object points as an intra-object feature and the relation feature between different objects as an inter-object feature. Based on these two features, we propose an end-to-end point cloud geometric relation network focusing on 3D object detection, which is termed as geometric relation network (GRNet). GRNet first extracts intra-object and inter-object features for each representative point using our proposed backbone network. Then, a centralization module with a scalable loss function is proposed to centralize each representative object point to its center. Next, proposal points are sampled from these shifted points, following a proposal feature pooling operation. Finally, an object-relation learning module is applied to predict bounding box parameters. Such parameters are the additive sum of prediction results from the relation-based inter-object feature and the aggregated intra-object feature. Our model achieves state-of-the-art 3D detection results with 59.1% mAP@0.25 and 39.1% mAP@0.5 on ScanNetV2 dataset, 58.4% mAP@0.25 and 34.9% mAP@0.5 on SUN RGB-D dataset.-
dc.languageeng-
dc.relation.ispartofISPRS Journal of Photogrammetry and Remote Sensing-
dc.subject3D object detection-
dc.subjectDeep learning-
dc.subjectGeometric relation-
dc.subjectIndoor mapping-
dc.subjectPoint cloud-
dc.subjectRGB-D-
dc.titleGRNet: Geometric relation network for 3D object detection from point clouds-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1016/j.isprsjprs.2020.05.008-
dc.identifier.scopuseid_2-s2.0-85085217744-
dc.identifier.volume165-
dc.identifier.spage43-
dc.identifier.epage53-
dc.identifier.isiWOS:000540373500004-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats