File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Improving Probe Localization for Freehand 3D Ultrasound Using Lightweight Cameras

TitleImproving Probe Localization for Freehand 3D Ultrasound Using Lightweight Cameras
Authors
Issue Date2025
Citation
Proceedings IEEE International Conference on Robotics and Automation, 2025, p. 15472-15478 How to Cite?
AbstractUltrasound (US) probe localization relative to the examined subject is essential for freehand 3D US imaging, which offers significant clinical value due to its affordability and unrestricted field of view. However, existing methods often rely on expensive tracking systems or bulky probes, while recent US image-based deep learning methods suffer from accumulated errors during probe maneuvering. To address these challenges, this study proposes a versatile, cost-effective probe pose localization method for freehand 3D US imaging, utilizing two lightweight cameras. To eliminate accumulated errors during US scans, we introduce PoseNet, which directly predicts the probe's 6 D pose relative to a preset world coordinate system based on camera observations. We first jointly train pose and camera image encoders based on pairs of 6 D pose and camera observations densely sampled in simulation. This will encourage each pair of probe pose and its corresponding camera observation to share the same representation in latent space. To ensure the two encoders handle unseen images and poses effectively, we incorporate a triplet loss that enforces smaller differences in latent features between nearby poses compared to distant ones. Then, the pose decoder uses the latent representation of the camera images to predict the probe's 6 D pose. To bridge the sim-to-real gap, in the real world, we use the trained image encoder and pose decoder for initial predictions, followed by an additional MLP layer to refine the estimated pose, improving accuracy. The results obtained from an arm phantom demonstrate the effectiveness of the proposed method, which notably surpasses state-of-the-art techniques, achieving average positional and rotational errors of 2.03 mm and 0.37°, respectively. Code:https://github.com/dianyeHuang/FreehandUS-Pose-Estimation.
Persistent Identifierhttp://hdl.handle.net/10722/365364
ISSN
2023 SCImago Journal Rankings: 1.620

 

DC FieldValueLanguage
dc.contributor.authorHuang, Dianye-
dc.contributor.authorNavab, Nassir-
dc.contributor.authorJiang, Zhongliang-
dc.date.accessioned2025-11-05T06:55:39Z-
dc.date.available2025-11-05T06:55:39Z-
dc.date.issued2025-
dc.identifier.citationProceedings IEEE International Conference on Robotics and Automation, 2025, p. 15472-15478-
dc.identifier.issn1050-4729-
dc.identifier.urihttp://hdl.handle.net/10722/365364-
dc.description.abstractUltrasound (US) probe localization relative to the examined subject is essential for freehand 3D US imaging, which offers significant clinical value due to its affordability and unrestricted field of view. However, existing methods often rely on expensive tracking systems or bulky probes, while recent US image-based deep learning methods suffer from accumulated errors during probe maneuvering. To address these challenges, this study proposes a versatile, cost-effective probe pose localization method for freehand 3D US imaging, utilizing two lightweight cameras. To eliminate accumulated errors during US scans, we introduce PoseNet, which directly predicts the probe's 6 D pose relative to a preset world coordinate system based on camera observations. We first jointly train pose and camera image encoders based on pairs of 6 D pose and camera observations densely sampled in simulation. This will encourage each pair of probe pose and its corresponding camera observation to share the same representation in latent space. To ensure the two encoders handle unseen images and poses effectively, we incorporate a triplet loss that enforces smaller differences in latent features between nearby poses compared to distant ones. Then, the pose decoder uses the latent representation of the camera images to predict the probe's 6 D pose. To bridge the sim-to-real gap, in the real world, we use the trained image encoder and pose decoder for initial predictions, followed by an additional MLP layer to refine the estimated pose, improving accuracy. The results obtained from an arm phantom demonstrate the effectiveness of the proposed method, which notably surpasses state-of-the-art techniques, achieving average positional and rotational errors of 2.03 mm and 0.37°, respectively. Code:https://github.com/dianyeHuang/FreehandUS-Pose-Estimation.-
dc.languageeng-
dc.relation.ispartofProceedings IEEE International Conference on Robotics and Automation-
dc.titleImproving Probe Localization for Freehand 3D Ultrasound Using Lightweight Cameras-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICRA55743.2025.11128230-
dc.identifier.scopuseid_2-s2.0-105016710375-
dc.identifier.spage15472-
dc.identifier.epage15478-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats