File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Unsupervised Light Field Depth Estimation via Multi-view Feature Matching with Occlusion Prediction

TitleUnsupervised Light Field Depth Estimation via Multi-view Feature Matching with Occlusion Prediction
Authors
KeywordsConvolutional neural networks
Costs
Estimation
Feature extraction
feature matching
Image edge detection
Light field
occlusion prediction
Training
Training data
unsupervised depth estimation
Issue Date2023
Citation
IEEE Transactions on Circuits and Systems for Video Technology, 2023 How to Cite?
AbstractDepth estimation from light field (LF) images is a fundamental step for numerous applications. Recently, learning-based methods have achieved higher accuracy and efficiency than the traditional methods. However, it is costly to obtain sufficient depth labels for supervised training. In this paper, we propose an unsupervised framework to estimate depth from LF images. First, we design a disparity estimation network (DispNet) with a coarse-to-fine structure to predict disparity maps from different view combinations. It explicitly performs multi-view feature matching to learn the correspondences effectively. As occlusions may cause the violation of photo-consistency, we introduce an occlusion prediction network (OccNet) to predict the occlusion maps, which are used as the element-wise weights of photometric loss to solve the occlusion issue and assist the disparity learning. With the disparity maps estimated by multiple input combinations, we then propose a disparity fusion strategy based on the estimated errors with effective occlusion handling to obtain the final disparity map with higher accuracy. Experimental results demonstrate that our method achieves superior performance on both the dense and sparse LF images, and also shows better robustness and generalization on the real-world LF images compared to the other methods.
Persistent Identifierhttp://hdl.handle.net/10722/330491
ISSN
2023 Impact Factor: 8.3
2023 SCImago Journal Rankings: 2.299

 

DC FieldValueLanguage
dc.contributor.authorZhang, Shansi-
dc.contributor.authorMeng, Nan-
dc.contributor.authorLam, Edmund Y.-
dc.date.accessioned2023-09-05T12:11:10Z-
dc.date.available2023-09-05T12:11:10Z-
dc.date.issued2023-
dc.identifier.citationIEEE Transactions on Circuits and Systems for Video Technology, 2023-
dc.identifier.issn1051-8215-
dc.identifier.urihttp://hdl.handle.net/10722/330491-
dc.description.abstractDepth estimation from light field (LF) images is a fundamental step for numerous applications. Recently, learning-based methods have achieved higher accuracy and efficiency than the traditional methods. However, it is costly to obtain sufficient depth labels for supervised training. In this paper, we propose an unsupervised framework to estimate depth from LF images. First, we design a disparity estimation network (DispNet) with a coarse-to-fine structure to predict disparity maps from different view combinations. It explicitly performs multi-view feature matching to learn the correspondences effectively. As occlusions may cause the violation of photo-consistency, we introduce an occlusion prediction network (OccNet) to predict the occlusion maps, which are used as the element-wise weights of photometric loss to solve the occlusion issue and assist the disparity learning. With the disparity maps estimated by multiple input combinations, we then propose a disparity fusion strategy based on the estimated errors with effective occlusion handling to obtain the final disparity map with higher accuracy. Experimental results demonstrate that our method achieves superior performance on both the dense and sparse LF images, and also shows better robustness and generalization on the real-world LF images compared to the other methods.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Circuits and Systems for Video Technology-
dc.subjectConvolutional neural networks-
dc.subjectCosts-
dc.subjectEstimation-
dc.subjectFeature extraction-
dc.subjectfeature matching-
dc.subjectImage edge detection-
dc.subjectLight field-
dc.subjectocclusion prediction-
dc.subjectTraining-
dc.subjectTraining data-
dc.subjectunsupervised depth estimation-
dc.titleUnsupervised Light Field Depth Estimation via Multi-view Feature Matching with Occlusion Prediction-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TCSVT.2023.3305978-
dc.identifier.scopuseid_2-s2.0-85168738597-
dc.identifier.eissn1558-2205-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats