File Download
 
Links for fulltext
(May Require Subscription)
 
Supplementary

Conference Paper: Self-calibrating depth from refraction
  • Basic View
  • Metadata View
  • XML View
TitleSelf-calibrating depth from refraction
 
AuthorsChen, Z1
Wong, KKY1
Matsushita, Y2
Zhu, X1
Liu, M1
 
KeywordsDepth map
Input image
Real world data
Self-calibrating
Cameras
 
Issue Date2011
 
PublisherIEEE. The Journal's web site is located at http://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000149
 
CitationThe 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6-13 November 2011. In Proceedings of IEEE ICCV, 2011, p. 635-642 [How to Cite?]
DOI: http://dx.doi.org/10.1109/ICCV.2011.6126298
 
AbstractIn this paper, we introduce a novel method for depth acquisition based on refraction of light. A scene is captured twice by a fixed perspective camera, with the first image captured directly by the camera and the second by placing a transparent medium between the scene and the camera. A depth map of the scene is then recovered from the displacements of scene points in the images. Unlike other existing depth from refraction methods, our method does not require the knowledge of the pose and refractive index of the transparent medium, but can recover them directly from the input images. We hence call our method self-calibrating depth from refraction. Experimental results on both synthetic and real-world data are presented, which demonstrate the effectiveness of the proposed method. © 2011 IEEE.
 
ISBN978-1-4577-1102-2
 
ISSN1550-5499
 
DOIhttp://dx.doi.org/10.1109/ICCV.2011.6126298
 
DC FieldValue
dc.contributor.authorChen, Z
 
dc.contributor.authorWong, KKY
 
dc.contributor.authorMatsushita, Y
 
dc.contributor.authorZhu, X
 
dc.contributor.authorLiu, M
 
dc.date.accessioned2011-08-26T14:30:34Z
 
dc.date.available2011-08-26T14:30:34Z
 
dc.date.issued2011
 
dc.description.abstractIn this paper, we introduce a novel method for depth acquisition based on refraction of light. A scene is captured twice by a fixed perspective camera, with the first image captured directly by the camera and the second by placing a transparent medium between the scene and the camera. A depth map of the scene is then recovered from the displacements of scene points in the images. Unlike other existing depth from refraction methods, our method does not require the knowledge of the pose and refractive index of the transparent medium, but can recover them directly from the input images. We hence call our method self-calibrating depth from refraction. Experimental results on both synthetic and real-world data are presented, which demonstrate the effectiveness of the proposed method. © 2011 IEEE.
 
dc.description.naturepublished_or_final_version
 
dc.description.otherThe 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6-13 November 2011. In Proceedings of IEEE ICCV, 2011, p. 635-642
 
dc.identifier.citationThe 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6-13 November 2011. In Proceedings of IEEE ICCV, 2011, p. 635-642 [How to Cite?]
DOI: http://dx.doi.org/10.1109/ICCV.2011.6126298
 
dc.identifier.doihttp://dx.doi.org/10.1109/ICCV.2011.6126298
 
dc.identifier.epage642
 
dc.identifier.hkuros191685
 
dc.identifier.isbn978-1-4577-1102-2
 
dc.identifier.issn1550-5499
 
dc.identifier.scopuseid_2-s2.0-84863036717
 
dc.identifier.spage635
 
dc.identifier.urihttp://hdl.handle.net/10722/137655
 
dc.languageeng
 
dc.publisherIEEE. The Journal's web site is located at http://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000149
 
dc.publisher.placeUnited States
 
dc.relation.ispartofIEEE International Conference on Computer Vision Proceedings
 
dc.rightsIEEE International Conference on Computer Vision Proceedings. Copyright © IEEE.
 
dc.rights©2011 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.
 
dc.rightsCreative Commons: Attribution 3.0 Hong Kong License
 
dc.subjectDepth map
 
dc.subjectInput image
 
dc.subjectReal world data
 
dc.subjectSelf-calibrating
 
dc.subjectCameras
 
dc.titleSelf-calibrating depth from refraction
 
dc.typeConference_Paper
 
<?xml encoding="utf-8" version="1.0"?>
<item><contributor.author>Chen, Z</contributor.author>
<contributor.author>Wong, KKY</contributor.author>
<contributor.author>Matsushita, Y</contributor.author>
<contributor.author>Zhu, X</contributor.author>
<contributor.author>Liu, M</contributor.author>
<date.accessioned>2011-08-26T14:30:34Z</date.accessioned>
<date.available>2011-08-26T14:30:34Z</date.available>
<date.issued>2011</date.issued>
<identifier.citation>The 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6-13 November 2011. In Proceedings of IEEE ICCV, 2011, p. 635-642</identifier.citation>
<identifier.isbn>978-1-4577-1102-2</identifier.isbn>
<identifier.issn>1550-5499</identifier.issn>
<identifier.uri>http://hdl.handle.net/10722/137655</identifier.uri>
<description.abstract>In this paper, we introduce a novel method for depth acquisition based on refraction of light. A scene is captured twice by a fixed perspective camera, with the first image captured directly by the camera and the second by placing a transparent medium between the scene and the camera. A depth map of the scene is then recovered from the displacements of scene points in the images. Unlike other existing depth from refraction methods, our method does not require the knowledge of the pose and refractive index of the transparent medium, but can recover them directly from the input images. We hence call our method self-calibrating depth from refraction. Experimental results on both synthetic and real-world data are presented, which demonstrate the effectiveness of the proposed method. &#169; 2011 IEEE.</description.abstract>
<language>eng</language>
<publisher>IEEE. The Journal&apos;s web site is located at http://ieeexplore.ieee.org/xpl/conhome.jsp?punumber=1000149</publisher>
<relation.ispartof>IEEE International Conference on Computer Vision Proceedings</relation.ispartof>
<rights>IEEE International Conference on Computer Vision Proceedings. Copyright &#169; IEEE.</rights>
<rights>&#169;2011 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.</rights>
<rights>Creative Commons: Attribution 3.0 Hong Kong License</rights>
<subject>Depth map</subject>
<subject>Input image</subject>
<subject>Real world data</subject>
<subject>Self-calibrating</subject>
<subject>Cameras</subject>
<title>Self-calibrating depth from refraction</title>
<type>Conference_Paper</type>
<description.nature>published_or_final_version</description.nature>
<identifier.doi>10.1109/ICCV.2011.6126298</identifier.doi>
<identifier.scopus>eid_2-s2.0-84863036717</identifier.scopus>
<identifier.hkuros>191685</identifier.hkuros>
<identifier.spage>635</identifier.spage>
<identifier.epage>642</identifier.epage>
<publisher.place>United States</publisher.place>
<description.other>The 2011 IEEE International Conference on Computer Vision (ICCV), Barcelona, Spain, 6-13 November 2011. In Proceedings of IEEE ICCV, 2011, p. 635-642</description.other>
<bitstream.url>http://hub.hku.hk/bitstream/10722/137655/1/Content.pdf</bitstream.url>
</item>
Author Affiliations
  1. The University of Hong Kong
  2. Microsoft Research Asia