File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)

Article: Proximity based automatic data annotation for autonomous driving

TitleProximity based automatic data annotation for autonomous driving
Authors
Issue Date27-Feb-2020
PublisherIEEE
Citation
IEEE/CAA Journal of Automatica Sinica, 2020, v. 7, n. 2, p. 395-404 How to Cite?
Abstract

The recent development in autonomous driving involves high-level computer vision and detailed road scene understanding. Today, most autonomous vehicles employ expensive high quality sensor-set such as light detection and ranging ( LIDAR ) and HD maps with high level annotations. In this paper, we propose a scalable and affordable data collection and annotation framework, image-to-map annotation proximity ( I2MAP ), for affordance learning in autonomous driving applications. We provide a new driving dataset using our proposed framework for driving scene affordance learning by calibrating the data samples with available tags from online database such as open street map ( OSM ) . Our benchmark consists of 40 000 images with more than 40 affordance labels under various day time and weather even with very challenging heavy snow. We implemented sample advanced driver-assistance systems ( ADAS ) functions by training our data with neural networks ( NN ) and cross-validate the results on benchmarks like KITTI and BDD100K, which indicate the effectiveness of our framework and training models.


Persistent Identifierhttp://hdl.handle.net/10722/353755
ISSN
2023 Impact Factor: 15.3
2023 SCImago Journal Rankings: 4.696

 

DC FieldValueLanguage
dc.contributor.authorSun, Chen-
dc.contributor.authorVianney, Jean M Uwabeza-
dc.contributor.authorLi, Ying-
dc.contributor.authorChen, Long-
dc.contributor.authorLi, Li-
dc.contributor.authorWang, Fei-Yue-
dc.date.accessioned2025-01-24T00:35:32Z-
dc.date.available2025-01-24T00:35:32Z-
dc.date.issued2020-02-27-
dc.identifier.citationIEEE/CAA Journal of Automatica Sinica, 2020, v. 7, n. 2, p. 395-404-
dc.identifier.issn2329-9266-
dc.identifier.urihttp://hdl.handle.net/10722/353755-
dc.description.abstract<p>The recent development in autonomous driving involves high-level computer vision and detailed road scene understanding. Today, most autonomous vehicles employ expensive high quality sensor-set such as light detection and ranging ( LIDAR ) and HD maps with high level annotations. In this paper, we propose a scalable and affordable data collection and annotation framework, image-to-map annotation proximity ( I2MAP ), for affordance learning in autonomous driving applications. We provide a new driving dataset using our proposed framework for driving scene affordance learning by calibrating the data samples with available tags from online database such as open street map ( OSM ) . Our benchmark consists of 40 000 images with more than 40 affordance labels under various day time and weather even with very challenging heavy snow. We implemented sample advanced driver-assistance systems ( ADAS ) functions by training our data with neural networks ( NN ) and cross-validate the results on benchmarks like KITTI and BDD100K, which indicate the effectiveness of our framework and training models.<br></p>-
dc.languageeng-
dc.publisherIEEE-
dc.relation.ispartofIEEE/CAA Journal of Automatica Sinica-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.titleProximity based automatic data annotation for autonomous driving-
dc.typeArticle-
dc.identifier.doi10.1109/JAS.2020.1003033-
dc.identifier.volume7-
dc.identifier.issue2-
dc.identifier.spage395-
dc.identifier.epage404-
dc.identifier.eissn2329-9274-
dc.identifier.issnl2329-9266-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats