File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Weakly Supervised Complementary Parts Models for Fine-Grained Image Classification from the Bottom Up

TitleWeakly Supervised Complementary Parts Models for Fine-Grained Image Classification from the Bottom Up
Authors
Issue Date2019
PublisherInstitute of Electrical and Electronics Engineers, Inc..
Citation
IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15-20 June, 2019. In Proceedings: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition: CVPR 2019 How to Cite?
AbstractGiven a training dataset composed of images and corresponding category labels, deep convolutional neural networks show a strong ability in mining discriminative parts for image classification. However, deep convolutional neural networks trained with image level labels only tend to focus on the most discriminative parts while missing other object parts, which could provide complementary information. In this paper, we approach this problem from a different perspective. We build complementary parts models in a weakly supervised manner to retrieve information suppressed by dominant object parts detected by convolutional neural networks. Given image level labels only, we first extract rough object instances by performing weakly supervised object detection and instance segmentation using Mask R-CNN and CRF-based segmentation. Then we estimate and search for the best parts model for each object instance under the principle of preserving as much diversity as possible. In the last stage, we build a bi-directional long short-term memory (LSTM) network to fuze and encode the partial information of these complementary parts into a comprehensive feature for image classification. Experimental results indicate that the proposed method not only achieves significant improvement over our baseline models, but also outperforms state-of-the-art algorithms by a large margin (6.7%, 2.8%, 5.2% respectively) on Stanford Dogs 120, Caltech-UCSD Birds 2011-200 and Caltech 256.
Persistent Identifierhttp://hdl.handle.net/10722/316285
ISBN
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorGE, W-
dc.contributor.authorLIN, X-
dc.contributor.authorYu, Y-
dc.date.accessioned2022-09-02T06:08:47Z-
dc.date.available2022-09-02T06:08:47Z-
dc.date.issued2019-
dc.identifier.citationIEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15-20 June, 2019. In Proceedings: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition: CVPR 2019-
dc.identifier.isbn9781728132938-
dc.identifier.urihttp://hdl.handle.net/10722/316285-
dc.description.abstractGiven a training dataset composed of images and corresponding category labels, deep convolutional neural networks show a strong ability in mining discriminative parts for image classification. However, deep convolutional neural networks trained with image level labels only tend to focus on the most discriminative parts while missing other object parts, which could provide complementary information. In this paper, we approach this problem from a different perspective. We build complementary parts models in a weakly supervised manner to retrieve information suppressed by dominant object parts detected by convolutional neural networks. Given image level labels only, we first extract rough object instances by performing weakly supervised object detection and instance segmentation using Mask R-CNN and CRF-based segmentation. Then we estimate and search for the best parts model for each object instance under the principle of preserving as much diversity as possible. In the last stage, we build a bi-directional long short-term memory (LSTM) network to fuze and encode the partial information of these complementary parts into a comprehensive feature for image classification. Experimental results indicate that the proposed method not only achieves significant improvement over our baseline models, but also outperforms state-of-the-art algorithms by a large margin (6.7%, 2.8%, 5.2% respectively) on Stanford Dogs 120, Caltech-UCSD Birds 2011-200 and Caltech 256.-
dc.languageeng-
dc.publisherInstitute of Electrical and Electronics Engineers, Inc..-
dc.relation.ispartofProceedings: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition: CVPR 2019-
dc.titleWeakly Supervised Complementary Parts Models for Fine-Grained Image Classification from the Bottom Up-
dc.typeConference_Paper-
dc.identifier.emailYu, Y: yzyu@cs.hku.hk-
dc.identifier.authorityYu, Y=rp01415-
dc.identifier.doi10.1109/CVPR.2019.00315-
dc.identifier.hkuros336349-
dc.identifier.spage760-
dc.identifier.epage769-
dc.identifier.isiWOS:000529484003021-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats