File Download

There are no files associated with this item.

Supplementary

Conference Paper: Transductive Zero-Shot Learning with Visual Structure Constraint

TitleTransductive Zero-Shot Learning with Visual Structure Constraint
Authors
Issue Date2019
PublisherMorgan Kaufmann Publishers.
Citation
33rd Conference on Neural Information Processing Systems (NeurIPS), Vancouver, Canada, December 8-14, 2019. In Advances in Neural Information Processing Systems (NeurIPS), v. 32, p. 9972-9982 How to Cite?
AbstractTo recognize objects of the unseen classes, most existing Zero-Shot Learning (ZSL) methods first learn a compatible projection function between the common semantic space and the visual space based on the data of source seen classes, then directly apply it to the target unseen classes. However, in real scenarios, the data distribution between the source and target domain might not match well, thus causing the well-known domain shift problem. Based on the observation that visual features of test instances can be separated into different clusters, we propose a new visual structure constraint on class centers for transductive ZSL, to improve the generality of the projection function (ie alleviate the above domain shift problem). Specifically, three different strategies (symmetric Chamfer-distance,Bipartite matching distance, and Wasserstein distance) are adopted to align the projected unseen semantic centers and visual cluster centers of test instances. We also propose a new training strategy to handle the real cases where many unrelated images exist in the test dataset, which is not considered in previous methods. Experiments on many widely used datasets demonstrate that the proposed visual structure constraint can bring substantial performance gain consistently and achieve state-of-the-art results.
Persistent Identifierhttp://hdl.handle.net/10722/316287

 

DC FieldValueLanguage
dc.contributor.authorWan, Z-
dc.contributor.authorChen, D-
dc.contributor.authorLi, Y-
dc.contributor.authorYan, X-
dc.contributor.authorZhang, J-
dc.contributor.authorYu, Y-
dc.contributor.authorLiao, J-
dc.date.accessioned2022-09-02T06:08:49Z-
dc.date.available2022-09-02T06:08:49Z-
dc.date.issued2019-
dc.identifier.citation33rd Conference on Neural Information Processing Systems (NeurIPS), Vancouver, Canada, December 8-14, 2019. In Advances in Neural Information Processing Systems (NeurIPS), v. 32, p. 9972-9982-
dc.identifier.urihttp://hdl.handle.net/10722/316287-
dc.description.abstractTo recognize objects of the unseen classes, most existing Zero-Shot Learning (ZSL) methods first learn a compatible projection function between the common semantic space and the visual space based on the data of source seen classes, then directly apply it to the target unseen classes. However, in real scenarios, the data distribution between the source and target domain might not match well, thus causing the well-known domain shift problem. Based on the observation that visual features of test instances can be separated into different clusters, we propose a new visual structure constraint on class centers for transductive ZSL, to improve the generality of the projection function (ie alleviate the above domain shift problem). Specifically, three different strategies (symmetric Chamfer-distance,Bipartite matching distance, and Wasserstein distance) are adopted to align the projected unseen semantic centers and visual cluster centers of test instances. We also propose a new training strategy to handle the real cases where many unrelated images exist in the test dataset, which is not considered in previous methods. Experiments on many widely used datasets demonstrate that the proposed visual structure constraint can bring substantial performance gain consistently and achieve state-of-the-art results.-
dc.languageeng-
dc.publisherMorgan Kaufmann Publishers.-
dc.relation.ispartofAdvances in Neural Information Processing Systems (NeurIPS)-
dc.titleTransductive Zero-Shot Learning with Visual Structure Constraint-
dc.typeConference_Paper-
dc.identifier.emailYu, Y: yzyu@cs.hku.hk-
dc.identifier.authorityYu, Y=rp01415-
dc.identifier.hkuros336352-
dc.identifier.volume32-
dc.identifier.spage9972-
dc.identifier.epage9982-
dc.publisher.placeUnited States-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats