File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Article: Visual event recognition in videos by learning from web data

TitleVisual event recognition in videos by learning from web data
Authors
Keywordsadaptive MKL
aligned space-time pyramid matching
cross-domain learning
domain adaptation
Event recognition
transfer learning
Issue Date2012
Citation
IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, v. 34, n. 9, p. 1667-1680 How to Cite?
AbstractWe propose a visual event recognition framework for consumer videos by leveraging a large amount of loosely labeled web videos (e.g., from YouTube). Observing that consumer videos generally contain large intraclass variations within the same type of events, we first propose a new method, called Aligned Space-Time Pyramid Matching (ASTPM), to measure the distance between any two video clips. Second, we propose a new transfer learning method, referred to as Adaptive Multiple Kernel Learning (A-MKL), in order to 1) fuse the information from multiple pyramid levels and features (i.e., space-time features and static SIFT features) and 2) cope with the considerable variation in feature distributions between videos from two domains (i.e., web video domain and consumer video domain). For each pyramid level and each type of local features, we first train a set of SVM classifiers based on the combined training set from two domains by using multiple base kernels from different kernel types and parameters, which are then fused with equal weights to obtain a prelearned average classifier. In A-MKL, for each event class we learn an adapted target classifier based on multiple base kernels and the prelearned average classifiers from this event class or all the event classes by minimizing both the structural risk functional and the mismatch between data distributions of two domains. Extensive experiments demonstrate the effectiveness of our proposed framework that requires only a small number of labeled consumer videos by leveraging web data. We also conduct an in-depth investigation on various aspects of the proposed method A-MKL, such as the analysis on the combination coefficients on the prelearned classifiers, the convergence of the learning algorithm, and the performance variation by using different proportions of labeled consumer videos. Moreover, we show that A-MKL using the prelearned classifiers from all the event classes leads to better performance when compared with A-MKL using the prelearned classifiers only from each individual event class. © 2012 IEEE.
Persistent Identifierhttp://hdl.handle.net/10722/321482
ISSN
2021 Impact Factor: 24.314
2020 SCImago Journal Rankings: 3.811
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorDuan, Lixin-
dc.contributor.authorXu, Dong-
dc.contributor.authorTsang, Ivor Wai Hung-
dc.contributor.authorLuo, Jiebo-
dc.date.accessioned2022-11-03T02:19:12Z-
dc.date.available2022-11-03T02:19:12Z-
dc.date.issued2012-
dc.identifier.citationIEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, v. 34, n. 9, p. 1667-1680-
dc.identifier.issn0162-8828-
dc.identifier.urihttp://hdl.handle.net/10722/321482-
dc.description.abstractWe propose a visual event recognition framework for consumer videos by leveraging a large amount of loosely labeled web videos (e.g., from YouTube). Observing that consumer videos generally contain large intraclass variations within the same type of events, we first propose a new method, called Aligned Space-Time Pyramid Matching (ASTPM), to measure the distance between any two video clips. Second, we propose a new transfer learning method, referred to as Adaptive Multiple Kernel Learning (A-MKL), in order to 1) fuse the information from multiple pyramid levels and features (i.e., space-time features and static SIFT features) and 2) cope with the considerable variation in feature distributions between videos from two domains (i.e., web video domain and consumer video domain). For each pyramid level and each type of local features, we first train a set of SVM classifiers based on the combined training set from two domains by using multiple base kernels from different kernel types and parameters, which are then fused with equal weights to obtain a prelearned average classifier. In A-MKL, for each event class we learn an adapted target classifier based on multiple base kernels and the prelearned average classifiers from this event class or all the event classes by minimizing both the structural risk functional and the mismatch between data distributions of two domains. Extensive experiments demonstrate the effectiveness of our proposed framework that requires only a small number of labeled consumer videos by leveraging web data. We also conduct an in-depth investigation on various aspects of the proposed method A-MKL, such as the analysis on the combination coefficients on the prelearned classifiers, the convergence of the learning algorithm, and the performance variation by using different proportions of labeled consumer videos. Moreover, we show that A-MKL using the prelearned classifiers from all the event classes leads to better performance when compared with A-MKL using the prelearned classifiers only from each individual event class. © 2012 IEEE.-
dc.languageeng-
dc.relation.ispartofIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.subjectadaptive MKL-
dc.subjectaligned space-time pyramid matching-
dc.subjectcross-domain learning-
dc.subjectdomain adaptation-
dc.subjectEvent recognition-
dc.subjecttransfer learning-
dc.titleVisual event recognition in videos by learning from web data-
dc.typeArticle-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/TPAMI.2011.265-
dc.identifier.pmid22201057-
dc.identifier.scopuseid_2-s2.0-84865579385-
dc.identifier.volume34-
dc.identifier.issue9-
dc.identifier.spage1667-
dc.identifier.epage1680-
dc.identifier.isiWOS:000306409100002-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats