File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis

TitleLearning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis
Authors
Issue Date2020
Citation
AAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, p. 8992-8999 How to Cite?
AbstractMultimodal language analysis often considers relationships between features based on text and those based on acoustical and visual properties. Text features typically outperform non-text features in sentiment analysis or emotion recognition tasks in part because the text features are derived from advanced language models or word embeddings trained on massive data sources while audio and video features are human-engineered and comparatively underdeveloped. Given that the text, audio, and video are describing the same utterance in different ways, we hypothesize that the multimodal sentiment analysis and emotion recognition can be improved by learning (hidden) correlations between features extracted from the outer product of text and audio (we call this text-based audio) and analogous text-based video. This paper proposes a novel model, the Interaction Canonical Correlation Network (ICCN), to learn such multimodal embeddings. ICCN learns correlations between all three modes via deep canonical correlation analysis (DCCA) and the proposed embeddings are then tested on several benchmark datasets and against other state-of-the-art multimodal embedding algorithms. Empirical results and ablation studies confirm the effectiveness of ICCN in capturing useful information from all three views.
Persistent Identifierhttp://hdl.handle.net/10722/341311

 

DC FieldValueLanguage
dc.contributor.authorSun, Zhongkai-
dc.contributor.authorSarma, Prathusha K.-
dc.contributor.authorSethares, William A.-
dc.contributor.authorLiang, Yingyu-
dc.date.accessioned2024-03-13T08:41:49Z-
dc.date.available2024-03-13T08:41:49Z-
dc.date.issued2020-
dc.identifier.citationAAAI 2020 - 34th AAAI Conference on Artificial Intelligence, 2020, p. 8992-8999-
dc.identifier.urihttp://hdl.handle.net/10722/341311-
dc.description.abstractMultimodal language analysis often considers relationships between features based on text and those based on acoustical and visual properties. Text features typically outperform non-text features in sentiment analysis or emotion recognition tasks in part because the text features are derived from advanced language models or word embeddings trained on massive data sources while audio and video features are human-engineered and comparatively underdeveloped. Given that the text, audio, and video are describing the same utterance in different ways, we hypothesize that the multimodal sentiment analysis and emotion recognition can be improved by learning (hidden) correlations between features extracted from the outer product of text and audio (we call this text-based audio) and analogous text-based video. This paper proposes a novel model, the Interaction Canonical Correlation Network (ICCN), to learn such multimodal embeddings. ICCN learns correlations between all three modes via deep canonical correlation analysis (DCCA) and the proposed embeddings are then tested on several benchmark datasets and against other state-of-the-art multimodal embedding algorithms. Empirical results and ablation studies confirm the effectiveness of ICCN in capturing useful information from all three views.-
dc.languageeng-
dc.relation.ispartofAAAI 2020 - 34th AAAI Conference on Artificial Intelligence-
dc.titleLearning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.scopuseid_2-s2.0-85106643433-
dc.identifier.spage8992-
dc.identifier.epage8999-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats