File Download
Supplementary

postgraduate thesis: Interpreting and analyzing neural networks for NLP : a knowledge management perspective

TitleInterpreting and analyzing neural networks for NLP : a knowledge management perspective
Authors
Issue Date2021
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Wu, Z. [吳志勇]. (2021). Interpreting and analyzing neural networks for NLP : a knowledge management perspective. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractLanguage technology has become pervasive in everyday life, with applications like Google Search, Amazon Alexa, Apple Siri, etc. The success of such Natural Language Processing (NLP) systems is built on the advance in deep neural networks, thus it also comes at the expense of model becoming less interpretable, i.e., such models are often perceived as ``black box''. The lack of model transparency not only hinders future research effort in advancing the field, but also limit these models' utility in applications that have certain requirement in terms of reliability, ethic, and legitimacy. Opening up the black box of neural NLP models has attracted attention from different NLP sub-field. This thesis offers a novel perspective and investigates methods to analyze how NLP models cope with real-world knowledge, since the ability to connect textual symbols with human knowledge and exploit these knowledge in decision making is the key to NLP models' success. Toward this goal, we first develop perturbed masking -- a parameter-free method for analyzing and interpreting a recent prevalent pre-trained language model -- BERT. The analysis quantifies the linguistic knowledge BERT captured and sheds light on its remarkable success in many NLP tasks. Second, we present a directly interpretable Multimodal Machine Translation (MMT) model, which consists of an interpretable component that quantifies its dependency on visual knowledge. Our findings stress the importance of interpretability in MMT and suggest potential directions for improvement. Third, we explore how interpretability can lead to a better model design and demonstrate with an application on knowledge-based question answering.
DegreeDoctor of Philosophy
SubjectNeural networks (Computer science)
Natural language processing (Computer science)
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/310288

 

DC FieldValueLanguage
dc.contributor.authorWu, Zhiyong-
dc.contributor.author吳志勇-
dc.date.accessioned2022-01-29T16:16:04Z-
dc.date.available2022-01-29T16:16:04Z-
dc.date.issued2021-
dc.identifier.citationWu, Z. [吳志勇]. (2021). Interpreting and analyzing neural networks for NLP : a knowledge management perspective. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/310288-
dc.description.abstractLanguage technology has become pervasive in everyday life, with applications like Google Search, Amazon Alexa, Apple Siri, etc. The success of such Natural Language Processing (NLP) systems is built on the advance in deep neural networks, thus it also comes at the expense of model becoming less interpretable, i.e., such models are often perceived as ``black box''. The lack of model transparency not only hinders future research effort in advancing the field, but also limit these models' utility in applications that have certain requirement in terms of reliability, ethic, and legitimacy. Opening up the black box of neural NLP models has attracted attention from different NLP sub-field. This thesis offers a novel perspective and investigates methods to analyze how NLP models cope with real-world knowledge, since the ability to connect textual symbols with human knowledge and exploit these knowledge in decision making is the key to NLP models' success. Toward this goal, we first develop perturbed masking -- a parameter-free method for analyzing and interpreting a recent prevalent pre-trained language model -- BERT. The analysis quantifies the linguistic knowledge BERT captured and sheds light on its remarkable success in many NLP tasks. Second, we present a directly interpretable Multimodal Machine Translation (MMT) model, which consists of an interpretable component that quantifies its dependency on visual knowledge. Our findings stress the importance of interpretability in MMT and suggest potential directions for improvement. Third, we explore how interpretability can lead to a better model design and demonstrate with an application on knowledge-based question answering.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshNeural networks (Computer science)-
dc.subject.lcshNatural language processing (Computer science)-
dc.titleInterpreting and analyzing neural networks for NLP : a knowledge management perspective-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2022-
dc.identifier.mmsid991044467224703414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats