File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: BioNumQA-BERT: Answering Biomedical Questions Using Numerical Facts with a Deep Language Representation Model

TitleBioNumQA-BERT: Answering Biomedical Questions Using Numerical Facts with a Deep Language Representation Model
Authors
KeywordsText Mining
Biomedical Question Answering
BERT
Numerical encoding
Issue Date2021
PublisherAssociation for Computing Machinery (ACM).
Citation
The 12th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM-BCB 2021). Virtual Conference, 1-4 August 2021 How to Cite?
AbstractBiomedical question answering (QA) is playing an increasingly significant role in medical knowledge translation. However, current biomedical QA datasets and methods have limited capacity, as they commonly neglect the role of numerical facts in biomedical QA. In this paper, we constructed BioNumQA, a novel biomedical QA dataset that answers research questions using relevant numerical facts for biomedical QA model training and testing. To leverage the new dataset, we designed a new method called BioNumQA-BERT by introducing a novel numerical encoding scheme into the popular biomedical language model BioBERT to represent the numerical values in the input text. Our experiments show that BioNumQABERT significantly outperformed other state-of-art models, including DrQA, BERT and BioBERT (39.0% vs 29.5%, 31.3% and 33.2%, respectively, in strict accuracy). To improve the generalization ability of BioNumQA-BERT, we further pretrained it on a large biomedical text corpus and achieved 41.5% strict accuracy. BioNumQA and BioNumQA-BERT establish a new baseline for biomedical QA. The dataset, source codes and pretrained model of BioNumQA-BERT are available at https://github.com/LeaveYeah/BioNumQA-BERT.
DescriptionBCB Session 6B: Ontologies & Databases
Persistent Identifierhttp://hdl.handle.net/10722/301148
ISBN
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorWu, Y-
dc.contributor.authorTing, HF-
dc.contributor.authorLam, TW-
dc.contributor.authorLuo, R-
dc.date.accessioned2021-07-27T08:06:50Z-
dc.date.available2021-07-27T08:06:50Z-
dc.date.issued2021-
dc.identifier.citationThe 12th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM-BCB 2021). Virtual Conference, 1-4 August 2021-
dc.identifier.isbn9781450384506-
dc.identifier.urihttp://hdl.handle.net/10722/301148-
dc.descriptionBCB Session 6B: Ontologies & Databases-
dc.description.abstractBiomedical question answering (QA) is playing an increasingly significant role in medical knowledge translation. However, current biomedical QA datasets and methods have limited capacity, as they commonly neglect the role of numerical facts in biomedical QA. In this paper, we constructed BioNumQA, a novel biomedical QA dataset that answers research questions using relevant numerical facts for biomedical QA model training and testing. To leverage the new dataset, we designed a new method called BioNumQA-BERT by introducing a novel numerical encoding scheme into the popular biomedical language model BioBERT to represent the numerical values in the input text. Our experiments show that BioNumQABERT significantly outperformed other state-of-art models, including DrQA, BERT and BioBERT (39.0% vs 29.5%, 31.3% and 33.2%, respectively, in strict accuracy). To improve the generalization ability of BioNumQA-BERT, we further pretrained it on a large biomedical text corpus and achieved 41.5% strict accuracy. BioNumQA and BioNumQA-BERT establish a new baseline for biomedical QA. The dataset, source codes and pretrained model of BioNumQA-BERT are available at https://github.com/LeaveYeah/BioNumQA-BERT.-
dc.languageeng-
dc.publisherAssociation for Computing Machinery (ACM).-
dc.relation.ispartofThe 12th ACM Conference on Bioinformatics, Computational Biology, and Health Informatics (ACM BCB 2021)-
dc.subjectText Mining-
dc.subjectBiomedical Question Answering-
dc.subjectBERT-
dc.subjectNumerical encoding-
dc.titleBioNumQA-BERT: Answering Biomedical Questions Using Numerical Facts with a Deep Language Representation Model-
dc.typeConference_Paper-
dc.identifier.emailTing, HF: hfting@cs.hku.hk-
dc.identifier.emailLam, TW: twlam@cs.hku.hk-
dc.identifier.emailLuo, R: rbluo@cs.hku.hk-
dc.identifier.authorityTing, HF=rp00177-
dc.identifier.authorityLam, TW=rp00135-
dc.identifier.authorityLuo, R=rp02360-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1145/3459930.3469557-
dc.identifier.scopuseid_2-s2.0-85112390963-
dc.identifier.hkuros323502-
dc.identifier.isiWOS:000722623700070-
dc.publisher.placeNew York-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats