Title | Towards Semantics-Enhanced Pre-Training: Can Lexicon Definitions Help Learning Sentence Meanings? |
Authors | Ren, Xuancheng Sun, Xu Wang, Houfeng Liu, Qun |
Affiliation | Peking Univ, Sch EECS, MOE Key Lab Computat Linguist, Beijing, Peoples R China Peking Univ, Ctr Data Sci, Beijing, Peoples R China Huawei Noahs Ark Lab, Beijing, Peoples R China |
Issue Date | 2021 |
Publisher | THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE |
Abstract | Self-supervised pre-training techniques, albeit relying on large amounts of text, have enabled rapid growth in learning language representations for natural language understanding. However, as radically empirical models on sentences, they are subject to the input data distribution, inevitably incorporating data bias and reporting bias, which may lead to inaccurate understanding of sentences. To address this problem, we propose to adopt a human learner's approach: when we cannot make sense of a word in a sentence, we often consult the dictionary for specific meanings; but can the same work for empirical models? In this work, we try to inform the pre-trained masked language models of word meanings for semantics-enhanced pre-training. To achieve a contrastive and holistic view of word meanings, a definition pair of two related words is presented to the masked language model such that the model can better associate a word with its crucial semantic features. Both intrinsic and extrinsic evaluations validate the proposed approach on semantics-orientated tasks, with an almost negligible increase of training data. |
URI | http://hdl.handle.net/20.500.11897/623200 |
ISBN | 978-1-57735-866-4 |
ISSN | 2159-5399 |
Indexed | CPCI-S(ISTP) |
Appears in Collections: | 信息科学技术学院 计算语言学教育部重点实验室 其他研究院 |