TY - JOUR
T1 - elBERto
T2 - Self-supervised commonsense learning for question answering
AU - Zhan, Xunlin
AU - Li, Yuan
AU - Dong, Xiao
AU - Liang, Xiaodan
AU - Hu, Zhiting
AU - Carin, Lawrence
N1 - Funding Information:
This work is supported by the National Natural Science Foundation of China (Grant No. 61976233 and U19A2073 ), Guangdong Provincial Natural Science Foundation of China (Grant No. 2019B1515120039 ), and the Office of Naval Research, United States under grant N00014-18-1-2871 .
Publisher Copyright:
© 2022 Elsevier B.V.
PY - 2022/12/22
Y1 - 2022/12/22
N2 - Commonsense question answering requires reasoning about everyday situations and causes and effects implicit in context. Typically, existing approaches first retrieve external evidence and then perform commonsense reasoning using these evidence. In this paper, we propose a Self-supervised Bidirectional Encoder Representation Learning of Commonsense (elBERto) framework, which is compatible with off-the-shelf QA model architectures. The framework comprises five self-supervised tasks to force the model to fully exploit the additional training signals from contexts containing rich commonsense. The tasks include a novel Contrastive Relation Learning task to encourage the model to distinguish between logically contrastive contexts, a new Jigsaw Puzzle task that requires the model to infer logical chains in long contexts, and three classic self-supervised learning(SSL) tasks to maintain pre-trained models’ language encoding ability. On the representative WIQA, CosmosQA, and ReClor datasets, elBERto outperforms all other methods using the same backbones and the same training set, including those utilizing explicit graph reasoning and external knowledge retrieval. Moreover, elBERto achieves substantial improvements on out-of-paragraph and no-effect questions where simple lexical similarity comparison does not help, indicating that it successfully learns commonsense and is able to leverage it when given dynamic context.
AB - Commonsense question answering requires reasoning about everyday situations and causes and effects implicit in context. Typically, existing approaches first retrieve external evidence and then perform commonsense reasoning using these evidence. In this paper, we propose a Self-supervised Bidirectional Encoder Representation Learning of Commonsense (elBERto) framework, which is compatible with off-the-shelf QA model architectures. The framework comprises five self-supervised tasks to force the model to fully exploit the additional training signals from contexts containing rich commonsense. The tasks include a novel Contrastive Relation Learning task to encourage the model to distinguish between logically contrastive contexts, a new Jigsaw Puzzle task that requires the model to infer logical chains in long contexts, and three classic self-supervised learning(SSL) tasks to maintain pre-trained models’ language encoding ability. On the representative WIQA, CosmosQA, and ReClor datasets, elBERto outperforms all other methods using the same backbones and the same training set, including those utilizing explicit graph reasoning and external knowledge retrieval. Moreover, elBERto achieves substantial improvements on out-of-paragraph and no-effect questions where simple lexical similarity comparison does not help, indicating that it successfully learns commonsense and is able to leverage it when given dynamic context.
KW - Commonsense reasoning
KW - Question answering
KW - Self-supervised learning
UR - http://www.scopus.com/inward/record.url?scp=85140333428&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2022.109964
DO - 10.1016/j.knosys.2022.109964
M3 - Article
AN - SCOPUS:85140333428
SN - 0950-7051
VL - 258
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 109964
ER -