TY - GEN
T1 - Document-Level Relation Extraction with Entity Enhancement and Context Refinement
AU - Zou, Meng
AU - Yang, Qiang
AU - Qu, Jianfeng
AU - Li, Zhixu
AU - Liu, An
AU - Zhao, Lei
AU - Chen, Zhigang
N1 - KAUST Repository Item: Exported on 2022-01-11
Acknowledgements: We are grateful to Heng Ye, Jiaan Wang and all reviews for their constructive comments. This work was supported by the National Key R&D Program of China (No. 2018AAA0101900), the Priority Academic Program Development of Jiangsu Higher Education Institutions, National Natural Science Foundation of China (Grant No. 62072323, 61632016, 62102276), Natural Science Foundation of Jiangsu Province (No. BK20191420).
PY - 2022/1/1
Y1 - 2022/1/1
N2 - Document-level Relation Extraction (DocRE) is the task of extracting relational facts mentioned in the entire document. Despite its popularity, there are still two major difficulties with this task: (i) How to learn more informative embeddings for entity pairs? (ii) How to capture the crucial context describing the relation between an entity pair from the document? To tackle the first challenge, we propose to encode the document with a task-specific pre-trained encoder, where three tasks are involved in pre-training. While one novel task is designed to learn the relation semantic from diverse expressions by utilizing relation-aware pre-training data, the other two tasks, Masked Language Modeling (MLM) and Mention Reference Prediction (MRP), are adopted to enhance the encoder’s capacity in text understanding and coreference capturing. For addressing the second challenge, we craft a hierarchical attention mechanism to refine the context for entity pairs, which considers the embeddings from the encoder as well as the sequential distance information of mentions in the given document. Extensive experimental study on the benchmark dataset DocRED verifies that our method achieves better performance than the baselines.
AB - Document-level Relation Extraction (DocRE) is the task of extracting relational facts mentioned in the entire document. Despite its popularity, there are still two major difficulties with this task: (i) How to learn more informative embeddings for entity pairs? (ii) How to capture the crucial context describing the relation between an entity pair from the document? To tackle the first challenge, we propose to encode the document with a task-specific pre-trained encoder, where three tasks are involved in pre-training. While one novel task is designed to learn the relation semantic from diverse expressions by utilizing relation-aware pre-training data, the other two tasks, Masked Language Modeling (MLM) and Mention Reference Prediction (MRP), are adopted to enhance the encoder’s capacity in text understanding and coreference capturing. For addressing the second challenge, we craft a hierarchical attention mechanism to refine the context for entity pairs, which considers the embeddings from the encoder as well as the sequential distance information of mentions in the given document. Extensive experimental study on the benchmark dataset DocRED verifies that our method achieves better performance than the baselines.
UR - http://hdl.handle.net/10754/674902
UR - https://link.springer.com/10.1007/978-3-030-91560-5_25
UR - http://www.scopus.com/inward/record.url?scp=85121930534&partnerID=8YFLogxK
U2 - 10.1007/978-3-030-91560-5_25
DO - 10.1007/978-3-030-91560-5_25
M3 - Conference contribution
SN - 9783030915599
SP - 347
EP - 362
BT - Web Information Systems Engineering – WISE 2021
PB - Springer International Publishing
ER -