TY - GEN
T1 - EIDER
T2 - Findings of the Association for Computational Linguistics: ACL 2022
AU - Xie, Yiqing
AU - Shen, Jiaming
AU - Li, Sha
AU - Mao, Yuning
AU - Han, Jiawei
N1 - Publisher Copyright:
© 2022 Association for Computational Linguistics.
PY - 2022
Y1 - 2022
N2 - Document-level relation extraction (DocRE) aims to extract semantic relations among entity pairs in a document. Typical DocRE methods blindly take the full document as input, while a subset of the sentences in the document, noted as the evidence, are often sufficient for humans to predict the relation of an entity pair. In this paper, we propose an evidence-enhanced framework, EIDER, that empowers DocRE by efficiently extracting evidence and effectively fusing the extracted evidence in inference. We first jointly train an RE model with a lightweight evidence extraction model, which is efficient in both memory and runtime. Empirically, even training the evidence model on silver labels constructed by our heuristic rules can lead to better RE performance. We further design a simple yet effective inference process that makes RE predictions on both extracted evidence and the full document, then fuses the predictions through a blending layer. This allows EIDER to focus on important sentences while still having access to the complete information in the document. Extensive experiments show that EIDER outperforms state-ofthe-art methods on three benchmark datasets (e.g., by 1.37/1.26 Ign F1/F1 on DocRED).
AB - Document-level relation extraction (DocRE) aims to extract semantic relations among entity pairs in a document. Typical DocRE methods blindly take the full document as input, while a subset of the sentences in the document, noted as the evidence, are often sufficient for humans to predict the relation of an entity pair. In this paper, we propose an evidence-enhanced framework, EIDER, that empowers DocRE by efficiently extracting evidence and effectively fusing the extracted evidence in inference. We first jointly train an RE model with a lightweight evidence extraction model, which is efficient in both memory and runtime. Empirically, even training the evidence model on silver labels constructed by our heuristic rules can lead to better RE performance. We further design a simple yet effective inference process that makes RE predictions on both extracted evidence and the full document, then fuses the predictions through a blending layer. This allows EIDER to focus on important sentences while still having access to the complete information in the document. Extensive experiments show that EIDER outperforms state-ofthe-art methods on three benchmark datasets (e.g., by 1.37/1.26 Ign F1/F1 on DocRED).
UR - http://www.scopus.com/inward/record.url?scp=85140827749&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85140827749&partnerID=8YFLogxK
U2 - 10.18653/v1/2022.findings-acl.23
DO - 10.18653/v1/2022.findings-acl.23
M3 - Conference contribution
AN - SCOPUS:85140827749
T3 - Proceedings of the Annual Meeting of the Association for Computational Linguistics
SP - 257
EP - 268
BT - ACL 2022 - 60th Annual Meeting of the Association for Computational Linguistics, Findings of ACL 2022
A2 - Muresan, Smaranda
A2 - Nakov, Preslav
A2 - Villavicencio, Aline
PB - Association for Computational Linguistics (ACL)
Y2 - 22 May 2022 through 27 May 2022
ER -