TY - GEN
T1 - Attention based document-level relation extraction with none class ranking loss
AU - Xu, Xiaolong
AU - Li, Chenbin
AU - Xiang, Haolong
AU - Qi, Lianyong
AU - Zhang, Xuyun
AU - Dou, Wanchun
PY - 2024
Y1 - 2024
N2 - Through document-level relation extraction (RE), the analysis of the global relation between entities in the text is feasible, and more comprehensive and accurate semantic information can be obtained. In document-level RE, the model needs to infer the implicit relations between two entities in different sentences. To obtain more semantic information, existing methods mainly focus on exploring entity representations. However, they ignore the correlations and indivisibility between relations, entities and contexts. Furthermore, current methods only independently estimate the cases of predefined relations, ignoring the case of “no relation”, which results in poor prediction. To address the above issues, we propose a document-level RE method based on attention mechanisms, which considers the case of “no relation”. Specifically, our approach leverages graph attention and multi-head attention networks to capture the correlations and indivisibility among relations, entities, and contexts, respectively. In addition, a novel multi-label loss function that promotes large margins in label confidence scores between each predefined class and the none class is employed to improve the prediction performance. Extensive experiments conducted on benchmarking datasets demonstrate that our proposed method outperforms the state-of-the-art baselines with higher accuracy.
AB - Through document-level relation extraction (RE), the analysis of the global relation between entities in the text is feasible, and more comprehensive and accurate semantic information can be obtained. In document-level RE, the model needs to infer the implicit relations between two entities in different sentences. To obtain more semantic information, existing methods mainly focus on exploring entity representations. However, they ignore the correlations and indivisibility between relations, entities and contexts. Furthermore, current methods only independently estimate the cases of predefined relations, ignoring the case of “no relation”, which results in poor prediction. To address the above issues, we propose a document-level RE method based on attention mechanisms, which considers the case of “no relation”. Specifically, our approach leverages graph attention and multi-head attention networks to capture the correlations and indivisibility among relations, entities, and contexts, respectively. In addition, a novel multi-label loss function that promotes large margins in label confidence scores between each predefined class and the none class is employed to improve the prediction performance. Extensive experiments conducted on benchmarking datasets demonstrate that our proposed method outperforms the state-of-the-art baselines with higher accuracy.
KW - Natural Language Processing: NLP: Information extraction
KW - Natural Language Processing: NLP: Embeddings
UR - http://www.scopus.com/inward/record.url?scp=85204304034&partnerID=8YFLogxK
U2 - 10.24963/ijcai.2024/726
DO - 10.24963/ijcai.2024/726
M3 - Conference proceeding contribution
AN - SCOPUS:85204304034
T3 - IJCAI International Joint Conference on Artificial Intelligence
SP - 6569
EP - 6577
BT - IJCAI 2024
A2 - Larson, Kate
PB - Association for Computing Machinery (ACM)
CY - New York, NY
T2 - International Joint Conference on Artificial Intelligence (33rd : 2024)
Y2 - 3 August 2024 through 9 August 2024
ER -