Deep neural network-based relation extraction: an overview
2022, Neural Computing and Applications
https://doi.org/10.1007/S00521-021-06667-3Abstract
Knowledge is a formal way of understanding the world, providing a human-level cognition and intelligence for the next-generation artificial intelligence (AI). One of the representations of knowledge is semantic relations between entities. An effective way to automatically acquire this important knowledge, called Relation Extraction (RE), a sub-task of information extraction, plays a vital role in Natural Language Processing (NLP). Its purpose is to identify semantic relations between entities from natural language text. To date, there are several studies for RE in previous works, which have documented these techniques based on Deep Neural Networks (DNNs) become a prevailing technique in this research. Especially, the supervised and distant supervision methods based on DNNs are the most popular and reliable solutions for RE. This article 1) introduces some general concepts, and further 2) gives a comprehensive overview of DNNs in RE from two points of view: supervised RE, which attempts to improve the standard RE systems, and distant supervision RE, which adopts DNNs to design sentence encoder and de-noise method. We further 3) cover some novel methods and recent trends as well as discuss possible future research directions for this task.
FAQs
AI
What explains the importance of positional relations in DNN-based RE?
The paper demonstrates that integrating position embeddings significantly improves sentence representation, enhancing information capture around entities. For instance, the position feature method achieves state-of-the-art performance in SemEval-2010-task8.
How have DNN-based methods evolved to address distant supervision challenges?
Recent advancements in DNN-based RE leverage hybrid models like PCNN with multi-instance learning to mitigate noisy labels in distant supervision. These approaches have reported improvements in classification accuracy by refining sentence encoding and reducing data noise.
When were distant supervision methods first proposed in the context of RE?
Distant supervision techniques were introduced in 2009 by Mintz et al. at the ACL conference, aiming to generate large-scale training datasets without manual annotation.
What are the main categories of DNN-based relation extraction methods?
DNN-based relation extraction is categorized into supervised methods, distant supervision, and semi-supervised methods, each tailored to specific data availability and accuracy requirements in NLP tasks.
How do different neural network architectures affect relation extraction performance?
The study indicates that varying architectures, such as CNNs or RNNs, provide distinct advantages; for instance, RNNs excel with long-distance dependencies, while CNNs are favored for capturing local structural features.
References (47)
- Kumar S (2017) A survey of deep learning methods for relation extraction. arXiv preprint arXiv:170503645
- Pawar S, Palshikar GK, Bhattacharyya P (2017) Relation extraction: A survey. arXiv preprint arXiv:171205191
- Hendrickx I, Kim SN, Kozareva Z, Nakov P, Ó Séaghdha D, Padó S, Pen- nacchiotti M, Romano L, Szpakowicz S (2009) Semeval-2010 task 8: Multi- ods in natural language processing (EMNLP), pp 1532-1543
- Zeng D, Liu K, Lai S, Zhou G, Zhao J, et al. (2014) Relation classification via convolutional deep neural network
- Nguyen TH, Grishman R (2015) Relation extraction: Perspective from convolutional neural networks. In: Proceedings of the 1st Workshop on Vector Space Modeling for Natural Language Processing, pp 39-48
- Santos CNd, Xiang B, Zhou B (2015) Classifying relations by ranking with convolutional neural networks. arXiv preprint arXiv:150406580
- Wang L, Cao Z, De Melo G, Liu Z (2016) Relation classification via multi- level attention cnns. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 1298-1307
- Zhang D, Wang D (2015) Relation classification via recurrent neural net- work. arXiv preprint arXiv:150801006
- Qin P, Xu W, Guo J (2017) Designing an adaptive attention mechanism for relation classification. In: 2017 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 4356-4362
- Zhang C, Cui C, Gao S, Nie X, Xu W, Yang L, Xi X, Yin Y (2019) Multi-gram cnn-based self-attention model for relation classification. IEEE Access 7:5343-5357
- Ren F, Zhou D, Liu Z, Li Y, Zhao R, Liu Y, Liang X (2018) Neural relation classification with text descriptions. In: Proceedings of the 27th International Conference on Computational Linguistics, pp 1167-1177
- Zhang L, Xiang F (2018) Relation classification via bilstm-cnn. In: Inter- national Conference on Data Mining and Big Data, Springer, pp 373-382
- Mooney RJ, Bunescu RC (2006) Subsequence kernels for relation extrac- tion. In: Advances in neural information processing systems, pp 171-178
- Xu Y, Mou L, Li G, Chen Y, Peng H, Jin Z (2015) Classifying relations via long short term memory networks along shortest dependency paths. In: proceedings of the 2015 conference on empirical methods in natural language processing, pp 1785-1794
- Cai R, Zhang X, Wang H (2016) Bidirectional recurrent convolutional neu- ral network for relation classification. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol 1, pp 756-765
- Guo X, Zhang H, Yang H, Xu L, Ye Z (2019) A single attention-based com- bination of cnn and rnn for relation classification. IEEE Access 7:12467- 12475
- Jin L, Song L, Zhang Y, Xu K, Ma Wy, Yu D (2020) Relation extraction exploiting full dependency forests. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 34, pp 8034-8041
- Hearst MA (1992) Automatic acquisition of hyponyms from large text cor- pora. In: Proceedings of the 14th conference on Computational linguistics- Volume 2, Association for Computational Linguistics, pp 539-545
- Berland M, Charniak E (1999) Finding parts in very large corpora. In: Proceedings of the 37th annual meeting of the Association for Computa- 206-218
- Lin Y, Liu Z, Sun M (2017) Neural relation extraction with multi-lingual attention. In: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp 34-43
- Lin Y, Shen S, Liu Z, Luan H, Sun M (2016) Neural relation extraction with selective attention over instances. In: Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), vol 1, pp 2124-2133
- Banerjee S, Tsioutsiouliklis K (2018) Relation extraction using multi- encoder lstm network on a distant supervised dataset. In: 2018 IEEE 12th International Conference on Semantic Computing (ICSC), IEEE, pp 235- 238
- Du J, Han J, Way A, Wan D (2018) Multi-level structured self- attentions for distantly supervised relation extraction. arXiv preprint arXiv:180900699
- Ji G, Liu K, He S, Zhao J (2017) Distant supervision for relation extrac- tion with sentence-level attention and entity descriptions. In: Thirty-First AAAI Conference on Artificial Intelligence
- Wang G, Zhang W, Wang R, Zhou Y, Chen X, Zhang W, Zhu H, Chen H (2018) Label-free distant supervision for relation extraction via knowledge graph embedding. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp 2246-2255
- Vashishth S, Joshi R, Prayaga SS, Bhattacharyya C, Talukdar P (2018) Reside: Improving distantly-supervised neural relation extraction using side information. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp 1257-1266
- Qin P, Xu W, Wang WY (2018) Robust distant supervision relation ex- traction via deep reinforcement learning. arXiv preprint arXiv:180509927
- Qin P, Xu W, Wang WY (2018) Dsgan: Generative adversarial training for distant supervision relation extraction. arXiv preprint arXiv:180509929
- Angeli G, Premkumar MJJ, Manning CD (2015) Leveraging linguistic structure for open domain information extraction. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), vol 1, pp 344-354
- Pavlick E, Rastogi P, Ganitkevitch J, Van Durme B, Callison-Burch C (2015) Ppdb 2.0: Better paraphrase ranking, fine-grained entailment re- lations, word embeddings, and style classification. In: Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 2: Short Papers), vol 2, pp 425-430
- Zhang N, Deng S, Sun Z, Chen X, Zhang W, Chen H (2018) Attention- based capsule networks with dynamic routing for relation extraction. arXiv preprint arXiv:181211321
- Ye H, Chao W, Luo Z, Li Z (2016) Jointly extracting relations with class ties via effective deep ranking. arXiv preprint arXiv:161207602
- Miwa M, Bansal M (2016) End-to-end relation extraction using lstms on sequences and tree structures. arXiv preprint arXiv:160100770
- Li F, Zhang M, Fu G, Ji D (2017) A neural joint model for entity and relation extraction from biomedical text. BMC bioinformatics 18(1):198
- Zheng S, Wang F, Bao H, Hao Y, Zhou P, Xu B (2017) Joint extraction of entities and relations based on a novel tagging scheme. arXiv preprint arXiv:170605075
- Xiao Y, Tan C, Fan Z, Xu Q, Zhu W (2020) Joint entity and relation extraction with a hybrid transformer and reinforcement learning based model. In: AAAI, pp 9314-9321
- Bethard S, Carpuat M, Cer D, Jurgens D, Nakov P, Zesch T (2016) Proceedings of the 10th international workshop on semantic evaluation (semeval-2016). In: Proceedings of the 10th International Workshop on Semantic Evaluation (SemEval-2016)
- Gábor K, Buscaldi D, Schumann AK, QasemiZadeh B, Zargayouna H, Charnois T (2018) Semeval-2018 task 7: Semantic relation extraction and classification in scientific papers. In: Proceedings of The 12th International Workshop on Semantic Evaluation, pp 679-688
- Zhang Y, Zhong V, Chen D, Angeli G, Manning CD (2017) Position- aware attention and supervised data improve slot filling. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Process- ing (EMNLP 2017), pp 35-45, URL https://nlp.stanford.edu/pubs/ zhang2017tacred.pdf
- Segura Bedmar I, Martinez P, Sánchez Cisneros D (2011) The 1st ddiextraction-2011 challenge task: Extraction of drug-drug interactions from biomedical texts
- Segura Bedmar I, Martínez P, Herrero Zazo M (2013) Semeval-2013 task 9: Extraction of drug-drug interactions from biomedical texts (ddiextraction 2013). Association for Computational Linguistics
- Di S, Shen Y, Chen L (2019) Relation extraction via domain-aware transfer learning. In: Proceedings of the 25th ACM SIGKDD International Con- ference on Knowledge Discovery & Data Mining, pp 1348-1357
- Sun C, Wu Y (2019) Distantly supervised entity relation extraction with adapted manual annotations. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 33, pp 7039-7046
- Zhang N, Deng S, Sun Z, Chen J, Zhang W, Chen H (2019) Transfer learn- ing for relation extraction via relation-gated adversarial learning. arXiv preprint arXiv:190808507
- Sahu SK, Christopoulou F, Miwa M, Ananiadou S (2019) Inter-sentence relation extraction with document-level graph convolutional neural net- work. arXiv preprint arXiv:190604684
- Guo Z, Zhang Y, Lu W (2019) Attention Guided Graph Convolutional Networks for Relation Extraction pp 241-251, DOI 10.18653/v1/p19-1024, 1906.07510
- Zhang Y, Qi P, Manning CD (2019) Graph Convolution over Pruned Dependency Trees Improves Relation Extraction (2005):2205-2215, DOI