Abstract
The task of relation extraction aims at classifying the semantic relations between entities in a text. When coupled with named-entity recognition these can be used as the building blocks for an information extraction procedure that results in the construction of a Knowledge Graph. While many NLP libraries support named-entity recognition, there is no off-the-shelf solution for relation extraction. In this paper, we evaluate and compare several state-of-the-art approaches on a subset of the FewRel data set as well as a manually annotated corpus. The custom corpus contains six relations from the area of market research and is available for public use. Our approach provides guidance for the selection of models and training data for relation extraction in realworld projects.
References (24)
- J. Albrecht, A. Belger, R. Blum, and R. Zimmermann, "Business Analytics on Knowledge Graphs for Market Trend Analysis," in Proceedings of LWDA 2019 (CEUR Workshop Proceedings 2454), Berlin, Germany, 2019. Available: http://ceur-ws.org/Vol-2454
- A. Belger, R. Budinich, R. Blum, M. Zablocki, and R. Zimmermann, "Market and Technology Monitoring driven by Knowledge Graphs," Fraunhofer SCS; Technische Hochschule Nürnberg Georg Simon Ohm, 2020.
- X. Han et al., "FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation," 2018. Available: https://arxiv.org/pdf/1810.10147
- J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," 2018. Available: https://arxiv.org/pdf/1810.04805
- A. Vaswani et al., "Attention is All you Need," in Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017, pp. 5998- 6008. Available: http://papers.nips.cc/paper/7181-attention-is-all-you-need
- J. Lee, S. Seo, and Y. S. Choi, "Semantic Relation Classification via Bidirectional LSTM Networks with Entity-aware Attention using Latent Entity Typing," Jan. 2019. Available: https://arxiv.org/pdf/ 1901.08163
- J. Pennington, R. Socher, and C. Manning, "Glove: Global Vectors for Word Representation,"Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532-1543, 2014, doi: 10.3115/v1/D14-1162.
- M. E. Peters et al., "Deep contextualized word representations," Feb. 2018. Available: https:// arxiv.org/pdf/1802.05365
- Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun, and Q. Liu, "ERNIE: Enhanced Language Representation with Informative Entities," 2019. Available: https://arxiv.org/pdf/1905.07129
- A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, "Translating Embeddings for Modeling Multi-relational Data," inAdvances in Neural Information Processing Systems 26: 27 th Annual Conference on Neural Information Processing Systems 2013: Proceedings of a meeting held December 5-8, Lake Tahoe, Nevada, United States, 2013, pp. 2787-2795. Available: http:// papers.nips.cc/paper/5071-translating-embeddings-for-modeling-multi-relational-data
- S. Wu and Y. He, "Enriching Pre-trained Language Model with Entity Information for Relation Classification," May. 2019. Available: https://arxiv.org/pdf/1905.08284
- L. B. Soares, N. FitzGerald, J. Ling, and T. Kwiatkowski, "Matching the Blanks: Distributional Similarity for Relation Learning," 2019. Available: https://arxiv.org/pdf/1906.03158
- N. Noy, Y. Gao, A. Jain, A. Narayanan, A. Patterson, and J. Tylor, "Industry-scale Knowledge Graphs: Lessons and Challenges,"ACM Queue, vol. 17, no. 2, pp. 1-28, 2019, doi: 10.1145/3329781.3332266.
- Y. Zhang, V. Zhong, D. Chen, G. Angeli, and C. D. Manning, "Position-aware Attention and Supervised Data Improve Slot Filling," inProceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Stroudsburg, PA, USA, 2017.
- S. Riedel, L. Yao, and A. McCallum, "Modeling Relations and Their Mentions without Labeled Text," inJoint European Conference on Machine Learning and Knowledge Discovery in Databases, 2010, pp. 148-163. Available: https://link.springer.com/chapter/10.1007/978-3-642-15939-8_10
- I. Hendrickx et al., "SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations between Pairs of Nominals," Proceedings of the 5 th International Workshop on Semantic Evaluation, pp. 33- 38, 2010.
- T. Gao et al., "FewRel 2.0: Towards More Challenging Few-Shot Relation Classification," in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9 th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 2019, pp. 6249-6254.
- F. Hamborg, N. Meuschke, C. Breitinger, and B. Gipp, "news-please: A Generic News Crawler and Extractor," inProceedings of the 15 th International Symposium of Information Science, 2017. Available: https://www.researchgate.net/publication/314072045_news-please_A_Generic_News_ Crawler_and_Extractor
- J.-C. Klie, M. Bugert, B. Boullosa, R. E. de Castilho, and I. Gurevych, "The INCEpTION Platform: Machine-Assisted and Knowledge-Oriented Interactive Annotation,"Proceedings of the 27 th International Conference on Computational Linguistics: System Demonstrations, pp. 5-9, 2018.
- A. Bastos et al., "RECON: Relation Extraction using Knowledge Graph Context in a Graph Neural Network," Sep. 2020. [Online]. Available: https://arxiv.org/pdf/2009.08694
- T. Nayak and H. T. Ng, "Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation Extraction," Nov. 2019. [Online]. Available: https://arxiv.org/pdf/1911.09886
- Y. Liu et al., "RoBERTa: A Robustly Optimized BERT Pretraining Approach," Jul. 2019. Available: https://arxiv.org/pdf/1907.11692
- Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations," Sep. 2019. Available: https://arxiv.org/pdf/ 1909.11942
- Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang, "SQuAD: 100, 000+ Questions for Machine Comprehension of Text,"CoRR, abs/1606.05250, 2016.