Academia.eduAcademia.edu

Outline

State-ofthe-Art Models for Relation Extraction

2021

https://doi.org/10.5121/CSIT.2021.112006

Abstract

The task of relation extraction aims at classifying the semantic relations between entities in a text. When coupled with named-entity recognition these can be used as the building blocks for an information extraction procedure that results in the construction of a Knowledge Graph. While many NLP libraries support named-entity recognition, there is no off-the-shelf solution for relation extraction. In this paper, we evaluate and compare several state-of-the-art approaches on a subset of the FewRel data set as well as a manually annotated corpus. The custom corpus contains six relations from the area of market research and is available for public use. Our approach provides guidance for the selection of models and training data for relation extraction in realworld projects.

References (24)

  1. J. Albrecht, A. Belger, R. Blum, and R. Zimmermann, "Business Analytics on Knowledge Graphs for Market Trend Analysis," in Proceedings of LWDA 2019 (CEUR Workshop Proceedings 2454), Berlin, Germany, 2019. Available: http://ceur-ws.org/Vol-2454
  2. A. Belger, R. Budinich, R. Blum, M. Zablocki, and R. Zimmermann, "Market and Technology Monitoring driven by Knowledge Graphs," Fraunhofer SCS; Technische Hochschule Nürnberg Georg Simon Ohm, 2020.
  3. X. Han et al., "FewRel: A Large-Scale Supervised Few-Shot Relation Classification Dataset with State-of-the-Art Evaluation," 2018. Available: https://arxiv.org/pdf/1810.10147
  4. J. Devlin, M.-W. Chang, K. Lee, and K. Toutanova, "BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding," 2018. Available: https://arxiv.org/pdf/1810.04805
  5. A. Vaswani et al., "Attention is All you Need," in Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, 2017, pp. 5998- 6008. Available: http://papers.nips.cc/paper/7181-attention-is-all-you-need
  6. J. Lee, S. Seo, and Y. S. Choi, "Semantic Relation Classification via Bidirectional LSTM Networks with Entity-aware Attention using Latent Entity Typing," Jan. 2019. Available: https://arxiv.org/pdf/ 1901.08163
  7. J. Pennington, R. Socher, and C. Manning, "Glove: Global Vectors for Word Representation,"Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 1532-1543, 2014, doi: 10.3115/v1/D14-1162.
  8. M. E. Peters et al., "Deep contextualized word representations," Feb. 2018. Available: https:// arxiv.org/pdf/1802.05365
  9. Z. Zhang, X. Han, Z. Liu, X. Jiang, M. Sun, and Q. Liu, "ERNIE: Enhanced Language Representation with Informative Entities," 2019. Available: https://arxiv.org/pdf/1905.07129
  10. A. Bordes, N. Usunier, A. Garcia-Duran, J. Weston, and O. Yakhnenko, "Translating Embeddings for Modeling Multi-relational Data," inAdvances in Neural Information Processing Systems 26: 27 th Annual Conference on Neural Information Processing Systems 2013: Proceedings of a meeting held December 5-8, Lake Tahoe, Nevada, United States, 2013, pp. 2787-2795. Available: http:// papers.nips.cc/paper/5071-translating-embeddings-for-modeling-multi-relational-data
  11. S. Wu and Y. He, "Enriching Pre-trained Language Model with Entity Information for Relation Classification," May. 2019. Available: https://arxiv.org/pdf/1905.08284
  12. L. B. Soares, N. FitzGerald, J. Ling, and T. Kwiatkowski, "Matching the Blanks: Distributional Similarity for Relation Learning," 2019. Available: https://arxiv.org/pdf/1906.03158
  13. N. Noy, Y. Gao, A. Jain, A. Narayanan, A. Patterson, and J. Tylor, "Industry-scale Knowledge Graphs: Lessons and Challenges,"ACM Queue, vol. 17, no. 2, pp. 1-28, 2019, doi: 10.1145/3329781.3332266.
  14. Y. Zhang, V. Zhong, D. Chen, G. Angeli, and C. D. Manning, "Position-aware Attention and Supervised Data Improve Slot Filling," inProceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Stroudsburg, PA, USA, 2017.
  15. S. Riedel, L. Yao, and A. McCallum, "Modeling Relations and Their Mentions without Labeled Text," inJoint European Conference on Machine Learning and Knowledge Discovery in Databases, 2010, pp. 148-163. Available: https://link.springer.com/chapter/10.1007/978-3-642-15939-8_10
  16. I. Hendrickx et al., "SemEval-2010 Task 8: Multi-Way Classification of Semantic Relations between Pairs of Nominals," Proceedings of the 5 th International Workshop on Semantic Evaluation, pp. 33- 38, 2010.
  17. T. Gao et al., "FewRel 2.0: Towards More Challenging Few-Shot Relation Classification," in Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9 th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 2019, pp. 6249-6254.
  18. F. Hamborg, N. Meuschke, C. Breitinger, and B. Gipp, "news-please: A Generic News Crawler and Extractor," inProceedings of the 15 th International Symposium of Information Science, 2017. Available: https://www.researchgate.net/publication/314072045_news-please_A_Generic_News_ Crawler_and_Extractor
  19. J.-C. Klie, M. Bugert, B. Boullosa, R. E. de Castilho, and I. Gurevych, "The INCEpTION Platform: Machine-Assisted and Knowledge-Oriented Interactive Annotation,"Proceedings of the 27 th International Conference on Computational Linguistics: System Demonstrations, pp. 5-9, 2018.
  20. A. Bastos et al., "RECON: Relation Extraction using Knowledge Graph Context in a Graph Neural Network," Sep. 2020. [Online]. Available: https://arxiv.org/pdf/2009.08694
  21. T. Nayak and H. T. Ng, "Effective Modeling of Encoder-Decoder Architecture for Joint Entity and Relation Extraction," Nov. 2019. [Online]. Available: https://arxiv.org/pdf/1911.09886
  22. Y. Liu et al., "RoBERTa: A Robustly Optimized BERT Pretraining Approach," Jul. 2019. Available: https://arxiv.org/pdf/1907.11692
  23. Z. Lan, M. Chen, S. Goodman, K. Gimpel, P. Sharma, and R. Soricut, "ALBERT: A Lite BERT for Self-supervised Learning of Language Representations," Sep. 2019. Available: https://arxiv.org/pdf/ 1909.11942
  24. Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, and Percy Liang, "SQuAD: 100, 000+ Questions for Machine Comprehension of Text,"CoRR, abs/1606.05250, 2016.