Generating Summaries for Scientific Paper Review
2021, ArXiv
Abstract
The review process is essential to ensure the quality of publications. Recently, the increase of submissions for top venues in machine learning and NLP has caused a problem of excessive burden on reviewers and has often caused concerns regarding how this may not only overload reviewers, but also may affect the quality of the reviews. An automatic system for assisting with the reviewing process could be a solution for ameliorating the problem. In this paper, we explore automatic review summary generation for scientific papers. We posit that neural language models have the potential to be valuable candidates for this task. In order to test this hypothesis, we release a new dataset of scientific papers and their reviews, collected from papers published in the NeurIPS conference from 2013 to 2020. We evaluate state of the art neural summarization models, present initial results on the feasibility of automatic review summary generation, and propose directions for the future.
References (26)
- Kritika Agrawal, Aakash Mittal, and Vikram Pudi. 2019. Scalable, semi-supervised extraction of struc- tured information from scientific literature. In Pro- ceedings of the Workshop on Extracting Structured Knowledge from Scientific Publications, pages 11- 20. Nouf Ibrahim Altmami and Mohamed El Bachir Menai. 2020. Automatic summarization of scientific articles: A survey. Journal of King Saud University- Computer and Information Sciences.
- Tom B Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, et al. 2020. Language models are few-shot learners. arXiv preprint arXiv:2005.14165.
- Muthu Kumar Chandrasekaran, Michihiro Yasunaga, Dragomir Radev, Dayne Freitag, and Min-Yen Kan. 2019. Overview and results: Cl-scisumm shared task 2019. arXiv preprint arXiv:1907.09854.
- Arman Cohan, Franck Dernoncourt, Doo Soon Kim, Trung Bui, Seokhwan Kim, Walter Chang, and Nazli Goharian. 2018. A discourse-aware attention model for abstractive summarization of long documents. arXiv preprint arXiv:1804.05685.
- Arman Cohan and Nazli Goharian. 2017. Contextu- alizing citations for scientific summarization using word embeddings and domain knowledge. Proceed- ings of the 40th International ACM SIGIR Confer- ence on Research and Development in Information Retrieval.
- Shrey Desai, Jiacheng Xu, and Greg Durrett. 2020. Compressive summarization with plausibility and salience modeling. In Proceedings of the 2020 Con- ference on Empirical Methods in Natural Language Processing (EMNLP), pages 6259-6274.
- Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. Bert: Pre-training of deep bidirectional transformers for language under- standing. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Tech- nologies, Volume 1 (Long and Short Papers), pages 4171-4186.
- Cong Duy Vu Hoang and Min-Yen Kan. 2010. To- wards automated related work summarization. In Coling 2010: Posters, pages 427-435.
- Jiaxin Ju, Ming Liu, Longxiang Gao, and Shirui Pan. 2020. Scisummpip: An unsupervised scien- tific paper summarization pipeline. arXiv preprint arXiv:2010.09190.
- Elena Lloret, María Teresa Romá-Ferri, and Manuel Palomar. 2013. Compendium: A text summariza- tion system for generating abstracts of research pa- pers. Data & Knowledge Engineering, 88:164-175.
- Ramesh Nallapati, Bowen Zhou, Cicero dos San- tos, Çaglar GuÌ ‡lçehre, and Bing Xiang. 2016. Abstractive text summarization using sequence-to- sequence RNNs and beyond. In Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, pages 280-290, Berlin, Ger- many. Association for Computational Linguistics.
- Shashi Narayan, Shay B. Cohen, and Mirella Lapata. 2018. Don't give me the details, just the summary! topic-aware convolutional neural networks for ex- treme summarization. In Proceedings of the 2018 Conference on Empirical Methods in Natural Lan- guage Processing, pages 1797-1807, Brussels, Bel- gium. Association for Computational Linguistics.
- Shashi Narayan, Shay B. Cohen, and Mirella Lapata. 2019. What is this article about? extreme summa- rization with topic-aware convolutional neural net- works. CoRR, abs/1907.08722.
- Vahed Qazvinian, Dragomir Radev, and Arzucan Özgür. 2010. Citation summarization through keyphrase extraction. In Proceedings of the 23rd in- ternational conference on computational linguistics (COLING 2010), pages 895-903.
- Weizhen Qi, Yu Yan, Yeyun Gong, Dayiheng Liu, Nan Duan, Jiusheng Chen, Ruofei Zhang, and Ming Zhou. 2020. Prophetnet: Predicting future n-gram for sequence-to-sequence pre-training. In Proceed- ings of the 2020 Conference on Empirical Methods in Natural Language Processing: Findings, pages 2401-2410.
- Anna Rogers and Isabelle Augenstein. 2020. What can we do to improve peer review in nlp? arXiv preprint arXiv:2010.03863.
- Francesco Ronzano and Horacio Saggion. 2016. An empirical assessment of citation information in sci- entific summarization. In international conference on applications of natural language to information systems, pages 318-325. Springer.
- Horacio Saggion. 2011. Learning predicate insertion rules for document abstracting. In International Conference on Intelligent Text Processing and Com- putational Linguistics, pages 301-312. Springer.
- Horacio Saggion and Guy Lapalme. 2000. Selective analysis for automatic abstracting: Evaluating in- dicativeness and acceptability. In RIAO, pages 747- 764. Citeseer.
- Abigail See, Peter J. Liu, and Christopher D. Manning. 2017. Get to the point: Summarization with pointer- generator networks. In Proceedings of the 55th An- nual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1073- 1083, Vancouver, Canada. Association for Compu- tational Linguistics.
- Cepi Slamet, AR Atmadja, DS Maylawati, RS Lestari, Wahyudin Darmalaksana, and Muhammad Ali Ramdhani. 2018. Automated text summarization for indonesian article using vector space model. In IOP Conference Series: Materials Science and Engineer- ing, volume 288, page 012037. IOP Publishing.
- Xiaoping Sun and Hai Zhuge. 2018. Summarization of scientific paper through reinforcement ranking on semantic link network. IEEE Access, 6:40611- 40625.
- Shansong Yang, Weiming Lu, Zhanjiang Zhang, Bao- gang Wei, and Wenjia An. 2016. Amplifying sci- entific paper's abstract by leveraging data-weighted reconstruction. Information Processing & Manage- ment, 52(4):698-719.
- Michihiro Yasunaga, Jungo Kasai, Rui Zhang, Alexan- der R Fabbri, Irene Li, Dan Friedman, and Dragomir R Radev. 2019. Scisummnet: A large an- notated corpus and content-impact models for sci- entific paper summarization with citation networks. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 33, pages 7386-7393.
- Jingqing Zhang, Yao Zhao, Mohammad Saleh, and Pe- ter Liu. 2020. Pegasus: Pre-training with extracted gap-sentences for abstractive summarization. In In- ternational Conference on Machine Learning, pages 11328-11339. PMLR.
- Tianyi Zhang, Varsha Kishore, Felix Wu, Kilian Q Weinberger, and Yoav Artzi. 2019. Bertscore: Eval- uating text generation with bert. In International Conference on Learning Representations. Hao Zheng and Mirella Lapata. 2019. Sentence cen- trality revisited for unsupervised summarization. In Proceedings of the 57th Annual Meeting of the Asso- ciation for Computational Linguistics, pages 6236- 6247, Florence, Italy. Association for Computa- tional Linguistics.