Abstract
A huge number of papers have been published about COVID-19. So much it’s overwhelming. Many papers appear on preprint servers such as arXiv before publication. Researchers and clinicians can get ahead of the curve by making use of these preprint papers, but how to tell what is worth reading? Could there be an automated recommendation mechanism? In this paper we address the question by experimenting with SPECTER document-level vector embedding which establishes the representations by incorporating state-of-the-art Transformer models, such as SciBERT, a BERT variant tailored to scientific text. Meanwhile, the dataset we choose to apply SPECTER embedding is the CORD-19 dataset.
This work was supported by JST (JPMJMS2033). The last author would like to thank Advanced Telecommunications Research Institute for his research visit there.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Beltagy, I., Lo, K., Cohan, A.: SciBERT: a pretrained language model for scientific text (2019)
Chen, Q., Allot, A., Lu, Z.: Keep up with the latest coronavirus research. Nature 579(7798), 193 (2020). https://doi.org/10.1038/d41586-020-00694-1, https://www.ncbi.nlm.nih.gov/pubmed/32157233
Chen, Q., Allot, A., Lu, Z.: LitCovid: an open database of COVID-19 literature. Nucleic Acids Res. 49, D1534–D1540 (2020)
Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding (2019)
Neumann, P.M.: The mathematical Writings of Évariste Galois. European Mathematical Society (2011)
Pedregosa, F., et al.: Scikit-learn: machine learning in Python. J. Mach. Learn. Res. 12, 2825–2830 (2011)
SCImago: SJR - SCImago Journal & Country Rank [Portal] (2021). http://www.scimagojr.com. Accessed 29 Apr 2021
Vaswani, A., et al.: Attention is all you need (2017)
Wang, L.L., Lo, K.: Text mining approaches for dealing with the rapidly expanding literature on COVID-19. Brief. Bioinform. 22(2), 781–799 (2020). https://doi.org/10.1093/bib/bbaa296
Acknowledgment
The authors are grateful to Ryohei Sasano for his help with the experimental part of this work.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Xu, T., Hinton, N., Bennett, M.T., Maruyama, Y. (2022). Natural Language Processing for Scientific Paper Evaluation: Comparing Human and Machine Judgements. In: Stephanidis, C., Antona, M., Ntoa, S., Salvendy, G. (eds) HCI International 2022 – Late Breaking Posters. HCII 2022. Communications in Computer and Information Science, vol 1655. Springer, Cham. https://doi.org/10.1007/978-3-031-19682-9_90
Download citation
DOI: https://doi.org/10.1007/978-3-031-19682-9_90
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-19681-2
Online ISBN: 978-3-031-19682-9
eBook Packages: Computer ScienceComputer Science (R0)