14
views
0
recommends
+1 Recommend
1 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: not found

      Scholarly literature mining with information retrieval and natural language processing: Preface

      editorial
      1 , , 2 , 3
      Scientometrics
      Springer International Publishing

      Read this article at

      ScienceOpenPublisherPMC
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Introduction This special issue features the work of authors originally coming from different communities: bibliometrics/scientometrics (SCIM), information retrieval (IR) and, as an emerging player gaining more relevance for both aforementioned fields, natural language processing (NLP). The work presented in their papers combine ideas from all these fields, having in common that they all are using the scholarly data well known in scientometrics and solving problems typical to scientometric research. They model and mine citations, as well as metadata of bibliographic records (authorships, titles, abstracts sometimes), which is common practice in SCIM. They also mine and process fulltexts (including in-text references and equations) which is common practice in IR and requires established NLP text mining techniques. IR collections are utilised to ensure reproducible evaluations; creating and sharing test collections in evaluation initiatives such as CLEF eHealth1 is common IR tradition that is also prominent in NLP, eg., by the CL-SciSumm shared task.2 From an IR perspective, surprisingly, scholarly information retrieval and recommendation, though gaining momentum, have not always been the focus of research in the past. Besides operating on a rich set of data for researchers in all three disciplines to play with, scholarly search poses challenges in particular for IR due to the complex information needs that require different approaches than known from, e.g., Web search, where information needs are simpler in many cases. As an example, the current COVID-19 crisis shows that hybrid SCIM/IR/NLP approaches are increasingly required to ensure researchers get access to important relevant and high-quality information, often only available on preprint servers, in a short period of time (Brainard 2020; Fraser et al. 2020; Kwon 2020; Palayew et al. 2020). These kinds of complex information needs pose challenges which have been recognised by the Information Retrieval community that quickly launched the TREC-COVID initiative run by NIST (Roberts et al. 2020), demonstrating the timeliness of our endeavour and this special issue. Working on scholarly material thus has incentives for researchers in Information Retrieval but we believe the challenges can only be tackled effectively by all three communities as a whole. The NLP community has initiated a similar activity with a dedicated workshop series NLP COVID-19 Workshop3 which is running at major NLP conferences (ACL & EMNLP) in 2020. With the surge of “scholarly big data” (Giles 2013), Bibliometrics and Information Retrieval in combination with NLP methods have seen a recent renaissance that resulted in a series of special issues: “Combining Bibliometrics and Information Retrieval” (Mayr and Scharnhorst 2015) in Scientometrics (2015). “Bibliometric-enhanced Information Retrieval” (Cabanac et al. 2018) in Scientometrics (2018). “Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries” (Mayr et al. 2018) in International Journal on Digital Libraries (2018). “Mining Scientific Papers: NLP-enhanced Bibliometrics” (Atanassova et al. 2019) in Frontiers in Research Metrics and Analytics (2019). Special issue papers This special issue on “Scholarly literature mining with Information Retrieval and Natural Language Processing” presents works intersecting Bibliometrics and Information Retrieval, utilising Natural Language Processing (NLP). The special issue was announced via an open call for papers4. In response to the CFP, we received 24 submissions which were reviewed by two to three reviewers (for overlapping papers, eg., IR and NLP, we selected reviewers from both domains). Eventually, the guest editors accepted 14 papers. Nine papers have been rejected and one paper was withdrawn by the authors during the reviewing rounds. In the following we provide an overview of the 14 papers organised into 3 clusters. We introduce the paper ordering of the special issue in Table 1. To generate a lightweight overview of the variety of the papers we identified the research Tasks and Area of Application, the used Corpus, Objects, and Methods of each contribution. The papers in this special issue appear in the following sequence. We decided to start with a set of more classical papers featuring scientometric methods like network analysis and bibliographic data from the Web of Science, Scopus or similar resources. The second set of papers is more IR oriented: papers mine fulltexts and they use techniques like embeddings and neural networks. The third cluster of papers contains NLP-oriented papers which are, for instance, specialised in summarisation and utilise scholarly documents. Cluster 1. SCIM with IR and NLP Lietz: Drawing impossible boundaries: field delineation of Social Network Science. Schneider et al.: Continued post-retraction citation of a fraudulent clinical trial report, eleven years after it was retracted for falsifying data. Kreutz et al.: Evaluating semantometrics from computer science publications. Haunschild & Marx: Discovering seminal works with marker papers. Lamirel et al.: An overview of the history of Science of Science in China based on the use of bibliographic and citation data: a new method of analysis based on clustering with feature maximization and contrast graphs. Cluster 2. IR and Text-mining of scholarly literature Nogueira et al.: Navigation-based candidate expansion and pretrained language models for citation recommendation. Greiner-Petter et al.: Math-word embedding in math search and semantic extraction. Carvallo et al.: Automatic document screening of medical literature using word and text embeddings in an active learning setting. Saier & Färber: unarXive: a large scholarly data set with publications’ full-text, annotated in-text citations, and links to metadata. Cluster 3. NLP-oriented papers on scholarly literature Zerva et al.: Cited text span identification for scientific summarisation using pre-trained encoders. La Quatra et al.: Exploiting pivot words to classify and summarize discourse facets of scientific papers. AbuRa’ed et al.: Automatic related work section generation: experiments in scientific document abstracting. Jimenez et al.: Automatic prediction of citability of scientific articles by stylometry of their titles and abstracts. Portenoy & West: Constructing and evaluating automated literature review systems. Table 1 Overview of the articles in this special issue Task Area of application Corpus Objects Methods Lietz Field delineation Social network science Web of science Metadata (title, abstract, keywords), references Clustering, network analysis Schneider, Ye, Hill, & Whitehorn Analysing citing papers of a retracted study Clinical science Google scholar, web of science Seed paper, citations, retraction notices Network analysis, citation context analysis, retraction status visibility analysis Kreutz, Sahitaj, & Schenkel Spotting seminal work; classifying papers Computer science DBLP Fulltext Classification using words, semantics, topics and publication years Haunschild & Marx Spotting seminal work Physics Microsoft academic, web of science References, time Reference publication year spectroscopy Lamirel, Chen, Cuxac, Al Shehabi, Dugué & Liu Mapping the evolution of a country’s scientific production Science in China China national knowledge infrastructure database metadata (title, abstract, authors), dictionary of Chinese names Clustering, topic modelling, network analysis Nogueira, Jiang, Cho, & Lin Ranking citation recommendations Computer science, biomedicine DBLP, open research, PubMed Fulltext Document ranking model, embeddings Greiner-Petter, Youssef, Ruas, Miller, Schubotz, Aizawa & Gipp Discovering mathematical term similarity and analogy and query expansions Mathematics arXiv Fulltext Embeddings Carvallo, Parra, Lobel, & Soto Paper screening for evidence-based medicine Medicine CLEF eHealth, Epistemonekos Fulltext Document ranking model, query expansion, embeddings Saier & Färber Dataset creation Fields of arXiv preprints arXiv, Microsoft academic graph Fulltext, in-text citations, linked data Data integration, descriptive statistics Zerva, Nghiem, Nguyen, & Ananiadou Paper summarization (from citations) Natural language processing CL-SciSumm Fulltext, in-text citations Neural networks La Quatra, Cagliero, & Baralis Discourse facet summarization Natural language processing CL-SciSumm Fulltext, in-text citations Neural networks AbuRa’ed, Saggion, Shvets, & Bravo Citation sentence production Text summarization ScisummNet, Open academic graph, microsoft academic graph, RWSData Fulltext Neural networks Jimenez, Avila, Dueñas, & Gelbukh Citation forecasting The scientific literature Scopus Metadata (title + abstract) Statistics, stylometry Portenoy & West Generation of a literature review of a field Community detection in graphs, misinformation studies, science communication Web of science References, paper titles Text similarity, supervised learning, embeddings We hope the selection of papers in this special issue will be interesting and enjoyable for researchers coming from all relevant fields and provides a starting point for future explorations in the field.5

          Related collections

          Most cited references9

          • Record: found
          • Abstract: not found
          • Article: not found

          Pandemic publishing poses a new COVID-19 challenge

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            How swamped preprint servers are blocking bad coronavirus research

            Diana Kwon (2020)
              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              TREC-COVID: Rationale and Structure of an Information Retrieval Shared Task for COVID-19

              Abstract TREC-COVID is an information retrieval (IR) shared task initiated to support clinicians and clinical research during the COVID-19 pandemic. IR for pandemics breaks many normal assumptions, which can be seen by examining nine important basic IR research questions related to pandemic situations. TREC-COVID differs from traditional IR shared task evaluations with special considerations for the expected users, IR modality considerations, topic development, participant requirements, assessment process, relevance criteria, evaluation metrics, iteration process, projected timeline, and the implications of data use as a post-task test collection. This article describes how all these were addressed for the particular requirements of developing IR systems under a pandemic situation. Finally, initial participation numbers are also provided, which demonstrate the tremendous interest the IR community has in this effort.
                Bookmark

                Author and article information

                Contributors
                guillaume.cabanac@univ-tlse3.fr
                ifrommholz@acm.org
                philipp.mayr@gesis.org
                Journal
                Scientometrics
                Scientometrics
                Scientometrics
                Springer International Publishing (Cham )
                0138-9130
                1588-2861
                17 November 2020
                : 1-6
                Affiliations
                [1 ]GRID grid.508721.9, Computer Science Department, IRIT UMR 5505 CNRS, , University of Toulouse, ; 118 Route de Narbonne, 31062 Toulouse Cedex 9, France
                [2 ]GRID grid.15034.33, ISNI 0000 0000 9882 7057, University of Bedfordshire, ; Luton, LU1 3JU UK
                [3 ]GRID grid.425053.5, ISNI 0000 0001 1013 1176, GESIS – Leibniz Institute for the Social Sciences, ; Cologne, Germany
                Author information
                http://orcid.org/0000-0003-3060-6241
                http://orcid.org/0000-0002-5622-5132
                http://orcid.org/0000-0002-6656-1658
                Article
                3763
                10.1007/s11192-020-03763-4
                7670972
                d30d0eeb-4fbc-4533-a9d4-76be98f43642
                © Akadémiai Kiadó, Budapest, Hungary 2020

                This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.

                History
                : 9 October 2020
                Categories
                Article

                Computer science
                Computer science

                Comments

                Comment on this article