Introduction

Both government agencies and academic institutions produce a vast volume of data that is openly accessible and free for public use. Thousands of scientific journal research articles have been published by analyzing these datasets. It is paramount to track how the datasets have been used within the research community to understand and verify the return on investments in these datasets produced by numerous publicly-funded research programs. Although the practice of journals requiring persistent and resolvable references to data has recently been improving, the non-citation of datasets remains a persistent and sizable challenge. As a result, existing bibliographic databases, such as Scopus and Web of Science, don’t have sufficient direct linkages built between the datasets used in an article with dataset DOIs, which uniquely indicate the original source of the data. Linking datasets with the publication increases the reproducibility and transparency of the research outcomes and enhances Findability, Accessibility, Interoperability, and Reusability (FAIR) principles in data management and stewardship (). Conducting analyses using Earth Science data often requires expert users who have an advanced understanding of scientific data formats and algorithms, as well as knowledge of the fundamental elements of Earth Science. However, this poses several challenges for data centers with diversified user communities, especially those from various geographic regions, and those possessing different levels of education, expertise, and research background. The authors’ institution has created a library of peer-reviewed publications that directly links data collections and their research applications. Making relevant publications a prominent aspect of the dataset documentation can guide non-expert users of the data in discovering and narrowing down the data and variables to meet their specific needs. This collaboration has developed an open-sourced method and tool that can automatically extract peer-reviewed publication records from bibliographic sources to create a publication library with the data-to-publication linkage.

Current State of Linking Datasets and Documents

The emergence of dataset citation in the late 1990s has been influenced by the emphasis on open science, data sharing, and reproducibility, giving credit to dataset creators and providing metrics for dataset use (; ; ; ). The first principles of data citations were established by CODATA-ICSTI (), indicating that dataset citations should have the same significance as other citations in the scholarly record. To be considered article-quality, datasets must have metadata, undergo peer review, be searchable and discoverable in databases, and be cited (; ). Despite these efforts, links between datasets and documents still need to be improved due to the low citation of datasets in the research literature (; ; ; , , ). Reports have shown that over 85% of datasets covered by Web of Science’s Data Citation Index were uncited (; ).

Dataset DOI registries such as DataCite () provide persistent identifiers for research data, making it easier to discover and cite datasets. They also provide metadata to help users understand the context and content of the datasets. The dataset registry metadata is used by a number of services to discover and link datasets, including the Web of Science’s Data Citation Index.

Initiative for Open Abstracts () and Initiative for Open Citations () are initiatives led by DataCite to promote data sharing and open science. I4OA advocates for open access to scholarly literature, while I4OC encourages publishers to make their citation metadata openly available. Another initiative promoting open science and data sharing is Scholix () which aims to link research data and literature by creating a framework for exchanging information between scholarly communication platforms. These initiatives contribute to the growing movement towards open science and data sharing, which is crucial for advancing scientific research and innovation.

Dataset repositories and data-sharing initiatives are important tools for promoting open data and making research more transparent and reproducible. By enabling researchers to easily find and reuse datasets, these initiatives help to increase the impact and visibility of research and facilitate collaboration between researchers in different fields.

In addition to Web of Science’s Data Citation Index, other major services linking datasets and documents are Scopus (), Crossref, and Google Scholar. Web of Science and Scopus are popular subscription-based bibliographic databases that cover a range of journals, books, and proceedings selected based on journal impact factor (Web of Science), or by independent subject boards or advisory experts (Scopus). Crossref is a non-profit organization that provides a Digital Object Identifier (DOI) registration service for scholarly content (). Crossref provides infrastructure for creating and registering DOIs for various research outputs, such as journal articles, books, book chapters, conference proceedings, reports, preprints, and datasets. Crossref also operates the Crossref OpenCitations Index (COCI), which is a database of open DOI-to-DOI citations that allows for the tracking of citation links between research publications (). The dataset registry DataCite became a bibliometric database when it started providing the DOIs of the documents citing the datasets (). DataCite dataset-to-document linkage is based on Crossref’s COCI connections.

In contrast to the bibliographic databases described above, Google Scholar indexes documents rather than their sources (; ). It has an undisclosed number of records, likely due to its constant improvement in indexing and searching (). In addition to its comprehensive coverage, Google Scholar includes dissertations, books, and conference proceedings that are not covered by Web of Science or Scopus. However, there are cases where Google Scholar indexes non-scholarly content, such as book reviews, low-impact documents, non-refereed sources, and duplicate records that cause extra citations (; ; ; ). The limited metadata provided by Google Scholar makes it difficult to use for bibliometric research (; ; ).

Recently emerging OpenAlex () is a free, open-source scientific knowledge graph. It contains metadata for hundreds of millions of scholarly entities, including works, authors, and venues. OpenAlex provides a free API that allows search over article titles, abstracts, and texts, making it a Google Scholar competitor. As of August 2023, the number of journal articles and books covered by OpenAlex is estimated to be 209 million, as opposed to Google Scholar, which is around 389 million. The OpenAlex API delivers exhaustive bibliographic details for the items retrieved, in contrast to Google Scholar which only provides URLs that require further processing to extract necessary bibliographic metadata.

The above-described bibliometric sources can be searched by the dataset DOI to find datasets linked to research documents. The dataset repository DataCite can also identify publications linked to datasets via DOIs of datasets registered with DataCite. These approaches require that datasets are cited with their DOI. Due to low dataset citation rates, automated approaches have been suggested to identify datasets using their mentions. Lane et al. () summarized machine learning algorithms for detecting socioeconomic sciences data collections. Duan et al. () proposed machine learning methods to identify data collection entities in Earth Science papers based on the names of instruments, missions, and major dataset variables. These approaches require access to the document texts, which can be complicated due to publisher paywalls, copyright restrictions, limited journal availability, language barriers, and publisher embargoes on articles. In addition, these automated approaches are still too immature to be the primary method of mapping data collections to research citations used in data center libraries, due to the ambiguity of referencing utilized datasets.

There are several citation management systems (CMS) currently available (). Among those, the most widely used are the commercial products EndNote and RefWorks, and freely available products Mendeley () and Zotero (, 2023). The Zotero server, zotero.org, is hosted on the Amazon Web Services (AWS) cloud. While Zotero server software is open source, the code only supports server-to-client interaction with the zotero.org server instance. There are several studies comparing these CMSs (). In the creation of a citation library at the Goddard Earth Sciences Data and Information Services Center (), many factors were considered in the selection of the CMS, including platform support, cost, the versatility of citation export and import, content sharing between team members, citation attachment (PDF) management, and the availability of an application programming interface (API), as well as versatility of citation tagging needed for linking citations and datasets.

To address the lack of dataset citations and obtain an exhaustive collection of document citations linked to NASA Earth Science datasets, we built a library of document citations based on Zotero CMS. The datasets used were all archived at the GES DISC, one of NASA’s Earth Observing System Data and Information System (EOSDIS) data archive centers (). To demonstrate that a search of Google Scholar can provide counts of document citations that exceed searches of nominal bibliometric databases, we used over 1,500 GES DISC public datasets. These datasets have been registered with DataCite following NASA’s Earth Science Data and Information System (ESDIS) Project DOI registration process () and have a common DOI prefix, 10.5067.

Study Contributions

The study proposes an automated, high-precision solution for creating a trusted dataset-document citation linkage. This solution relies on a controlled dictionary that manages metadata for datasets and is used to search over Google Scholar. The proposed methodology for processing Google Scholar results allows for retaining deduplicated citations of specific document types such as books, book chapters, theses, journal papers, and conference proceedings. Collected document citations complement the dataset-document linkage which is created by searching Google Scholar and major bibliometric databases for dataset DOIs. Collected documents are added to the dataset landing pages and made available to the dataset users.

In addition, the study proposes a document-to-dataset linking approach that links various versions of datasets to all documents that reference any of those versions (and no versions in many cases). This approach ensures that research associated with prior dataset versions is not lost and is directly available to data users.

Automated Citation Collection by Dataset DOI search

We utilized available Application Programming Interfaces (APIs) from Crossref, Scopus, and DataCite to automate the retrieval of documents citing the datasets. These APIs provide a list of document DOIs linked to the dataset DOIs. As the Web of Science API is not available, we used the Web of Science web interface to acquire citations and their references for a given dataset DOI. To improve efficiency, we searched the Web interface for the dataset DOI prefix, 10.5067, and then processed resulting citations and their references to obtain dataset–document linkages. We used the subscription API, SerpAPI, from Google Scholar, which returns URLs of documents linked to the dataset DOIs. To determine the citations of these documents, we employed the Zotero translation service. To streamline the result merging process, only documents with DOIs were retained. We queried Crossref to determine the types of documents, which include books, book chapters, research articles, conference proceedings, theses, and reports. We excluded duplicate content by filtering out discussion papers and preprints. Figure 1 provides a diagram describing the dataset–document citation-acquiring process.

Figure 1 

Flow diagram of obtaining dataset–document linkages from bibliographic sources.

As of the middle of 2023, the total count of document citations found by all considered bibliographic sources is 2,724. Figure 2 shows a fraction of citations found by each individual source from the total citation count. As seen in Figure 2, 79% of the citations found by all considered sources are returned by Google Scholar.

Figure 2 

Citations found in each bibliographic source as a percentage of total citations.

Figure 3 shows the unique citation fraction contributed to the total citation count by an individual bibliographic source. As seen in Figure 3, Google Scholar contributes 18% of citations that are not provided by any other sources.

Figure 3 

Percentage of unique citations per bibliographic source.

Our results demonstrate that out of all considered bibliographic sources, Google Scholar can return the largest number of citations with the largest count of citations not found in any other sources. This can be explained by several factors. Web of Science and Scopus have limited coverage due to their selectiveness. Crossref only recently started gaining access to citation references. DataCite relies on Crossref for the citation index, so it is even more limited than Crossref. Google Scholar indexes vast amounts of citations including international and less selective journals that might not be covered by Scopus and Web of Science.

Automated Citation Collection from Google Scholar

In our work, we use Google Scholar for keyword searches because of its vast citation coverage, availability, and low cost of its API. With an increasing number of open-access publications, Google Scholar gains access to those document texts. So instead of using difficult-to-access texts from publishers such as Elsevier and Springer, we use Google Scholar.

The objective of the keyword search is to automatically select keywords for the datasets so that these keywords return citations that are linked to these datasets with high certainty. While the datasets can be described by many keywords, the selection of the keywords should be done to obtain as many citations as possible, with minimal erroneous results due to dataset name ambiguity. The keywords should thus be specific enough to find publications linked to the dataset, and not to the same word that has the dataset name but which is not related to that dataset. For example, if we search for the Ozone Monitoring Instrument (OMI) bromine oxide (BrO) dataset ‘OMBRO’ () we will get a lot of publications in the Portuguese or Spanish language (where ‘ombro’ means ‘shoulder’) that did not use this dataset.

The GES DISC dataset collection metadata are stored in NASA’s Earth Observing System Data and Information System (EOSDIS) Common Metadata Repository (), which is governed by the Global Change Master Directory () Keywords ontology. GCMD Keywords is a hierarchical set of controlled Earth Science vocabularies that allows for a consistent description of Earth Science datasets. The GCMD Keywords are organized into twelve sets, each describing Earth science keywords, platforms, instruments, data centers, locations, projects, services, and data resolution. Using these keywords and the dataset’s unique name can help narrow down document citations. For the Google Scholar query, we use the dataset’s essential metadata, such as the dataset’s unique name defined by Earth Science Data Type, or ESDT, (), and the dataset’s project, instrument, and mission defined by GCMD Keywords. ESDT is widely used at NASA data archives to uniquely identify datasets and thus is used by many researchers in publications to refer to these datasets. In addition, we add the ‘NASA’ keyword to further eliminate possible ambiguities. The query is thus assembled as ‘<Dataset ESDT> (<Dataset Platform> | <Dataset Instrument> | <Dataset Project>) NASA’. When searching for the dataset OMBRO (), the Google Scholar query is: ‘OMBRO’ (‘Aura’ | ‘OMI’) ‘NASA’, where ‘Aura’ is the satellite platform carrying the ‘OMI’ instrument that collected the data that was used to produce the ‘OMBRO’ dataset.

Figure 4 presents the document citation counts obtained from bibliometric sources searched by dataset DOIs, and by Google Scholar search via a combination of the dataset ESDT and GCMD Keywords. As can be seen in this figure, the DOI search produced just a few results in earlier years, as datasets had not yet been widely assigned DOIs. In these years the documents are still found by the Google Scholar keywords search. As the number of citations found by dataset DOI search grows over the years, so does the number of citations found by the dataset keywords. This growth tendency shows that the quantity of documents citing datasets by DOI is approaching the number of documents citing datasets by their ESDT name.

Figure 4 

Yearly counts of document citations found by dataset DOI and Keyword search.

The overlap of documents citing datasets by both DOI and mentioning them by ESDT is examined as shown in Figure 5. As seen in Figure 5, the overlap exists, but the quantities of the documents citing datasets by only DOI and only ESDT exceed their overlap. This can be explained by ESDT not being included in the dataset citation, as shown in the current citation for the ‘OMBRO’ dataset:

Figure 5 

Yearly counts of document citations found exclusively by dataset DOI and Keyword search and document citations found by both DOI and Keywords.

Chance, K. (), OMI/Aura Bromine Monoxide (BrO) Total Column 1-orbit L2 Swath 13 × 24 km V003, Greenbelt, MD, USA, Goddard Earth Sciences Data and Information Services Center (GES DISC), 10.5067/Aura/OMI/DATA2006

In citations prior to the DOI’s assignment to this dataset, the ‘OMBRO’ dataset ESDT name was used:

Suleiman, R. M., Chance, K., Liu, X., González Abad, G., Kurosu, T. P., Hendrick, F., & Theys, N. (2018). OMI total bromine monoxide (OMBRO) data product: Algorithm, retrieval and measurement comparisons. Atmospheric Measurement Techniques Discuss. https://doi.org/10.5194/amt-2018-1

Validation of Google Scholar Results

When calculating precision, we considered all articles published in 2021, totaling 945 papers. Out of these, 847 papers, or roughly 90%, had their PDF files available. These PDFs were then converted to ASCII format using the Linux ‘pdftotext’ utility. Each ASCII file was rigorously checked for Short Names, a process guided by findings from Google Scholar. The search within the ASCII files included both specific terms and their variations. Notably, 20 ASCII files did not yield any results from the automated term search, prompting a manual review. Our analysis revealed that Google Scholar’s matching capability was superior. For instance, in the paper titled ‘Aerosol-Cloud Interaction with Summer Precipitation over Major Cities in Eritrea,’ Google Scholar adeptly detected ‘TRMM_3B42_Daily,’ even though it was referenced in the paper as ‘(TRMM) 3B42 (daily).’ Furthermore, a mere three files, albeit relevant to the overarching theme, failed to clearly specify the datasets they were built upon. With these findings, we concluded that the precision of our methodology stands at an impressive 99.6%.

Comparing Google Scholar’s Search Results with OpenAlex

Google Scholar keyword results were compared with OpenAlex output by searching the latter with the same keyword combinations. For this, OpenAlex search, like Google Scholar’s, was limited to papers in English language documents with document types such as books, book chapters, dissertations, and journal and proceedings articles, excluding pre-prints and peer reviews. Only documents that have DOIs were counted. The difference in searches was that for the OpenAlex search the keyword ‘NASA’ was not added. Instead, we relied on the OpenAlex ‘relevance’ score and limited the results to ones having a score higher than 0.99. Relying on the relevance score rather than additional keywords allowed us to produce more hits that were evaluated for relevance by OpenAlex. Since Google Scholar does not have a relevance score, omitting ‘NASA’ or similar keywords indicating that the data were acquired from the archive produces erroneous results for some of the datasets. As we consider precision as a factor more important than recall, our search emphasis is on fewer results with better precision. Similar to Google Scholar keyword search results, the precision of OpenAlex results was evaluated by examining the documents published in 2018. In 2018 there were 412 documents, for which 381 PDF files were found. Similar to Google Scholar evaluation, PDF files were converted to ASCII files, which in turn were searched for the dataset name keywords. All 381 files contained the keywords, which allowed us to conclude 100% accuracy of the OpenAlex keyword search.

The results reported in Figure 6 show that up to the year 2018, OpenAlex slightly outperformed Google Scholar, with its performance significantly degrading after the year 2019.

Figure 6 

Yearly counts of document citations found exclusively by Google Scholar and OpenAlex Keyword search and document citations found by both Google Scholar and OpenAlex.

These results show that as OpenAlex continues improving it should be reevaluated in the future as an alternative solution to Google Scholar. We also looked at using OpenAlex for the referenced dataset DOI search and as it produced insignificant results, we concluded that it is not yet a viable alternative to Google Scholar.

Automated Citation Collection Applications

Data providers and science teams that archive datasets at the data center often collect citations of publications about their data collection algorithms, data validation, and overview, as well as publications about applications that use their datasets. A science team’s web interface commonly provides access to these citations. It is possible that some of these data citations aren’t directly relevant to the datasets because they may be related to general data science, instrumentation, or data from similar projects.

Publications that refer to data collections may also acknowledge the services that were used to acquire those data, such as subsetting, reformatting, visualizing, or analyzing the data collections. NASA Giovanni (), developed by GES DISC, has been a popular service for data analysis and visualization of NASA Earth datasets since early 2000. GES DISC regularly collects research publications that mention the usage of the NASA Giovanni service by exhaustive keyword searches of Google Scholar. By the beginning of 2023, around 3,000 research publications have been collected. As GES DISC collects the papers that cite the Giovanni service, the publications are reviewed to determine from which data collections the variables were derived so proper attribution can be assigned to the science teams who created these datasets, and to also list these publications on corresponding dataset web pages at the GES DISC user website.

Using our proposed automated methods, we analyzed how many publications can be found and automatically attributed to datasets, compared to how many publications must be collected using general keywords and manually reviewed to make such attribution. Figure 6 shows counts of citations that were found automatically by DOI and keyword search, their overlap, and counts of publications that were found by generic search and which required manual review.

Figure 7 demonstrates that a significant count of publications that use GES DISC data still cannot be automatically harvested from bibliometric sources because these documents do not refer to the datasets by DOIs or ESDT names, while still mentioning the NASA Giovanni service used to analyze data. Nevertheless, we can see that at least a quarter of the publications that use data from the Giovanni service can automatically be attributed to the individual datasets.

Figure 7 

Citation counts for dataset attribution to documents using the Giovanni service.

Figure 8 shows how the proposed method can be used to identify datasets in publications collected by several GES DISC data provider teams: Atmospheric Infrared Sounder (), OMI (), the Global Precipitation Measurement () mission, and the Orbiting Carbon Observatory (). The data in Figure 8 covers statistics for research documents published in 2021. As is seen in Figure 8, the proportion of the documents where datasets can be identified with automated search is different. It may be influenced by several factors, one of them being how long the instrument has been operating and providing the data: older missions (AIRS since 2002 and OMI since 2004) have more datasets identified than newer missions (both GPM and OCO since 2014).

Figure 8 

Fraction of data provider and Giovanni service documents that can be automatically found and attributed to datasets. The publication year for all these documents is 2021.

Automated search makes it possible to find publications in addition to the publications collected by data providers since automated search finds document citations independently from bibliographic sources. Figure 9 presents the proportion of the document citations that are found by the automated search compared to the total documents collected by the data providers and their search. This proportion varies significantly between the data providers and can be explained by both the data provider publication search methods and resources spent on publication search and review. As publication collection can take significant time to search for and review publications, our method — while it does not find the majority of the publications — can still be used as an automated method to find a relatively constant fraction of publications reliably linked to the datasets.

Figure 9 

Fraction of document citations that are found by an automated search in addition to data provider-found citations. The publication year of all these documents is 2021.

GES DISC Document Citation Management

The GES DISC goal is to present collected document citations to archive users in such a way that users would have direct access to all document citations that cite a particular dataset, and the users will have direct access to the datasets when they are presented with the document citation. Several factors complicate this linkage. First, the dataset DOIs are assigned to specific dataset versions. When a dataset is mentioned in the document by its ESDT, it is not always known which version the document authors have used — thus the document cannot be linked to a specific dataset DOI with absolute certainty. Second, older versions of the datasets are retired from the archive, and the DOIs of retired dataset versions are redirected to point to the DOIs of the versions that are distributed to the public. Thus, the publications which used retired versions need to be made available to the users. Third, more than one version of the dataset can be distributed to the users by the archive at the same time, and a newer dataset version usually would have much fewer citations, especially at the beginning of its availability. Thus, if document citations are linked to the specific version of the datasets, users may miss out on the research literature that cited the prior dataset versions. To address these issues, the GES DISC approach is to link all found documents to the dataset ESDT and present users with all of these documents for each version of the dataset on that dataset version’s dataset landing page (DSL), where DSL is the webpage to which the dataset DOI resolves. This approach is presented in Figure 10.

Figure 10 

Linking publications to datasets at GES DISC.

This GES DISC approach to citation linkage lets users see a full history of research literature that used the dataset, regardless of its version. To track previous versions of the dataset, GES DISC implemented a DOI version history that lists all previous DOIs for deprecated dataset versions on the DSL as shown in Figure 11. Documents can be matched with the exact dataset versions used in them when these dataset versions were cited in documents using this DOI history. The version history can help narrow down the versions of datasets that might have been used in documents that didn’t cite exact dataset versions.

Figure 11 

Snapshot of the dataset version history in Dataset Landing Page for the ‘NCALDAS_NOAH0125_D’ dataset ().

Dataset registry DataCite links dataset DOIs only to research documents that cited those DOIs. The DataCite has no infrastructure to link datasets to the literature that cited datasets for versions that did not have DOIs, or when the datasets were mentioned and the version that was used was uncertain.

As datasets progress through subsequent versions, the linkage of the documents to the public dataset versions has to be managed. Once a version is retired, the documents are unlinked from it, and once the new version becomes available, all documents linked to that dataset are linked to that version.

Dataset version management related to document linkage is managed separately from document citation collection. Figure 12 presents the diagram of GES DISC document citation management. The document citations are collected from various sources and by different means, and ingested into the Zotero citation management system. Citations are added to Zotero using the Zotero desktop client by import in either bibliographic formats such as BibTeX or by the citation’s DOI. In addition, the Zotero Web browser plugin, called ‘Zotero connector’, is used to load citations directly from the Web browser into the Zotero client.

Figure 12 

Flow diagram of the GES DISC document citation management system.

The Zotero client is connected to the Zotero server, which provides multiple capabilities; it allows for unlimited document citation storage along with citation tagging. The citation library is backed up at the Zotero server and can be shared between individual clients. The library content on the server can be accessed through its API. GES DISC uses the Python API client library, PyZotero, to manage the library content for export and citation tagging. GES DISC uses Zotero tags to add the following information to document citations: dataset ESDT name, dataset DOI when cited, and the bibliometric sources where document citation was retrieved.

The GES DISC User Interface (UI) back-end system retrieves citation library content from Zotero by means of the PyZotero API. The UI back-end system’s task is to link all collected citations to DSLs of the datasets that are currently distributed to the public. For this, the UI backend maintains a list of public dataset DOIs and corresponding dataset ESDT names. The UI is regularly updated to link document citations using dataset ESDT names they are tagged with to the public dataset DOIs. This linkage is further used by the GES DISC UI system to list linked documents on the DSLs, as shown in Figure 13.

Figure 13 

Snapshot of Dataset Landing Page for ‘OMBRO’ dataset ().

Figure 13 presents the DSL of the ‘OMBRO’ dataset () with the partial content of the ‘References’ tab. There are two types of ‘References’ that are displayed on GES DISC DSL pages: ‘Data Collection References’ and ‘Related Publications’. The former are the citations of the publications that the dataset provider advises the GES DISC to supply. These publications typically describe the measurement instrument, dataset algorithm, and validation. ‘Related Publications’ are the publications that are collected by GES DISC that use the dataset for research or investigative purposes. The disclaimer provided for these citations states ‘The majority of publications using GES DISC data do NOT include the data version id. As such, publications below may not use the most recent processing version of the data.’ Under each publication there are URLs to DSL pages of linked datasets.

At the GES DISC website, the document citations are also provided at the dedicated search interface as shown in Figure 14. As on the DSL pages under each document citation, there are URLs of corresponding dataset DSLs. The document citations on this interface can be refined by the document publication year, type, and journal. Citations also can be sorted by publication year and author names, as well as downloaded in BibTex format that allows for easy export to the user’s own citation management system. The additional capability of this interface is the search of the publications by the dataset and free text keywords. Free text search on the publications interface opens the possibility of data discovery by their application. For example, if a user searches for ‘landslide’, then the search results will list documents that have this keyword appear in their title and/or abstract and which very likely relate to the research related to landslides. Direct linkage of these citations to the datasets allows the user to ‘discover’ datasets that were used for research on landslides.

Figure 14 

Snapshot of GES DISC Publications interface https://disc.gsfc.nasa.gov/information/publications.

Summary

We developed an automated approach to obtain and process large quantities of citations from Google Scholar and to accurately link them with the datasets used in the cited research. To develop this approach, we first analyzed all major bibliographical resources such as Web of Science, Scopus, Crossref, and Google Scholar. We concluded that the latter not only covers the majority of citations that can be found in other sources but also contains citations missing from the rest. Based on this knowledge, we then developed a method of obtaining document citations from Google Scholar using the subscription-based SerpAPI. The major components of our method consist of composing dataset-specific keyword search queries, and subsequent application of open-source Zotero translation software, for converting obtained URLs into proper document citations. These valid citations are then further processed to deduplicate them and retain citations with desired document types. We evaluated the precision of our method for all citations published in 2021 by verifying that those papers indeed referenced or mentioned the datasets. To showcase the comparative efficacy of our method against other available solutions, we implemented a dataset keyword search using OpenAlex. Our findings indicate that OpenAlex stands as a promising tool, adept at facilitating efficient text searches within publications. We illustrated the capacity of our method to enhance the efficiency of collection citations, utilizing a control sample gathered from dataset provider science teams as a benchmark. Our results show that for some of the data provider citation websites, as many as 40% of their citations can be found by our automated approach. Finally, we demonstrate the workflow of how the document-dataset linked citations are collected at the GES DISC and how potential data users benefit from this process.

Computer Code Availability

The Google Scholar Dataset Citing Documents Search () tool is available from GitHub. The document citations collected for the GES DISC archive are maintained in the NASA GES DISC Zotero library ().