Mr-Vicky-01/Repository-Dataset dataset hosted on Hugging Face and contributed by the HF Datasets community
The NSF Public Access Repository contains an initial collection of journal publications and the final accepted version of the peer-reviewed manuscript or the version of record. To do this, NSF draws upon services provided by the publisher community including the Clearinghouse of Open Research for the United States, CrossRef, and International Standard Serial Number. When clicking on a Digital Object Identifier number, you will be taken to an external site maintained by the publisher. Some full text articles may not be available without a charge during the embargo, or administrative interval. Some links on this page may take you to non-federal websites. Their policies may differ from this website.
Data collected from major Canadian and international research data repositories cover data storage, preservation, metadata, interchange, data file types, and other standard features used in the retention and sharing of research data. The outputs of this project primarily aim to assist in the establishment of recommended minimum requirements for a Canadian research data infrastructure. The committee also aims to further develop guidelines and criteria for the assessment and selection o f repositories for deposit of Canadian research data by researchers, data managers, librarians, archivists etc.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In the age of digital transformation, scientific and social interest for data and data products is constantly on the rise. The quantity as well as the variety of digital research data is increasing significantly. This raises the question about the governance of this data. For example, how to store the data so that it is presented transparently, freely accessible and subsequently available for re-use in the context of good scientific practice. Research data repositories provide solutions to these issues.
Considering the variety of repository software, it is sometimes difficult to identify a fitting solution for a specific use case. For this purpose a detailed analysis of existing software is needed. Presented table of requirements can serve as a starting point and decision-making guide for choosing the most suitable for your purposes repository software. This table is dealing as a supplementary material for the paper "How to choose a research data repository software? Experience report." (persistent identifier to the paper will be added as soon as paper is published).
Overview This directory was developed to provide discovery information for anyone looking for publicly accessible repositories that house geological materials in the U.S. and Canada. In addition, this resource is intended to be a tool to facilitate a community of practice. The need for the directory was identified during planning for and follow-up from a drill core repository webinar series in Spring 2020 for public repository curators and staff in the U.S. and Canada hosted by the Minnesota Geological Survey and the Minnesota Department of Natural Resources. Additional supporting sponsors included the U.S. Geological Survey National Geological and Geophysical Data Preservation Program and the Association of American State Geologists Data Preservation Committee. The 10-part webinar series provided overviews of state, provincial, territorial, and national repositories that house drill core, other geoscience materials, and data. When the series concluded a small working group of the participants continued to meet to facilitate the development and production of a directory of repositories that maintain publicly-accessible geological materials throughout the U.S. and Canada. The group used previous directory efforts described in the next section, Summary of Historical Repository Directory Compilation Efforts, as guides for content during development. The working group prepared and compiled responses from a call for repository information and characterization. This directory is planned to be a living resource for the geoscience community with updates every other year to accommodate changes. The updates will facilitated through versioned updates of this data release. Summary of Historical Repository Directory Compilation Efforts 1957 – Sample and Core Repositories of the United States, Alaska, and Canada. Published by AAPG. Committee on Preservation of Samples and Cores. 13 members from industry, academia, and government. 1977 – Well-Sample and Core Repositories of the Unites States and Canada, C.K. Fisher; M.P. Krupa, USGS Open file report 77-567.USGS wanted to update the original index. Includes a map showing core repositories by “State” “University” “Commercial” and “Federal”. Also includes a “Brief Statement of Requirements for the Preservation of Subsurface Material and Data” and referral to state regulations for details on preserved materials. 1984 - Nonprofit Sample and Core Repositories Open to the Public in the United States – USGS Circular 942. James Schmoker, Thomas Michalski, Patricia Worl. The survey was conducted by a questionnaire mailed to repository curators. Information on additions, corrections, and deletions to earlier (1957,1977) directories from state geologists, each state office of the Water Resources Division of the U.S. Geological Survey, additional government agencies and colleagues were also used. 1997 - The National Directory of Geoscience Data Repositories, edited by Nicholas H. Claudy – American Geologic Institute. To prepare the directory, questionnaires were mailed to state geologists, more than 60 geological societies, private-sector data centers selected from oil and gas directories, and to the membership committee of the American Association of Petroleum Geologists, one of AGI's member societies. The directory contains 124 repository listings, organized alphabetically by state. 2002 – National Research Council 2002. Geoscience Data and Collections: National resources in Peril. Washington, D.C.: The National Academies Press 2005 – The National Geological and Geophysical Data Preservation Program (NGGDPP) of the United States Geological Survey (USGS) was established by The Energy Policy Act of 2005, and reauthorized in the Consolidated Appropriations Act, 2021, “to preserve and expose the Nation’s geoscience collections (samples, logs, maps, data) to promote their discovery and use for research and resource development”. The Program provides “technical and financial assistance to state geological surveys and U.S. Department of the Interior (DOI) bureaus” to archive “geological, geophysical, and engineering data, maps, photographs, samples, and other physical specimens”. Metadata records describing the preserved assets are cataloged in the National Digital Catalog (NDC). References American Association of Petroleum Geologists, 1957, Sample and core repositories of the United States, Alaska, and Canada: American Association of Petroleum Geologists, Committee on Preservation of Samples and Cores, 29 p. American Association of Petroleum Geologists, 2018, US Geological Sample and Data Repositories: American Association of Petroleum Geologists, Preservation of Geoscience Data Committee, Unpublished, (Contact: AAPG Preservation of Geoscience Data Committee) American Geological Institute, 1997, National Geoscience Data Repository System, Phase II. Final report, January 30, 1995--January 28, 1997. United States. https://doi.org/10.2172/598388 American Geological Institute, 1997, National Directory of Geoscience Data Repositories, Claudy, N. H., (ed.), 91pp. Claudy N., Stevens D., 1997, AGI Publishes first edition of national directory of geoscience data repositories. American Geological Institute Spotlight, https://www.agiweb.org/news/datarep2.html Consolidated Appropriations Act, 2021 (Public Law 116-260, Sec.7002) Davidson, E. D., Jr., 1981, A look at core and sample libraries: Bureau of Economic Geology, The University of Texas at Austin, 4 p. and Appendix. Deep Carbon Observatory (DCO) Data Portal, Scientific Collections, https://info.deepcarbon.net/vivo/scientific-collections; Keyword Search: sample repository, https://info.deepcarbon.net/vivo/scientific-collections?source=%7B%22query%22%3A%7B%22query_string%22%3A%7B%22query%22%3A%22sample%20repository%20%22%2C%22default_operator%22%3A%22OR%22%7D%7D%2C%22sort%22%3A%5B%7B%22_score%22%3A%7B%22order%22%3A%22asc%22%7D%7D%5D%2C%22from%22%3A0%2C%22size%22%3A200%7D: Accessed September 29, 2021 Fisher, C. K., and Krupa, M. P., 1977, Well-sample and core repositories of the United States and Canada: U.S. Geological Survey Open-File Report 77-567, 73 p. https://doi.org/10.3133/ofr77567 Fogwill, W.D., 1985, Drill Core Collection and Storage Systems in Canada, Manitoba Energy & Mines. https://www.ngsc-cptgs.com/files/PGJSpecialReport_1985_V03b.pdf Goff, S., and Heiken, G., eds., 1982, Workshop on core and sample curation for the National Continental Scientific Drilling Program: Los Alamos National Laboratory, May 5-6, 1981, LA-9308-C, 31 p. https://www.osti.gov/servlets/purl/5235532 Lonsdale, J. T., 1953, On the preservation of well samples and cores: Oklahoma City Geological Society Shale Shaker, v. 3, no. 7, p. 4. National Geological and Geophysical Data Preservation Program. https://www.usgs.gov/core-science-systems/national-geological-and-geophysical-data-preservation-program National Research Council. 2002. Geoscience Data and Collections: National Resources in Peril. Washington, DC: The National Academies Press, 107 p. https://doi.org/10.17226/10348 Pow, J. R., 1969, Core and sample storage in western Canada: Bulletin of Canadian Petroleum Geology, v. 17, no. 4, p. 362-369. DOI: 10.35767/gscpgbull.17.4.362 Ramdeen, S., 2015. Preservation challenges for geological data at state geological surveys, GeoResJ 6 (2015) 213-220, https://doi.org/10.1016/j.grj.2015.04.002 Schmoker, J. W., Michalski, T. C., and Worl, P. B., 1984, Nonprofit sample and core repositories of the United States: U.S. Geological Survey Circular 942. https://doi.org/10.3133/cir942 Schmoker, J. W., Michalski, T. C., and Worl, P. B., 1984, Addresses, telephone numbers, and brief descriptions of publicly available, nonprofit sample and core repositories of the United States: U.S. Geological Survey Open-File Report 84-333, 13 p. (Superseded by USGS Circular 942) https://doi.org/10.3133/ofr84333 The Energy Policy Act of 2005 (Public Law 109-58, Sec. 351) The National Digital Catalog (NDC). https://www.usgs.gov/core-science-systems/national-geological-and-geophysical-data-preservation-program/national-digital U.S. Bureau of Mines, 1978, CORES Operations Manual: Bureau of Mines Core Repository System: U.S. Bureau of Mines Information Circular IC 8784, 118 p. https://digital.library.unt.edu/ark:/67531/metadc170848/
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
In this work, based on GitHub Archive project and repository mining tools, we process all available data into concise and structured format to generate GitHub developer behavior and repository evolution dataset. With the self-configurable interactive analysis tool provided by us, it will give us a macroscopic view of open source ecosystem evolution.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Bear Lake Data Repository (BLDR) is an active archive, containing a growing compilation of biological, chemical, and physical datasets collected from Bear Lake and its surrounding watershed. The datasets herein have been digitized from historical records and reports, extracted from papers and theses, and obtained from public and private entities, including the United States Geological Survey, PacifiCorp, and, inter alia, Ecosystems Research Institute.
Contributions are welcome. The BLDR accepts biological, chemical, or physical datasets obtained at Bear Lake, irrespective of funding source. There is no submission size limit at present—workarounds will be found if submissions exceed Hydroshare limits (20 GB). Contributions are published with an open access license and will serve many use cases. The current repository steward, Bear Lake Watch, will advise on submissions and make accepted contributions available promptly.
Metadata files are provided for each dataset, however, contact with original contributor(s) is encouraged for questions and additional details prior to data usage. The BLDR and its contributors shall not be liable for any damages resulting from misinterpretation or misuse of the data or metadata.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
and how these can be affected by the presence of human body nearby.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains 189 survey responses from a respository/ data managers' survey where we explored the current status, needs and challenges of research data repositories.
The Administrative Data Repository (ADR) was established to provide support for the administrative data elements relative to multiple categories of a person entity such as demographic and eligibility information. Although initially focused on the computing needs of the Veterans Health Administration, the ADR is positioned to provide identity management and demographics support for all IT systems within the Department of Veterans Affairs.
To address the increasing complexity of network management and the limitations of data repositories in handling the various network operational data, this paper proposes a novel repository design that uniformly represents network operational data while allowing for a multiple abstractions access to the information. This smart repository simplifies network management functions by enabling network verification directly within the repository. The data is organized in a knowledge graph compatible with any general-purpose graph database, offering a comprehensive and extensible network repository. Performance evaluations confirm the feasibility of the proposed design. The repository's ability to natively support 'what-if' scenario evaluation is demonstrated by verifying Border Gateway Protocol (BGP) route policies and analyzing forwarding behavior with virtual Traceroute.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
This dataset was extracted from a set of metadata files harvested from the DataCite metadata store (http://search.datacite.org/ui) during December 2015. Metadata records for items with a resourceType of dataset were collected. 1,647,949 total records were collected. This dataset contains four files: 1) readme.txt: a readme file. 2) language-results.csv: A CSV file containing three columns: DOI, DOI prefix, and language text contents 3) language-counts.csv: A CSV file containing counts for unique language text content values. 4) language-grouped-counts.txt: A text file containing the results of manually grouping these language codes.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains article metadata and information about Open Science Indicators for approximately 139,000 research articles published in PLOS journals from 1 January 2018 to 30 March 2025 and a set of approximately 28,000 comparator articles published in non-PLOS journals. This is the tenth release of this dataset, which will be updated with new versions on an annual basis.This version of the Open Science Indicators dataset shares the indicators seen in the previous versions as well as fully operationalised protocols and study registration indicators, which were previously only shared in preliminary forms. The v10 dataset focuses on detection of five Open Science practices by analysing the XML of published research articles:Sharing of research data, in particular data shared in data repositoriesSharing of codePosting of preprintsSharing of protocolsSharing of study registrationsThe dataset provides data and code generation and sharing rates, the location of shared data and code (whether in Supporting Information or in an online repository). It also provides preprint, protocol and study registration sharing rates as well as details of the shared output, such as publication date, URL/DOI/Registration Identifier and platform used. Additional data fields are also provided for each article analysed. This release has been run using an updated preprint detection method (see OSI-Methods-Statement_v10_Jul25.pdf for details). Further information on the methods used to collect and analyse the data can be found in Documentation.Further information on the principles and requirements for developing Open Science Indicators is available in https://doi.org/10.6084/m9.figshare.21640889.Data folders/filesData Files folderThis folder contains the main OSI dataset files PLOS-Dataset_v10_Jul25.csv and Comparator-Dataset_v10_Jul25.csv, which containdescriptive metadata, e.g. article title, publication data, author countries, is taken from the article .xml filesadditional information around the Open Science Indicators derived algorithmicallyand the OSI-Summary-statistics_v10_Jul25.xlsx file contains the summary data for both PLOS-Dataset_v10_Jul25.csv and Comparator-Dataset_v10_Jul25.csv.Documentation folderThis file contains documentation related to the main data files. The file OSI-Methods-Statement_v10_Jul25.pdf describes the methods underlying the data collection and analysis. OSI-Column-Descriptions_v10_Jul25.pdf describes the fields used in PLOS-Dataset_v10_Jul25.csv and Comparator-Dataset_v10_Jul25.csv. OSI-Repository-List_v1_Dec22.xlsx lists the repositories and their characteristics used to identify specific repositories in the PLOS-Dataset_v10_Jul25.csv and Comparator-Dataset_v10_Jul25.csv repository fields.The folder also contains documentation originally shared alongside the preliminary versions of the protocols and study registration indicators in order to give fuller details of their detection methods.Contact details for further information:Iain Hrynaszkiewicz, Director, Open Research Solutions, PLOS, ihrynaszkiewicz@plos.org / plos@plos.orgLauren Cadwallader, Open Research Manager, PLOS, lcadwallader@plos.org / plos@plos.orgAcknowledgements:Thanks to Allegra Pearce, Tim Vines, Asura Enkhbayar, Scott Kerr and parth sarin of DataSeer for contributing to data acquisition and supporting information.
This repository contains Privacy Impact Assessments (PIA) that have been vetted/approved. Section 208 of the Electronic Government Act of 2002 (E-Gov Act) requires federal agencies to conduct a Privacy Impact Assessment (PIA) on information technology systems that either (1) collect, maintain, and/or disseminate Personally Identifiable Information (PII); or (2) make substantial changes to existing technology that manages PII.. A PIA is an analysis of how PII is collected, stored, protected, shared, and managed. Its purpose is to demonstrate that the information system owners and/or developers have incorporated privacy protections throughout the lifecycle of the system. Absent a legitimate security or sensitivity concern demonstrated by the agency, the E-Gov Act requires agencies to make PIAs publicly available.The Department of Veterans Affairs (VA) Office of Privacy, FOIA, and Records Management strive to be the leader among Federal Information Technology (IT) organizations in protecting the privacy of Veterans' personal data. Our goal is to respond to mandates to meet the needs of Veterans, while improving the protection of their personal information.To view a PIA for a particular system, go to the heading labeled 'System Operating with a Current/Valid FY2013 PIA' and select the name of the program/project of interest.
BRADS is a repository for data and biospecimens from population health research initiatives and clinical or interventional trials designed and implemented by NICHD’s Division of Intramural Population Health Research (DIPHR). Topics include human reproduction and development, pregnancy, child health and development, and women’s health. The website is maintained by DIPHR.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Our dataset "repository_survey" summarizes a comprehensive survey of over 150 data repositories, characterizing their metadata documentation and standardization, data curation and validation, and tracking of dataset use in the literature. In addition, "survey_model_evaluation" includes our findings on model evaluation for four methodological repositories. The data are associated with our paper "Towards an Ideological Data Repository: Lessons and Recommendations," where individual columns and survey methods are described.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
A facet is a metadata element, usually from a controlled list, that provides counts of records in a query result with particular values for the metadata element. The DataCite JSON Response includes data on a variety of facets for each query done using the DataCite API.
DataCite Commons uses facets on repository pages to provide an overview of repositories. For example, the Metadata Game Changers Commons page shows publication year, work types, licenses, creators and contributors, and some other facets as graphics and lists.
The facets provided by DataCite can be used to 1) understand characteristics of DataCite metadata, 2) understand some aspects of repository completeness, and 3) provide overviews of repositories.
DataCite includes facets and facet values in all query results, so they are a useful tool for answering some "big picture" questions about DataCite metadata. Some of these questions were explored during 2022 in DataCite Facets: Understanding DataCite Usage using a tool called DataCite Facets.
DataCite facets can be used to provide overviews of any DataCite Repository and understand some characteristics of the repositories. They can also be used, in some cases, to provide insights into some aspects of repository completeness.
Many useful repository measures focus on completeness of the metadata, i.e., the portion of records in the repository that include some metadata element. The DataCite facet data can provide some insight into completeness, but we must keep in mind that the facet data are limited to top ten values for most facets (except for published and resourceTypes, which can be > 10). The blog DataCite Facets and Metadata Completeness describes how some facets can be used to provide insights into metadata completeness.
This dataset provides selected facets downloaded using the DataCite API and associated statistics as a comma-separated-value (CSV) file.
Column definitions:
The dataset includes a number of columns for the selected facets:
Statistic |
Description |
number |
The number of facet values |
max |
The number of occurrences of the most common facet value |
common |
The most common facet value |
total |
The total number of records in the top 10, i.e. the total listed in the facets |
homogeneity (HI) |
An indicator of homogeneity of the facet: maximum count / total count (0.1 = uniform, 1.0 = single item) |
coverage |
The % of all records covered by the top 10 (numbers close to 100% are good) |
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data representing intermediate and final results of the systematic mapping study entitled "Warm-Starting and Quantum Computing: A Systematic Mapping Study". The data is prepared in a folder structure. It is recommended to change the files view to "Tree" to get a better overview of the files. This comprises (file prefix: 01) the exact search queries executed in the database search phase and details on the execution of these queries, (02) the search results of the database search, (03) the list of included publications from the database search, (04) the list of included publications from the snowballing activity, (05) the consolidated final list of included publications, (06) the information extracted from these publications, including the properties of identified techniques, (07) the analysis of these techniques based on their properties and the resulting categorization.
The NIH Common Data Elements (CDE) Repository has been designed to provide access to structured human and machine-readable definitions of data elements that have been recommended or required by NIH Institutes and Centers and other organizations for use in research and for other purposes. Visit the NIH CDE Resource Portal for contextual information about the repository.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This toolkit is intended to assist cultural heritage organizations in choosing materials to send to distributed digital preservation (DDP) systems, networks in geographically-dispersed locations designed to perform preservation actions.
Mr-Vicky-01/Repository-Dataset dataset hosted on Hugging Face and contributed by the HF Datasets community