https://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/FJAG0Xhttps://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/FJAG0X
Background Data sharing is commonly seen as beneficial for science, but is not yet common practice. Research funding agencies are known to play a key role in promoting data sharing, but German funders’ data sharing policies appear to lag behind in international comparison. This study aims to answer the question of how German data sharing experts inside and outside funding agencies perceive and evaluate German funders’ data sharing policies and overall efforts to promote data sharing. Methods This study is based on sixteen guideline-structured interviews with representatives of German funding agencies and German research data experts from other organisations, who shared their perceptions of German’ funders efforts to promote data sharing. By applying the method of qualitative content analysis to our interview data, we categorise and describe noteworthy aspects of the German data sharing policy landscape and illustrate our findings with interview passages. Research data This dataset contains summaries from interviews with data sharing and funding policy experts from German funding agencies and what we call "stakeholder organisations" (e.g., universities, research data infrastructure providers, etc.). We asked the interviewees about their perspectives on German funders' data sharing policies, for example regarding the actual status quo, their expectations about the potential role that funders can play in promoting data sharing, as well as general developments in this area. Supplement_1_Interview_guideline_funders.pdf and Supplement_2_Interview_guideline_stakeholders.pdf provide supplemental information in the form of the (german) interview guidelines used in this study. Supplement_3_Transcription_and_coding_guideline.pdf lays out the rules we followed in our transcription and coding process. Supplement_4_Category_system.pdf describes the underlying category system of the qualitative content analysis we conducted.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Survey period: 08 April - 08 May, 2014 Top 10 Impact Factor journals in each of 22 categories
Figures https://doi.org/10.6084/m9.figshare.6857273.v1
Article https://doi.org/10.20651/jslis.62.1_20 https://doi.org/10.15068/00158168
Public Domain Mark 1.0https://creativecommons.org/publicdomain/mark/1.0/
License information was derived automatically
This policy applies to SPREP’s own data as well as data held by SPREP on behalf of government agencies and partners within the Pacific.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Over the last decade, there have been significant changes in data sharing policies and in the data sharing environment faced by life science researchers. Using data from a 2013 survey of over 1600 life science researchers, we analyze the effects of sharing policies of funding agencies and journals. We also examine the effects of new sharing infrastructure and tools (i.e., third party repositories and online supplements). We find that recently enacted data sharing policies and new sharing infrastructure and tools have had a sizable effect on encouraging data sharing. In particular, third party repositories and online supplements as well as data sharing requirements of funding agencies, particularly the NIH and the National Human Genome Research Institute, were perceived by scientists to have had a large effect on facilitating data sharing. In addition, we found a high degree of compliance with these new policies, although noncompliance resulted in few formal or informal sanctions. Despite the overall effectiveness of data sharing policies, some significant gaps remain: about one third of grant reviewers placed no weight on data sharing plans in their reviews, and a similar percentage ignored the requirements of material transfer agreements. These patterns suggest that although most of these new policies have been effective, there is still room for policy improvement.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Raw data from the article entitled, "Data sharing policies of journals in life, health, and physical sciences indexed in Journal Citation Reports"
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Data sharing is crucial to the advancement of science because it facilitates collaboration, transparency, reproducibility, criticism, and re-analysis. Publishers are well-positioned to promote sharing of research data by implementing data sharing policies. While there is an increasing trend toward requiring data sharing, not all journals mandate that data be shared at the time of publication. In this study, we extended previous work to analyze the data sharing policies of 447 journals across several scientific disciplines, including biology, clinical sciences, mathematics, physics, and social sciences. Our results showed that only a small percentage of journals require data sharing as a condition of publication, and that this varies across disciplines and Impact Factors. Both Impact Factors and discipline are associated with the presence of a data sharing policy. Our results suggest that journals with higher Impact Factors are more likely to have data sharing policies; use shared data in peer review; require deposit of specific data types into publicly available data banks; and refer to reproducibility as a rationale for sharing data. Biological science journals are more likely than social science and mathematics journals to require data sharing.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains Data Availability Statements from 47,593 papers published in PLOS ONE between March 2014 (when the policy went into effect) and May 2016, analyzed for type of statement.
This dataset collects the slides that were presented at the Data Collaborations Across Boundaries session in SciDataCon 2022, part of the International Data Week.
The following session proposal was prepared by Tyng-Ruey Chuang and submitted to SciDataCon 2022 organizers for consideration on 2022-02-28. The proposal was accepted on 2022-03-28. Six abstracts were submitted and accepted to this session. Five presentations were delivered online in a virtual session on 2022-06-21.
Data Collaborations Across Boundaries
There are many good stories about data collaborations across boundaries. We need more. We also need to share the lessons each of us has learned from collaborating with parties and communities not in our familiar circles.
By boundaries, we mean not just the regulatory borders in between the nation states about data sharing but the various barriers, readily conceivable or not, that hinder collaboration in aggregating, sharing, and reusing data for social good. These barriers to collaboration exist between the academic disciplines, between the economic players, and between the many user communities, just to name a few. There are also cross-domain barriers, for example those that lay among data practitioners, public administrators, and policy makers when they are articulating the why, what, and how of "open data" and debating its economic significance and fair distribution. This session aims to bring together experiences and thoughts on good data practices in facilitating collaborations across boundaries and domains.
The success of Wikipedia proves that collaborative content production and service, by ways of copyleft licenses, can be sustainable when coordinated by a non-profit and funded by the general public. Collaborative code repositories like GitHub and GitLab demonstrate the enormous value and mass scale of systems-facilitated integration of user contributions that run across multiple programming languages and developer communities. Research data aggregators and repositories such as GBIF, GISAID, and Zenodo have served numerous researchers across academic disciplines. Citizen science projects and platforms, for instance eBird, Galaxy Zoo, and Taiwan Roadkill Observation Network (TaiRON), not only collect data from diverse communities but also manage and release datasets for research use and public benefit (e.g. TaiRON datasets being used to improve road design and reduce animal mortality). At the same time large scale data collaborations depend on standards, protocols, and tools for building registries (e.g. Archival Resource Key), ontologies (e.g. Wikidata and schema.org), repositories (e.g. CKAN and Omeka), and computing services (e.g. Jupyter Notebook). There are many types of data collaborations. The above lists only a few.
This session proposal calls for contributions to bring forward lessons learned from collaborative data projects and platforms, especially about those that involve multiple communities and/or across organizational boundaries. Presentations focusing on the following (non-exclusive) topics are sought after:
Support mechanisms and governance structures for data collaborations across organizations/communities.
Data policies --- such as data sharing agreements, memorandum of understanding, terms of use, privacy policies, etc. --- for facilitating collaborations across organizations/communities.
Traditional and non-traditional funding sources for data collaborations across multiple parties; sustainability of data collaboration projects, platforms, and communities.
Data workflows --- collection, processing, aggregation, archiving, and publishing, etc. --- designed with considerations of (external) collaboration.
Collaborative web platforms for data acquisition, curation, analysis, visualization, and education.
Examples and insights from data trusts, data coops, as well as other formal and informal forms of data stewardship.
Debates on the pros and cons of centralized, distributed, and/or federated data services.
Practical lessons learned from data collaboration stories: failure, success, incidence, unexpected turn of event, aftermath, etc. (no story is too small!).
The journals’ author guidelines and/or editorial policies were examined on whether they take a stance with regard to the availability of the underlying data of the submitted article. The mere explicated possibility of providing supplementary material along with the submitted article was not considered as a research data policy in the present study. Furthermore, the present article excluded source codes or algorithms from the scope of the paper and thus policies related to them are not included in the analysis of the present article.
For selection of journals within the field of neurosciences, Clarivate Analytics’ InCites Journal Citation Reports database was searched using categories of neurosciences and neuroimaging. From the results, journals with the 40 highest Impact Factor (for the year 2017) indicators were extracted for scrutiny of research data policies. Respectively, the selection journals within the field of physics was created by performing a similar search with the categories of physics, applied; physics, atomic, molecular & chemical; physics, condensed matter; physics, fluids & plasmas; physics, mathematical; physics, multidisciplinary; physics, nuclear and physics, particles & fields. From the results, journals with the 40 highest Impact Factor indicators were again extracted for scrutiny. Similarly, the 40 journals representing the field of operations research were extracted by using the search category of operations research and management.
Journal-specific data policies were sought from journal specific websites providing journal specific author guidelines or editorial policies. Within the present study, the examination of journal data policies was done in May 2019. The primary data source was journal-specific author guidelines. If journal guidelines explicitly linked to the publisher’s general policy with regard to research data, these were used in the analyses of the present article. If journal-specific research data policy, or lack of, was inconsistent with the publisher’s general policies, the journal-specific policies and guidelines were prioritized and used in the present article’s data. If journals’ author guidelines were not openly available online due to, e.g., accepting submissions on an invite-only basis, the journal was not included in the data of the present article. Also journals that exclusively publish review articles were excluded and replaced with the journal having the next highest Impact Factor indicator so that each set representing the three field of sciences consisted of 40 journals. The final data thus consisted of 120 journals in total.
‘Public deposition’ refers to a scenario where researcher deposits data to a public repository and thus gives the administrative role of the data to the receiving repository. ‘Scientific sharing’ refers to a scenario where researcher administers his or her data locally and by request provides it to interested reader. Note that none of the journals examined in the present article required that all data types underlying a submitted work should be deposited into a public data repositories. However, some journals required public deposition of data of specific types. Within the journal research data policies examined in the present article, these data types are well presented by the Springer Nature policy on “Availability of data, materials, code and protocols” (Springer Nature, 2018), that is, DNA and RNA data; protein sequences and DNA and RNA sequencing data; genetic polymorphisms data; linked phenotype and genotype data; gene expression microarray data; proteomics data; macromolecular structures and crystallographic data for small molecules. Furthermore, the registration of clinical trials in a public repository was also considered as a data type in this study. The term specific data types used in the custom coding framework of the present study thus refers to both life sciences data and public registration of clinical trials. These data types have community-endorsed public repositories where deposition was most often mandated within the journals’ research data policies.
The term ‘location’ refers to whether the journal’s data policy provides suggestions or requirements for the repositories or services used to share the underlying data of the submitted works. A mere general reference to ‘public repositories’ was not considered a location suggestion, but only references to individual repositories and services. The category of ‘immediate release of data’ examines whether the journals’ research data policy addresses the timing of publication of the underlying data of submitted works. Note that even though the journals may only encourage public deposition of the data, the editorial processes could be set up so that it leads to either publication of the research data or the research data metadata in conjunction to publishing of the submitted work.
This dataset contains anonymized transcripts of interviews performed during a qualitative interview study with members of 16 research funding agencies. Sample: Funding agencies were located mainly, but not exclusively, in Europe. Funding agencies were selected using a purposive sampling strategy that aimed to acquire sufficient representation of (a) national and international agencies; (b) public and philanthropic agencies; and (c) agencies in continental European and the anglophone world. Suitable funding agencies were selected through various means, such as a list of health research funding organizations according to their annual expenditure on health research (https://www.healthresearchfunders.org/health-research-funding-organizations/) and Science Europe Working Groups. Background: Open science policy documents have emphasized the need to install more incentives for data sharing. These incentives are often understood as being reputational or financial. Additionally, there are other policy measures that could be taken, such as data sharing mandates. Aim: To document views of funding agencies on (1) potential alterations to recognition systems in academia; (2) incentives to enhance the sharing of cohort data; (3) data sharing policies in terms of the governance of cohort data; (4) other potential interactions between science policy and data sharing platforms for cohorts. Our study focused on the sharing of patient- and population-based cohorts through data infrastructures.
The public sharing of primary research datasets potentially benefits the research community but is not yet common practice. In this pilot study, we analyzed whether data sharing frequency was associated with funder and publisher requirements, journal impact factor, or investigator experience and impact. Across 397 recent biomedical microarray studies, we found investigators were more likely to publicly share their raw dataset when their study was published in a high-impact journal and when the first or last authors had high levels of career experience and impact. We estimate the USA's National Institutes of Health (NIH) data sharing policy applied to 19% of the studies in our cohort; being subject to the NIH data sharing plan requirement was not found to correlate with increased data sharing behavior in multivariate logistic regression analysis. Studies published in journals that required a database submission accession number as a condition of publication were more likely to share their data, but this trend was not statistically significant. These early results will inform our ongoing larger analysis, and hopefully contribute to the development of more effective data sharing initiatives. Earlier version presented at ASIS&T and ISSI Pre-Conference: Symposium on Informetrics and Scientometrics 2009
Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically
A manually curated registry of data policies from research funders, journal publishers, societies, and other organisations. These are linked to the databases and standards that they recommend for use
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Examples of data sharing policies.
https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=hdl:1902.29/11582https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=hdl:1902.29/11582
The rapidly growing number of digital research data and developing scientific methods has challenged science policy makers and research funders to seek tools for increasing openness and sharing of research data. In some countries, academic research funders have adopted formal data policies to guide data management and data sharing in new research projects. The International Federation of Data Organizations for Social Science (IFDO) conducted an international web-based survey of data professionals to gather information on the institutional policies of social sciences and humanities in different countries and to collect and locate sources for a more detailed analysis. This dataset contains additional responses collected beyond the Wave I data collection period.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
It is the survey data on the data sharing policy to editors in Korea.
Use this guide to find information on Tempe data policy and standards.Open Data PolicyEthical Artificial Intelligence (AI) PolicyEvaluation PolicyExpedited Data Sharing PolicyData Sharing Agreement (General)Data Sharing Agreement (GIS)Data Quality Standard and ChecklistDisaggregated Data StandardsData and Analytics Service Standard
A survey conducted in December 2021 in the United States found that ** percent of U.S. consumers did not think they had any control over preventing companies from sharing their personal information with third parties. Another ** percent said they didn't believe they could prevent the companies from collecting information about them.
https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/2.2/customlicense?persistentId=doi:10.15139/S3/12157https://dataverse-staging.rdmc.unc.edu/api/datasets/:persistentId/versions/2.2/customlicense?persistentId=doi:10.15139/S3/12157
This study consists of data files that code the data availability policies of top-20 academic journals in the fields of Business & Finance, Economics, International Relations, Political Science, and Sociology. Journals that were ranked as top-20 titles based on 2003-vintage ISI Impact Factor scores were coded on their data policies in 2003 and on their data policies in 2015. In addition, journals that were ranked as top-20 titles based on most recent ISI Impact Factor scores were likewise coded on their data polices in 2015. The included Stata .do file imports the contents of each of the Excel files, cleans and labels the data, and produces two tables: one comparing the data policies of 2003-vintage top-20 journals in 2003 those journals' policies in 2015, and one comparing the data policies of 2003-vintage top-20 journals in 2003 to the data policies of current top-20 journals in 2015.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
This dataset contains survey responses from 3,257 researchers in 20 different disciplines. The survey was conducted in 2020 as part of a PhD research project and explored data generation, sharing and reuse practices across these disciplines.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
The data underlying scientific papers should be accessible to researchers both now and in the future, but how best can we ensure that these data are available? Here we examine the effectiveness of four approaches to data archiving: no stated archiving policy, recommending (but not requiring) archiving, and two versions of mandating data deposition at acceptance. We control for differences between data types by trying to obtain data from papers that use a single, widespread population genetic analysis, STRUCTURE. At one extreme, we found that mandated data archiving policies that require the inclusion of a data availability statement in the manuscript improve the odds of finding the data online almost 1000-fold compared to having no policy. However, archiving rates at journals with less stringent policies were only very slightly higher than those with no policy at all. We also assessed the effectiveness of asking for data directly from authors and obtained over half of the requested datasets, albeit with ∼8 d delay and some disagreement with authors. Given the long-term benefits of data accessibility to the academic community, we believe that journal-based mandatory data archiving policies and mandatory data availability statements should be more widely adopted.
https://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/FJAG0Xhttps://heidata.uni-heidelberg.de/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.11588/DATA/FJAG0X
Background Data sharing is commonly seen as beneficial for science, but is not yet common practice. Research funding agencies are known to play a key role in promoting data sharing, but German funders’ data sharing policies appear to lag behind in international comparison. This study aims to answer the question of how German data sharing experts inside and outside funding agencies perceive and evaluate German funders’ data sharing policies and overall efforts to promote data sharing. Methods This study is based on sixteen guideline-structured interviews with representatives of German funding agencies and German research data experts from other organisations, who shared their perceptions of German’ funders efforts to promote data sharing. By applying the method of qualitative content analysis to our interview data, we categorise and describe noteworthy aspects of the German data sharing policy landscape and illustrate our findings with interview passages. Research data This dataset contains summaries from interviews with data sharing and funding policy experts from German funding agencies and what we call "stakeholder organisations" (e.g., universities, research data infrastructure providers, etc.). We asked the interviewees about their perspectives on German funders' data sharing policies, for example regarding the actual status quo, their expectations about the potential role that funders can play in promoting data sharing, as well as general developments in this area. Supplement_1_Interview_guideline_funders.pdf and Supplement_2_Interview_guideline_stakeholders.pdf provide supplemental information in the form of the (german) interview guidelines used in this study. Supplement_3_Transcription_and_coding_guideline.pdf lays out the rules we followed in our transcription and coding process. Supplement_4_Category_system.pdf describes the underlying category system of the qualitative content analysis we conducted.