100+ datasets found
  1. o

    Long Term Development Statement (SPEN_002) Data Quality Checks

    • spenergynetworks.opendatasoft.com
    Updated Mar 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Long Term Development Statement (SPEN_002) Data Quality Checks [Dataset]. https://spenergynetworks.opendatasoft.com/explore/dataset/spen_data_quality_ltds/
    Explore at:
    Dataset updated
    Mar 28, 2025
    Description

    This data table provides the detailed data quality assessment scores for the Long Term Development Statement dataset. The quality assessment was carried out on 31st March. At SPEN, we are dedicated to sharing high-quality data with our stakeholders and being transparent about its' quality. This is why we openly share the results of our data quality assessments. We collaborate closely with Data Owners to address any identified issues and enhance our overall data quality; to demonstrate our progress we conduct annual assessments of our data quality in line with the dataset refresh rate. To learn more about our approach to how we assess data quality, visit Data Quality - SP Energy Networks.We welcome feedback and questions from our stakeholders regarding this process. Our Open Data Team is available to answer any enquiries or receive feedback on the assessments. You can contact them via our Open Data mailbox at opendata@spenergynetworks.co.uk.The first phase of our comprehensive data quality assessment measures the quality of our datasets across three dimensions. Please refer to the data table schema for the definitions of these dimensions. We are now in the process of expanding our quality assessments to include additional dimensions to provide a more comprehensive evaluation and will update the data tables with the results when available.DisclaimerThe data quality assessment may not represent the quality of the current dataset that is published on the Open Data Portal. Please check the date of the latest quality assessment and compare to the 'Modified' date of the corresponding dataset. The data quality assessments will be updated on either a quarterly or annual basis, dependent on the update frequency of the dataset. This information can be found in the dataset metadata, within the Information tab. If you require a more up to date quality assessment, please contact the Open Data Team at opendata@spenergynetworks.co.uk and a member of the team will be in contact.

  2. o

    Single Digital View (SPEN_020) Data Quality Checks

    • spenergynetworks.opendatasoft.com
    Updated Mar 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Single Digital View (SPEN_020) Data Quality Checks [Dataset]. https://spenergynetworks.opendatasoft.com/explore/dataset/spen_data_quality_single_digital_view/
    Explore at:
    Dataset updated
    Mar 28, 2025
    Description

    This data table provides the detailed data quality assessment scores for the Single Digital View dataset. The quality assessment was carried out on the 31st of March. At SPEN, we are dedicated to sharing high-quality data with our stakeholders and being transparent about its' quality. This is why we openly share the results of our data quality assessments. We collaborate closely with Data Owners to address any identified issues and enhance our overall data quality. To demonstrate our progress we conduct, at a minimum, bi-annual assessments of our data quality - for datasets that are refreshed more frequently than this, please note that the quality assessment may be based on an earlier version of the dataset. To learn more about our approach to how we assess data quality, visit Data Quality - SP Energy Networks.We welcome feedback and questions from our stakeholders regarding this process. Our Open Data Team is available to answer any enquiries or receive feedback on the assessments. You can contact them via our Open Data mailbox at opendata@spenergynetworks.co.uk.The first phase of our comprehensive data quality assessment measures the quality of our datasets across three dimensions. Please refer to the data table schema for the definitions of these dimensions. We are now in the process of expanding our quality assessments to include additional dimensions to provide a more comprehensive evaluation and will update the data tables with the results when available.DisclaimerThe data quality assessment may not represent the quality of the current dataset that is published on the Open Data Portal. Please check the date of the latest quality assessment and compare to the 'Modified' date of the corresponding dataset. The data quality assessments will be updated on either a quarterly or annual basis, dependent on the update frequency of the dataset. This information can be found in the dataset metadata, within the Information tab. If you require a more up to date quality assessment, please contact the Open Data Team at opendata@spenergynetworks.co.uk and a member of the team will be in contact.

  3. A&I - Data Quality

    • catalog.data.gov
    • data.transportation.gov
    • +5more
    Updated Jul 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Federal Motor Carrier Safety Administration (2024). A&I - Data Quality [Dataset]. https://catalog.data.gov/dataset/ai-data-quality
    Explore at:
    Dataset updated
    Jul 30, 2024
    Dataset provided by
    Federal Motor Carrier Safety Administrationhttps://www.fmcsa.dot.gov/
    Description

    Data Quality identifies FMCSA resources for evaluating, monitoring, and improving the quality of data submitted by States to the Motor Carrier Management Information System (MCMIS).

  4. KPMG Virtual Internship - Data Quality Assessment

    • kaggle.com
    zip
    Updated Jan 15, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Isa Zeynalov (2023). KPMG Virtual Internship - Data Quality Assessment [Dataset]. https://www.kaggle.com/datasets/isazeynalov/kpmg-virtual-internship-data-quality-assessment
    Explore at:
    zip(2604276 bytes)Available download formats
    Dataset updated
    Jan 15, 2023
    Authors
    Isa Zeynalov
    Description

    Dataset

    This dataset was created by Isa Zeynalov

    Contents

  5. Linking Data for Mothers and Babies in De-Identified Electronic Health Data

    • plos.figshare.com
    docx
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Katie Harron; Ruth Gilbert; David Cromwell; Jan van der Meulen (2023). Linking Data for Mothers and Babies in De-Identified Electronic Health Data [Dataset]. http://doi.org/10.1371/journal.pone.0164667
    Explore at:
    docxAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Katie Harron; Ruth Gilbert; David Cromwell; Jan van der Meulen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    ObjectiveLinkage of longitudinal administrative data for mothers and babies supports research and service evaluation in several populations around the world. We established a linked mother-baby cohort using pseudonymised, population-level data for England.Design and SettingRetrospective linkage study using electronic hospital records of mothers and babies admitted to NHS hospitals in England, captured in Hospital Episode Statistics between April 2001 and March 2013.ResultsOf 672,955 baby records in 2012/13, 280,470 (42%) linked deterministically to a maternal record using hospital, GP practice, maternal age, birthweight, gestation, birth order and sex. A further 380,164 (56%) records linked using probabilistic methods incorporating additional variables that could differ between mother/baby records (admission dates, ethnicity, 3/4-character postcode district) or that include missing values (delivery variables). The false-match rate was estimated at 0.15% using synthetic data. Data quality improved over time: for 2001/02, 91% of baby records were linked (holding the estimated false-match rate at 0.15%). The linked cohort was representative of national distributions of gender, gestation, birth weight and maternal age, and captured approximately 97% of births in England.ConclusionProbabilistic linkage of maternal and baby healthcare characteristics offers an efficient way to enrich maternity data, improve data quality, and create longitudinal cohorts for research and service evaluation. This approach could be extended to linkage of other datasets that have non-disclosive characteristics in common.

  6. o

    Curtailment (SPEN_009) Data Quality Checks

    • spenergynetworks.opendatasoft.com
    Updated Mar 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Curtailment (SPEN_009) Data Quality Checks [Dataset]. https://spenergynetworks.opendatasoft.com/explore/dataset/spen_data_quality_curtailment/
    Explore at:
    Dataset updated
    Mar 28, 2025
    Description

    This data table provides the detailed data quality assessment scores for the Curtailment dataset. The quality assessment was carried out on the 31st of March. At SPEN, we are dedicated to sharing high-quality data with our stakeholders and being transparent about its' quality. This is why we openly share the results of our data quality assessments. We collaborate closely with Data Owners to address any identified issues and enhance our overall data quality. To demonstrate our progress we conduct, at a minimum, bi-annual assessments of our data quality - for datasets that are refreshed more frequently than this, please note that the quality assessment may be based on an earlier version of the dataset. To learn more about our approach to how we assess data quality, visit Data Quality - SP Energy Networks.We welcome feedback and questions from our stakeholders regarding this process. Our Open Data Team is available to answer any enquiries or receive feedback on the assessments. You can contact them via our Open Data mailbox at opendata@spenergynetworks.co.uk.The first phase of our comprehensive data quality assessment measures the quality of our datasets across three dimensions. Please refer to the data table schema for the definitions of these dimensions. We are now in the process of expanding our quality assessments to include additional dimensions to provide a more comprehensive evaluation and will update the data tables with the results when available.DisclaimerThe data quality assessment may not represent the quality of the current dataset that is published on the Open Data Portal. Please check the date of the latest quality assessment and compare to the 'Modified' date of the corresponding dataset. The data quality assessments will be updated on either a quarterly or annual basis, dependent on the update frequency of the dataset. This information can be found in the dataset metadata, within the Information tab. If you require a more up to date quality assessment, please contact the Open Data Team at opendata@spenergynetworks.co.uk and a member of the team will be in contact.

  7. MULTI-SITE EVALUATION OF A DATA QUALITY TOOL FOR BIG DATA IN HEALTHCARE

    • figshare.com
    xlsx
    Updated Jan 20, 2016
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Vojtech Huser (2016). MULTI-SITE EVALUATION OF A DATA QUALITY TOOL FOR BIG DATA IN HEALTHCARE [Dataset]. http://doi.org/10.6084/m9.figshare.1497942.v4
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Jan 20, 2016
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Vojtech Huser
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Evaluation of data quality in large healthcare datasets.

    abstract: Data quality and fitness for analysis are crucial if outputs of big data analyses should be trusted by the public and the research community. Here we analyze the output from a data quality tool called Achilles Heel as it was applied to 24 datasets across seven different organizations. We highlight 12 data quality rules that identified issues in at least 10 of the 24 datasets and provide a full set of 71 rules identified in at least one dataset. Achilles Heel is developed by Observational Health Data Sciences and Informatics (OHDSI) community and is a freely available software that provides a useful starter set of data quality rules. Our analysis represents the first data quality comparison of multiple datasets across several countries in America, Europe and Asia.

  8. d

    Data Quality Utility

    • catalog.data.gov
    • data.virginia.gov
    Updated Sep 8, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Administration for Children and Families (2025). Data Quality Utility [Dataset]. https://catalog.data.gov/dataset/data-quality-utility
    Explore at:
    Dataset updated
    Sep 8, 2025
    Dataset provided by
    Administration for Children and Families
    Description

    The Data Quality Utility performs comprehensive checks on AFCARS data to help title IV-E agencies assess and improve data quality. Metadata-only record linking to the original dataset. Open original dataset below.

  9. f

    Questionnaire for Evaluating the Effects of Data, System, and Service...

    • datasetcatalog.nlm.nih.gov
    • data.4tu.nl
    • +1more
    Updated Apr 20, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Purwanto, Arie; Zuiderwijk, Anneke; Janssen, M. F. W. H. A. (2020). Questionnaire for Evaluating the Effects of Data, System, and Service Quality on Citizens' Trust [Dataset]. https://datasetcatalog.nlm.nih.gov/dataset?q=0000463146
    Explore at:
    Dataset updated
    Apr 20, 2020
    Authors
    Purwanto, Arie; Zuiderwijk, Anneke; Janssen, M. F. W. H. A.
    Description

    The dataset contains the questions asked in the survey over which quantitative data was collected to evaluate the effects of data quality, system quality, and service quality on citizens' trust.

  10. Data from: Evaluating the Quality of Survey and Administrative Data with...

    • tandf.figshare.com
    zip
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    D. L. Oberski; A. Kirchner; S. Eckman; F. Kreuter (2023). Evaluating the Quality of Survey and Administrative Data with Generalized Multitrait-Multimethod Models [Dataset]. http://doi.org/10.6084/m9.figshare.4742170.v3
    Explore at:
    zipAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    Taylor & Francishttps://taylorandfrancis.com/
    Authors
    D. L. Oberski; A. Kirchner; S. Eckman; F. Kreuter
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Administrative data are increasingly important in statistics, but, like other types of data, may contain measurement errors. To prevent such errors from invalidating analyses of scientific interest, it is therefore essential to estimate the extent of measurement errors in administrative data. Currently, however, most approaches to evaluate such errors involve either prohibitively expensive audits or comparison with a survey that is assumed perfect. We introduce the “generalized multitrait-multimethod” (GMTMM) model, which can be seen as a general framework for evaluating the quality of administrative and survey data simultaneously. This framework allows both survey and administrative data to contain random and systematic measurement errors. Moreover, it accommodates common features of administrative data such as discreteness, nonlinearity, and nonnormality, improving similar existing models. The use of the GMTMM model is demonstrated by application to linked survey-administrative data from the German Federal Employment Agency on income from of employment, and a simulation study evaluates the estimates obtained and their robustness to model misspecification. Supplementary materials for this article are available online.

  11. TBDQ: A Pragmatic Task-Based Method to Data Quality Assessment and...

    • plos.figshare.com
    jpeg
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Reza Vaziri; Mehran Mohsenzadeh; Jafar Habibi (2023). TBDQ: A Pragmatic Task-Based Method to Data Quality Assessment and Improvement [Dataset]. http://doi.org/10.1371/journal.pone.0154508
    Explore at:
    jpegAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Reza Vaziri; Mehran Mohsenzadeh; Jafar Habibi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Organizations are increasingly accepting data quality (DQ) as a major key to their success. In order to assess and improve DQ, methods have been devised. Many of these methods attempt to raise DQ by directly manipulating low quality data. Such methods operate reactively and are suitable for organizations with highly developed integrated systems. However, there is a lack of a proactive DQ method for businesses with weak IT infrastructure where data quality is largely affected by tasks that are performed by human agents. This study aims to develop and evaluate a new method for structured data, which is simple and practical so that it can easily be applied to real world situations. The new method detects the potentially risky tasks within a process, and adds new improving tasks to counter them. To achieve continuous improvement, an award system is also developed to help with the better selection of the proposed improving tasks. The task-based DQ method (TBDQ) is most appropriate for small and medium organizations, and simplicity in implementation is one of its most prominent features. TBDQ is case studied in an international trade company. The case study shows that TBDQ is effective in selecting optimal activities for DQ improvement in terms of cost and improvement.

  12. a

    07.2 Assessing Data Quality using ArcGIS Data

    • training-iowadot.opendata.arcgis.com
    • hub.arcgis.com
    • +1more
    Updated Feb 23, 2017
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Iowa Department of Transportation (2017). 07.2 Assessing Data Quality using ArcGIS Data [Dataset]. https://training-iowadot.opendata.arcgis.com/documents/c6c18d21a59a44588933122e2695022d
    Explore at:
    Dataset updated
    Feb 23, 2017
    Dataset authored and provided by
    Iowa Department of Transportation
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In this seminar, the presenter introduces essential concepts of ArcGIS Data Reviewer and highlights automated and semi-automated methods to streamline and expedite data validation.This seminar was developed to support the following:ArcGIS Desktop 10.3 (Basic, Standard, or Advanced)ArcGIS Server 10.3 Workgroup (Standard Or Advanced)ArcGIS Data Reviewer for DesktopArcGIS Data Reviewer for Server

  13. O

    Data Quality Score

    • data.winnipeg.ca
    csv, xlsx, xml
    Updated Nov 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Data Quality Score [Dataset]. https://data.winnipeg.ca/w/73sq-j2qi/swpr-bv7p?cur=nRy3UZz8x5A
    Explore at:
    xml, xlsx, csvAvailable download formats
    Dataset updated
    Nov 24, 2025
    License

    Open Government Licence - Canada 2.0https://open.canada.ca/en/open-government-licence-canada
    License information was derived automatically

    Description

    This dataset is a meta-data evaluation of the public datasets on the Open Data portal. Each public dataset is evaluated based on a variety of topics, and assigned a score between 0 and 100.
    The datasets are assigned a meta data attribute based on the following scores: • 0-70: Bronze • 71-80: Silver • 81-100: Gold For more information about the method by which the score is calculated, please visit the following PDF: http://wpgopendata.blob.core.windows.net/documents/Data-Quality-Score-Documentation.pdf

  14. Table 1_The development and evaluation of a quality assessment framework for...

    • frontiersin.figshare.com
    • figshare.com
    docx
    Updated Jun 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Laura A. Bardon; Grace Bennett; Michelle Weech; Faustina Hwang; Eve F. A. Kelly; Julie A. Lovegrove; Panče Panov; Siân Astley; Paul Finglas; Eileen R. Gibney (2025). Table 1_The development and evaluation of a quality assessment framework for reuse of dietary intake data: an FNS-Cloud study.docx [Dataset]. http://doi.org/10.3389/fnut.2025.1519401.s001
    Explore at:
    docxAvailable download formats
    Dataset updated
    Jun 6, 2025
    Dataset provided by
    Frontiers Mediahttp://www.frontiersin.org/
    Authors
    Laura A. Bardon; Grace Bennett; Michelle Weech; Faustina Hwang; Eve F. A. Kelly; Julie A. Lovegrove; Panče Panov; Siân Astley; Paul Finglas; Eileen R. Gibney
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    A key aim of the FNS-Cloud project (grant agreement no. 863059) was to overcome fragmentation within food, nutrition and health data through development of tools and services facilitating matching and merging of data to promote increased reuse. However, in an era of increasing data reuse, it is imperative that the scientific quality of data analysis is maintained. Whilst it is true that many datasets can be reused, questions remain regarding whether they should be, thus, there is a need to support researchers making such a decision. This paper describes the development and evaluation of the FNS-Cloud data quality assessment tool for dietary intake datasets. Markers of quality were identified from the literature for dietary intake, lifestyle, demographic, anthropometric, and consumer behavior data at all levels of data generation (data collection, underlying data sources used, dataset management and data analysis). These markers informed the development of a quality assessment framework, which comprised of decision trees and feedback messages relating to each quality parameter. These fed into a report provided to the researcher on completion of the assessment, with considerations to support them in deciding whether the dataset is appropriate for reuse. This quality assessment framework was transformed into an online tool and a user evaluation study undertaken. Participants recruited from three centres (N = 13) were observed and interviewed while using the tool to assess the quality of a dataset they were familiar with. Participants positively rated the assessment format and feedback messages in helping them assess the quality of a dataset. Several participants quoted the tool as being potentially useful in training students and inexperienced researchers in the use of secondary datasets. This quality assessment tool, deployed within FNS-Cloud, is openly accessible to users as one of the first steps in identifying datasets suitable for use in their specific analyses. It is intended to support researchers in their decision-making process of whether previously collected datasets under consideration for reuse are fit their new intended research purposes. While it has been developed and evaluated, further testing and refinement of this resource would improve its applicability to a broader range of users.

  15. Z

    Linked Data Quality Assessment for Datasets on the LOD Cloud

    • data.niaid.nih.gov
    Updated Jan 24, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Debattista, Jeremy; Lange, Christoph; Auer, Sören (2020). Linked Data Quality Assessment for Datasets on the LOD Cloud [Dataset]. https://data.niaid.nih.gov/resources?id=ZENODO_633197
    Explore at:
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    University of Bonn
    Authors
    Debattista, Jeremy; Lange, Christoph; Auer, Sören
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    For more up to date quality metadata, please visit https://w3id.org/lodquator

    This dataset is a collection of TRiG files with quality metadata for different datasets on the LOD cloud. Each dataset was assessed for

    The length of URIs

    Usage of RDF primitives

    Re-use of existing terms

    Usage of undefined terms

    Usage of blank nodes

    Indication for different serialisation formats

    Usage of multiple languages

    This data dump is part of the empirical study conducted for the paper "Are LOD Cloud Datasets Well Represented? A Data Representation Quality Survey."

    For more information visit http://jerdeb.github.io/lodqa

  16. HIFLD OPEN Data Quality Assessment Areas (USACE IENC)

    • datalumos.org
    Updated Oct 21, 2025
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    US Army Corps of Engineers (USACE) (2025). HIFLD OPEN Data Quality Assessment Areas (USACE IENC) [Dataset]. http://doi.org/10.3886/E239077V1
    Explore at:
    Dataset updated
    Oct 21, 2025
    Dataset provided by
    United States Army Corps of Engineershttp://www.usace.army.mil/
    United States Department of Homeland Security
    License

    https://creativecommons.org/share-your-work/public-domain/pdmhttps://creativecommons.org/share-your-work/public-domain/pdm

    Time period covered
    Apr 8, 2025
    Area covered
    United States
    Description

    The USACE IENCs coverage area consists of 7,260 miles across 21 rivers primarily located in the Central United States. IENCs apply to inland waterways that are maintained for navigation by USACE for shallow-draft vessels (e.g., maintained at a depth of 9-14 feet, dependent upon the waterway project authorization). Generally, IENCs are produced for those commercially navigable waterways which the National Oceanic and Atmospheric Administration (NOAA) does not produce Electronic Navigational Charts (ENCs). However, Special Purpose IENCs may be produced in agreement with NOAA. IENC POC: IENC_POC@usace.army.mil

  17. H

    Replication Data for: Assessing Data Quality: An Approach and An Application...

    • dataverse.harvard.edu
    • search.dataone.org
    Updated Sep 29, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Kelly McMann; Daniel Pemstein; Brigitte Seim; Jan Teorell; Staffan Lindberg (2022). Replication Data for: Assessing Data Quality: An Approach and An Application [Dataset]. http://doi.org/10.7910/DVN/BXV4AT
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Sep 29, 2022
    Dataset provided by
    Harvard Dataverse
    Authors
    Kelly McMann; Daniel Pemstein; Brigitte Seim; Jan Teorell; Staffan Lindberg
    License

    https://dataverse.harvard.edu/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.7910/DVN/BXV4AThttps://dataverse.harvard.edu/api/datasets/:persistentId/versions/1.1/customlicense?persistentId=doi:10.7910/DVN/BXV4AT

    Description

    Political scientists routinely face the challenge of assessing the quality (validity and reliability) of measures in order to use them in substantive research. While stand-alone assessment tools exist, researchers rarely combine them comprehensively. Further, while a large literature informs data producers, data consumers lack guidance on how to assess existing measures for use in substantive research. We delineate a three-component practical approach to data quality assessment that integrates complementary multi-method tools to assess: 1) content validity; 2) the validity and reliability of the data generation process; and 3) convergent validity. We apply our quality assessment approach to the corruption measures from the Varieties of Democracy (V-Dem) project, both illustrating our rubric and unearthing several quality advantages and disadvantages of the V-Dem measures, compared to other existing measures of corruption.

  18. Quality Assessment of the 2014 to 2019 NSDUH Public Use Files

    • data.virginia.gov
    • catalog.data.gov
    html
    Updated Sep 6, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Substance Abuse and Mental Health Services Administration (2025). Quality Assessment of the 2014 to 2019 NSDUH Public Use Files [Dataset]. https://data.virginia.gov/dataset/quality-assessment-of-the-2014-to-2019-nsduh-public-use-files
    Explore at:
    htmlAvailable download formats
    Dataset updated
    Sep 6, 2025
    Dataset provided by
    Substance Abuse and Mental Health Services Administrationhttps://www.samhsa.gov/
    Description

    Explore the data quality of the 2014-2019 National Survey on Drug Use and Health (NSDUH) Public Use Files (PUFs) and its comparability with the NSDUH Restricted Use Files (RUFs). This report demonstrates the overall quality of the NSDUH PUFs and the statistical disclosure control techniques used to create them.Chapters:Describes NSDUH and lays out the objective of the report.Presents an overview of the NSDUH disclosure concerns, briefly discusses the disclosure technique known as Micro Agglomeration, Substitution, Subsampling, and Calibration (MASSC), and provides a summary of this study’s quality assessment and research methods.Discusses how some of the detailed tables based on RUF data were selected and replicated using PUF data.Describes the data quality assessment results and findings.Summarizes the conclusions.There are also five appendices that support the analysis further.Aprevious reportanalyzed the PUF data from 2002-2013.

  19. U

    Fish and selected physical and chemical water-quality data for the U.S....

    • data.usgs.gov
    • search.dataone.org
    • +1more
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Michael Meador; James Kiesler, Fish and selected physical and chemical water-quality data for the U.S. Geological Survey Midwest Stream Quality Assessment [Dataset]. http://doi.org/10.5066/F7JS9NJZ
    Explore at:
    Dataset provided by
    United States Geological Surveyhttp://www.usgs.gov/
    Authors
    Michael Meador; James Kiesler
    License

    U.S. Government Workshttps://www.usa.gov/government-works
    License information was derived automatically

    Time period covered
    Jun 23, 2013 - Sep 11, 2013
    Description

    In 2013, the first of several Regional Stream Quality Assessments (RSQA) was done in the Midwest United States. The Midwest Stream Quality Assessment (MSQA) was a collaborative study by the U.S. Geological Survey National Water Quality Assessment and the U.S. Environmental Protection Agency National Rivers and Streams Assessment. One of the objectives of the RSQA, and thus the MSQA, is to characterize relations between stream ecology and water-quality stressors to determine the relative effects of these stressors on aquatic biota in streams. Data required to meet this objective included fish species and abundance data and physical and chemical water-quality characteristics of the ecological reaches of the sites that were sampled. This dataset comprises 135 fish species, 39,920 fish, 10 selected water-quality stressor metrics, and six selected fish community stressor response variables for 98 sites sampled for the MSQA.

  20. m

    Reliability and applicability of the revised Cochrane risk-of-bias tool for...

    • data.mendeley.com
    Updated Jul 6, 2020
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Rita Banzi (2020). Reliability and applicability of the revised Cochrane risk-of-bias tool for randomised trials (RoB 2): low inter-rater reliability and challenges in application [Dataset]. http://doi.org/10.17632/d46f5hsvnz.1
    Explore at:
    Dataset updated
    Jul 6, 2020
    Authors
    Rita Banzi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Dataset of the paper: Reliability and applicability of the revised Cochrane risk-of-bias tool for randomised trials (RoB 2): low inter-rater reliability and challenges in application

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
(2025). Long Term Development Statement (SPEN_002) Data Quality Checks [Dataset]. https://spenergynetworks.opendatasoft.com/explore/dataset/spen_data_quality_ltds/

Long Term Development Statement (SPEN_002) Data Quality Checks

Explore at:
Dataset updated
Mar 28, 2025
Description

This data table provides the detailed data quality assessment scores for the Long Term Development Statement dataset. The quality assessment was carried out on 31st March. At SPEN, we are dedicated to sharing high-quality data with our stakeholders and being transparent about its' quality. This is why we openly share the results of our data quality assessments. We collaborate closely with Data Owners to address any identified issues and enhance our overall data quality; to demonstrate our progress we conduct annual assessments of our data quality in line with the dataset refresh rate. To learn more about our approach to how we assess data quality, visit Data Quality - SP Energy Networks.We welcome feedback and questions from our stakeholders regarding this process. Our Open Data Team is available to answer any enquiries or receive feedback on the assessments. You can contact them via our Open Data mailbox at opendata@spenergynetworks.co.uk.The first phase of our comprehensive data quality assessment measures the quality of our datasets across three dimensions. Please refer to the data table schema for the definitions of these dimensions. We are now in the process of expanding our quality assessments to include additional dimensions to provide a more comprehensive evaluation and will update the data tables with the results when available.DisclaimerThe data quality assessment may not represent the quality of the current dataset that is published on the Open Data Portal. Please check the date of the latest quality assessment and compare to the 'Modified' date of the corresponding dataset. The data quality assessments will be updated on either a quarterly or annual basis, dependent on the update frequency of the dataset. This information can be found in the dataset metadata, within the Information tab. If you require a more up to date quality assessment, please contact the Open Data Team at opendata@spenergynetworks.co.uk and a member of the team will be in contact.

Search
Clear search
Close search
Google apps
Main menu