13 datasets found
  1. d

    Johns Hopkins COVID-19 Case Tracker

    • data.world
    • kaggle.com
    csv, zip
    Updated Dec 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Associated Press (2025). Johns Hopkins COVID-19 Case Tracker [Dataset]. https://data.world/associatedpress/johns-hopkins-coronavirus-case-tracker
    Explore at:
    zip, csvAvailable download formats
    Dataset updated
    Dec 3, 2025
    Authors
    The Associated Press
    Time period covered
    Jan 22, 2020 - Mar 9, 2023
    Area covered
    Description

    Updates

    • Notice of data discontinuation: Since the start of the pandemic, AP has reported case and death counts from data provided by Johns Hopkins University. Johns Hopkins University has announced that they will stop their daily data collection efforts after March 10. As Johns Hopkins stops providing data, the AP will also stop collecting daily numbers for COVID cases and deaths. The HHS and CDC now collect and visualize key metrics for the pandemic. AP advises using those resources when reporting on the pandemic going forward.

    • April 9, 2020

      • The population estimate data for New York County, NY has been updated to include all five New York City counties (Kings County, Queens County, Bronx County, Richmond County and New York County). This has been done to match the Johns Hopkins COVID-19 data, which aggregates counts for the five New York City counties to New York County.
    • April 20, 2020

      • Johns Hopkins death totals in the US now include confirmed and probable deaths in accordance with CDC guidelines as of April 14. One significant result of this change was an increase of more than 3,700 deaths in the New York City count. This change will likely result in increases for death counts elsewhere as well. The AP does not alter the Johns Hopkins source data, so probable deaths are included in this dataset as well.
    • April 29, 2020

      • The AP is now providing timeseries data for counts of COVID-19 cases and deaths. The raw counts are provided here unaltered, along with a population column with Census ACS-5 estimates and calculated daily case and death rates per 100,000 people. Please read the updated caveats section for more information.
    • September 1st, 2020

      • Johns Hopkins is now providing counts for the five New York City counties individually.
    • February 12, 2021

      • The Ohio Department of Health recently announced that as many as 4,000 COVID-19 deaths may have been underreported through the state’s reporting system, and that the "daily reported death counts will be high for a two to three-day period."
      • Because deaths data will be anomalous for consecutive days, we have chosen to freeze Ohio's rolling average for daily deaths at the last valid measure until Johns Hopkins is able to back-distribute the data. The raw daily death counts, as reported by Johns Hopkins and including the backlogged death data, will still be present in the new_deaths column.
    • February 16, 2021

      - Johns Hopkins has reconciled Ohio's historical deaths data with the state.

      Overview

    The AP is using data collected by the Johns Hopkins University Center for Systems Science and Engineering as our source for outbreak caseloads and death counts for the United States and globally.

    The Hopkins data is available at the county level in the United States. The AP has paired this data with population figures and county rural/urban designations, and has calculated caseload and death rates per 100,000 people. Be aware that caseloads may reflect the availability of tests -- and the ability to turn around test results quickly -- rather than actual disease spread or true infection rates.

    This data is from the Hopkins dashboard that is updated regularly throughout the day. Like all organizations dealing with data, Hopkins is constantly refining and cleaning up their feed, so there may be brief moments where data does not appear correctly. At this link, you’ll find the Hopkins daily data reports, and a clean version of their feed.

    The AP is updating this dataset hourly at 45 minutes past the hour.

    To learn more about AP's data journalism capabilities for publishers, corporations and financial institutions, go here or email kromano@ap.org.

    Queries

    Use AP's queries to filter the data or to join to other datasets we've made available to help cover the coronavirus pandemic

    Interactive

    The AP has designed an interactive map to track COVID-19 cases reported by Johns Hopkins.

    @(https://datawrapper.dwcdn.net/nRyaf/15/)

    Interactive Embed Code

    <iframe title="USA counties (2018) choropleth map Mapping COVID-19 cases by county" aria-describedby="" id="datawrapper-chart-nRyaf" src="https://datawrapper.dwcdn.net/nRyaf/10/" scrolling="no" frameborder="0" style="width: 0; min-width: 100% !important;" height="400"></iframe><script type="text/javascript">(function() {'use strict';window.addEventListener('message', function(event) {if (typeof event.data['datawrapper-height'] !== 'undefined') {for (var chartId in event.data['datawrapper-height']) {var iframe = document.getElementById('datawrapper-chart-' + chartId) || document.querySelector("iframe[src*='" + chartId + "']");if (!iframe) {continue;}iframe.style.height = event.data['datawrapper-height'][chartId] + 'px';}}});})();</script>
    

    Caveats

    • This data represents the number of cases and deaths reported by each state and has been collected by Johns Hopkins from a number of sources cited on their website.
    • In some cases, deaths or cases of people who've crossed state lines -- either to receive treatment or because they became sick and couldn't return home while traveling -- are reported in a state they aren't currently in, because of state reporting rules.
    • In some states, there are a number of cases not assigned to a specific county -- for those cases, the county name is "unassigned to a single county"
    • This data should be credited to Johns Hopkins University's COVID-19 tracking project. The AP is simply making it available here for ease of use for reporters and members.
    • Caseloads may reflect the availability of tests -- and the ability to turn around test results quickly -- rather than actual disease spread or true infection rates.
    • Population estimates at the county level are drawn from 2014-18 5-year estimates from the American Community Survey.
    • The Urban/Rural classification scheme is from the Center for Disease Control and Preventions's National Center for Health Statistics. It puts each county into one of six categories -- from Large Central Metro to Non-Core -- according to population and other characteristics. More details about the classifications can be found here.

    Johns Hopkins timeseries data - Johns Hopkins pulls data regularly to update their dashboard. Once a day, around 8pm EDT, Johns Hopkins adds the counts for all areas they cover to the timeseries file. These counts are snapshots of the latest cumulative counts provided by the source on that day. This can lead to inconsistencies if a source updates their historical data for accuracy, either increasing or decreasing the latest cumulative count. - Johns Hopkins periodically edits their historical timeseries data for accuracy. They provide a file documenting all errors in their timeseries files that they have identified and fixed here

    Attribution

    This data should be credited to Johns Hopkins University COVID-19 tracking project

  2. f

    Low Heart Rate Variability in a 2-Minute Electrocardiogram Recording Is...

    • plos.figshare.com
    • datasetcatalog.nlm.nih.gov
    docx
    Updated May 30, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ankit Maheshwari; Faye L. Norby; Elsayed Z. Soliman; Selcuk Adabag; Eric A. Whitsel; Alvaro Alonso; Lin Y. Chen (2023). Low Heart Rate Variability in a 2-Minute Electrocardiogram Recording Is Associated with an Increased Risk of Sudden Cardiac Death in the General Population: The Atherosclerosis Risk in Communities Study [Dataset]. http://doi.org/10.1371/journal.pone.0161648
    Explore at:
    docxAvailable download formats
    Dataset updated
    May 30, 2023
    Dataset provided by
    PLOS ONE
    Authors
    Ankit Maheshwari; Faye L. Norby; Elsayed Z. Soliman; Selcuk Adabag; Eric A. Whitsel; Alvaro Alonso; Lin Y. Chen
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Low heart rate variability (HRV) has been linked to increased total mortality in the general population; however, the relationship between low HRV and sudden cardiac death (SCD) is less well-characterized. The goal of this study was to evaluate the relationship between low HRV and SCD in a community-based cohort. Our cohort consisted of 12,543 participants from the Atherosclerosis Risk in Communities (ARIC) study. HRV measures were derived from 2-minute electrocardiogram recordings obtained during the baseline exam (1987–89). Time domain measurements included the standard deviation of all normal RR intervals (SDNN) and the root mean squared successive difference (r-MSSD). Frequency domain measurements included low frequency power (LF) and high frequency (HF) power. During a median follow-up of 13 years, 215 SCDs were identified from physician adjudication of all coronary heart disease deaths through 2001. In multivariable adjusted Cox proportional hazards models, each standard deviation decrement in SDNN, LF, and HF were associated with 24%, 27% and 16% increase in SCD risk, respectively. Low HRV is independently associated with increased risk of SCD in the general population.

  3. Deaths, by month

    • www150.statcan.gc.ca
    • gimi9.com
    • +2more
    Updated Feb 19, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Government of Canada, Statistics Canada (2025). Deaths, by month [Dataset]. http://doi.org/10.25318/1310070801-eng
    Explore at:
    Dataset updated
    Feb 19, 2025
    Dataset provided by
    Government of Canadahttp://www.gg.ca/
    Statistics Canadahttps://statcan.gc.ca/en
    Area covered
    Canada
    Description

    Number and percentage of deaths, by month and place of residence, 1991 to most recent year.

  4. Metallica Spotify Dataset

    • kaggle.com
    zip
    Updated Nov 26, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2022). Metallica Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/metallica-spotify-dataset
    Explore at:
    zip(78945 bytes)Available download formats
    Dataset updated
    Nov 26, 2022
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API for the band Metallica. Metallica is one of the most successful bands in the music industry's history. Starting out in California, Metallica has gone on to play all over the world.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions

    I hope you find this data to be useful, enjoy!

  5. Dataset related to this study.

    • plos.figshare.com
    bin
    Updated Jun 6, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Erean Shigign Malka; Tarekegn Solomon; Dejene Hailu Kassa; Besfat Berihun Erega; Derara Girma Tufa (2024). Dataset related to this study. [Dataset]. http://doi.org/10.1371/journal.pone.0302665.s002
    Explore at:
    binAvailable download formats
    Dataset updated
    Jun 6, 2024
    Dataset provided by
    PLOShttp://plos.org/
    Authors
    Erean Shigign Malka; Tarekegn Solomon; Dejene Hailu Kassa; Besfat Berihun Erega; Derara Girma Tufa
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    IntroductionThe largest risk of child mortality occurs within the first week after birth. Early neonatal mortality remains a global public health concern, especially in sub-Saharan African countries. More than 75% of neonatal death occurs within the first seven days of birth, but there are limited prospective follow- up studies to determine time to death, incidence and predictors of death in Ethiopia particularly in the study area. The study aimed to determine incidence and predictors of early neonatal mortality among neonates admitted to the neonatal intensive care unit of Addis Ababa public hospitals, Ethiopia 2021.MethodsInstitutional prospective cohort study was conducted in four public hospitals found in Addis Ababa City, Ethiopia from June 7th, 2021 to July 13th, 2021. All early neonates consecutively admitted to the corresponding neonatal intensive care unit of selected hospitals were included in the study and followed until 7 days-old. Data were coded, cleaned, edited, and entered into Epi data version 3.1 and then exported to STATA software version 14.0 for analysis. The Kaplan Meier survival curve with log- rank test was used to compare survival time between groups. Moreover, both bi-variable and multivariable Cox proportional hazard regression model was used to identify the predictors of early neonatal mortality. All variables having P-value ≤0.2 in the bi-variable analysis model were further fitted to the multivariable model. The assumption of the model was checked graphically and using a global test. The goodness of fit of the model was performed using the Cox-Snell residual test and it was adequate.ResultsA total of 391 early neonates with their mothers were involved in this study. The incidence rate among admitted early neonates was 33.25 per 1000 neonate day’s observation [95% confidence interval (CI): 26.22, 42.17]. Being preterm birth [adjusted hazard ratio (AHR): 6.0 (95% CI 2.02, 17.50)], having low fifth minute Apgar score [AHR: 3.93 (95% CI; 1.5, 6.77)], low temperatures [AHR: 2.67 (95%CI; 1.41, 5.02)] and, resuscitating of early neonate [AHR: 2.80 (95% CI; 1.51,5.10)] were associated with increased hazard of early neonatal death. However, early neonatal crying at birth [AHR: 0.48 (95%CI; 0.26, 0.87)] was associated with reduced hazard of death.ConclusionsEarly neonatal mortality is high in Addis Ababa public Hospitals. Preterm birth, low five-minute Apgar score, hypothermia and crying at birth were found to be independent predictors of early neonatal death. Good care and attention to neonate with low Apgar scores, premature, and hypothermic neonates.

  6. f

    Data from: S1 Dataset -

    • plos.figshare.com
    xlsx
    Updated Jul 10, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Leandra Nagler; Carmen Eißmann; Marita Wasenitz; Franz Bahlmann; Ammar Al Naimi (2024). S1 Dataset - [Dataset]. http://doi.org/10.1371/journal.pone.0306877.s001
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Jul 10, 2024
    Dataset provided by
    PLOS ONE
    Authors
    Leandra Nagler; Carmen Eißmann; Marita Wasenitz; Franz Bahlmann; Ammar Al Naimi
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The prevalence of overweight and obese people worldwide has dramatically increased in the last decades and is yet to peak. At the same time and partly due to obesity and associated assisted reproduction, twinning rates showed a clear rise in the last years. Adverse fetomaternal outcomes are known to occur in singleton and twin pregnancies in overweight and obese women. However, the impact of the obesity levels as defined by the World Health Organization on the outcomes of twin pregnancies has not been thoroughly studied. Therefore, the purpose of this study is to examine how maternal overweight, and the level of obesity affect fetomaternal outcomes in twin pregnancies, hypothesizing a higher likelihood for adverse outcomes with overweight and each obesity level. This is a retrospective cohort study with 2,349 twin pregnancies that delivered at the Buergerhospital Frankfurt, Germany between 2005 and 2020. The mothers were divided into exposure groups depending on their pre-gestational body mass index; these were normal weight (reference group), overweight and obesity levels I, II, and III. A multivariate logistic regression analysis was performed to assess the influence of overweight and obesity on gestational diabetes mellitus, preeclampsia, postpartum hemorrhage, intrauterine fetal death, and a five-minutes Apgar score below seven. The adjusted odds ratio for gestational diabetes compared to normal weight mothers were 1.47, 2.79, 4.05, and 6.40 for overweight and obesity levels I, II and III respectively (p = 0.015 for overweight and p < 0.001 for each obesity level). Maternal BMI had a significant association with the risk of preeclampsia (OR 1.04, p = 0.028). Overweight and obesity did not affect the odds of postpartum hemorrhage, fetal demise, or a low Apgar score. While maternal overweight and obesity did not influence the fetal outcomes in twin pregnancies, they significantly increased the risk of gestational diabetes and preeclampsia, and that risk is incremental with increasing level of obesity.

  7. Elton John Spotify Dataset

    • kaggle.com
    zip
    Updated Nov 26, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2022). Elton John Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/elton-john-spotify-dataset
    Explore at:
    zip(85928 bytes)Available download formats
    Dataset updated
    Nov 26, 2022
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for Elton John. At first it may look like there are song duplicates but I checked and all song IDs are unique.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions Data cleaning

    I hope you find this data to be useful, enjoy!

  8. Tame Impala Spotify Dataset

    • kaggle.com
    zip
    Updated Aug 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2024). Tame Impala Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/tame-impala-spotify-dataset
    Explore at:
    zip(4956 bytes)Available download formats
    Dataset updated
    Aug 1, 2024
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for Tame Impala. I set up the dataset to update monthly so that if any albums get added it will get added to the dataset too. At first it may look like there are song duplicates but I checked and all song IDs are unique.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions Data cleaning

    I hope you find this data to be useful, enjoy!

  9. Taylor Swift Spotify Dataset

    • kaggle.com
    zip
    Updated Aug 3, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2024). Taylor Swift Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/taylor-swift-spotify-dataset/
    Explore at:
    zip(37834 bytes)Available download formats
    Dataset updated
    Aug 3, 2024
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for Taylor Swift. I set up the dataset to update monthly so that if any albums get added it will get added to the dataset too. At first it may look like there are song duplicates but I checked and all song IDs are unique.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions Data cleaning

    I hope you find this data to be useful, enjoy!

  10. The Beatles Spotify Dataset

    • kaggle.com
    zip
    Updated Nov 24, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2022). The Beatles Spotify Dataset [Dataset]. https://www.kaggle.com/jarredpriester/the-beatles-spotify-dataset
    Explore at:
    zip(49118 bytes)Available download formats
    Dataset updated
    Nov 24, 2022
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for The Beatles.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions

    I hope you find this data to be useful, enjoy!

  11. Ed Sheeran Spotify Dataset

    • kaggle.com
    zip
    Updated Aug 2, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2024). Ed Sheeran Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/ed-sheeran-spotify-dataset
    Explore at:
    zip(17711 bytes)Available download formats
    Dataset updated
    Aug 2, 2024
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for Ed Sheeran. I set up the dataset to update monthly so that if any albums get added it will get added to the dataset too. At first it may look like there are song duplicates but I checked and all song IDs are unique.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions Data cleaning

    I hope you find this data to be useful, enjoy!

  12. Rolling Stones Spotify Dataset

    • kaggle.com
    zip
    Updated Nov 27, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jarred Priester (2022). Rolling Stones Spotify Dataset [Dataset]. https://www.kaggle.com/datasets/jarredpriester/rolling-stones-spotify-dataset/data
    Explore at:
    zip(107026 bytes)Available download formats
    Dataset updated
    Nov 27, 2022
    Authors
    Jarred Priester
    Description

    This dataset consist of data from Spotify's API on all albums listed on Spotify for the Rolling Stones. At first it may look like there are song duplicates but I checked and all song IDs are unique.

    The columns in this dataset are:

    name - the name of the song

    album - the name of the album

    release_date - the day month and year the album was released

    track number - the order the song appears on the album

    id - the Spotify id for the song

    uri - the Spotify uri for the song

    acousticness - A confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic.

    danceability - Danceability describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable.

    energy - Energy is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy.

    instrumentalness - Predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0.

    liveness - Detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live.

    loudness - The overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db.

    speechiness - detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks.

    tempo - The overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration.

    valence - A measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry).

    popularity - the popularity of the song from 0 to 100

    duration_ms - The duration of the track in milliseconds.

    Possible ways to use this data:

    Data exploration Data visualization Recommendation systems Cluster analysis Popularity predictions Data cleaning

    I hope you find this data to be useful, enjoy!

  13. Spotify and Youtube

    • kaggle.com
    zip
    Updated Mar 20, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Salvatore Rastelli (2023). Spotify and Youtube [Dataset]. https://www.kaggle.com/datasets/salvatorerastelli/spotify-and-youtube/code
    Explore at:
    zip(9386454 bytes)Available download formats
    Dataset updated
    Mar 20, 2023
    Authors
    Salvatore Rastelli
    License

    https://creativecommons.org/publicdomain/zero/1.0/https://creativecommons.org/publicdomain/zero/1.0/

    Area covered
    YouTube
    Description

    Dataset of songs of various artist in the world and for each song is present: - Several statistics of the music version on spotify, including the number of streams; - Number of views of the official music video of the song on youtube.

    Content

    It includes 26 variables for each of the songs collected from spotify. These variables are briefly described next: - Track: name of the song, as visible on the Spotify platform. - Artist: name of the artist. - Url_spotify: the Url of the artist. - Album: the album in wich the song is contained on Spotify. - Album_type: indicates if the song is relesead on Spotify as a single or contained in an album. - Uri: a spotify link used to find the song through the API. - Danceability: describes how suitable a track is for dancing based on a combination of musical elements including tempo, rhythm stability, beat strength, and overall regularity. A value of 0.0 is least danceable and 1.0 is most danceable. - Energy: is a measure from 0.0 to 1.0 and represents a perceptual measure of intensity and activity. Typically, energetic tracks feel fast, loud, and noisy. For example, death metal has high energy, while a Bach prelude scores low on the scale. Perceptual features contributing to this attribute include dynamic range, perceived loudness, timbre, onset rate, and general entropy. - Key: the key the track is in. Integers map to pitches using standard Pitch Class notation. E.g. 0 = C, 1 = C♯/D♭, 2 = D, and so on. If no key was detected, the value is -1. - Loudness: the overall loudness of a track in decibels (dB). Loudness values are averaged across the entire track and are useful for comparing relative loudness of tracks. Loudness is the quality of a sound that is the primary psychological correlate of physical strength (amplitude). Values typically range between -60 and 0 db. - Speechiness: detects the presence of spoken words in a track. The more exclusively speech-like the recording (e.g. talk show, audio book, poetry), the closer to 1.0 the attribute value. Values above 0.66 describe tracks that are probably made entirely of spoken words. Values between 0.33 and 0.66 describe tracks that may contain both music and speech, either in sections or layered, including such cases as rap music. Values below 0.33 most likely represent music and other non-speech-like tracks. - Acousticness: a confidence measure from 0.0 to 1.0 of whether the track is acoustic. 1.0 represents high confidence the track is acoustic. - Instrumentalness: predicts whether a track contains no vocals. "Ooh" and "aah" sounds are treated as instrumental in this context. Rap or spoken word tracks are clearly "vocal". The closer the instrumentalness value is to 1.0, the greater likelihood the track contains no vocal content. Values above 0.5 are intended to represent instrumental tracks, but confidence is higher as the value approaches 1.0. - Liveness: detects the presence of an audience in the recording. Higher liveness values represent an increased probability that the track was performed live. A value above 0.8 provides strong likelihood that the track is live. - Valence: a measure from 0.0 to 1.0 describing the musical positiveness conveyed by a track. Tracks with high valence sound more positive (e.g. happy, cheerful, euphoric), while tracks with low valence sound more negative (e.g. sad, depressed, angry). - Tempo: the overall estimated tempo of a track in beats per minute (BPM). In musical terminology, tempo is the speed or pace of a given piece and derives directly from the average beat duration. - Duration_ms: the duration of the track in milliseconds. - Stream: number of streams of the song on Spotify. - Url_youtube: url of the video linked to the song on Youtube, if it have any. - Title: title of the videoclip on youtube. - Channel: name of the channel that have published the video. - Views: number of views. - Likes: number of likes. - Comments: number of comments. - Description: description of the video on Youtube. - Licensed: Indicates whether the video represents licensed content, which means that the content was uploaded to a channel linked to a YouTube content partner and then claimed by that partner. - official_video: boolean value that indicates if the video found is the official video of the song.

    Notes

    These datas are heavily dependent on the time they were collected, which is in this case the 7th of February, 2023.

  14. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
The Associated Press (2025). Johns Hopkins COVID-19 Case Tracker [Dataset]. https://data.world/associatedpress/johns-hopkins-coronavirus-case-tracker

Johns Hopkins COVID-19 Case Tracker

Johns Hopkins' county-level COVID-19 case and death data, paired with population and rates per 100,000

Explore at:
10 scholarly articles cite this dataset (View in Google Scholar)
zip, csvAvailable download formats
Dataset updated
Dec 3, 2025
Authors
The Associated Press
Time period covered
Jan 22, 2020 - Mar 9, 2023
Area covered
Description

Updates

  • Notice of data discontinuation: Since the start of the pandemic, AP has reported case and death counts from data provided by Johns Hopkins University. Johns Hopkins University has announced that they will stop their daily data collection efforts after March 10. As Johns Hopkins stops providing data, the AP will also stop collecting daily numbers for COVID cases and deaths. The HHS and CDC now collect and visualize key metrics for the pandemic. AP advises using those resources when reporting on the pandemic going forward.

  • April 9, 2020

    • The population estimate data for New York County, NY has been updated to include all five New York City counties (Kings County, Queens County, Bronx County, Richmond County and New York County). This has been done to match the Johns Hopkins COVID-19 data, which aggregates counts for the five New York City counties to New York County.
  • April 20, 2020

    • Johns Hopkins death totals in the US now include confirmed and probable deaths in accordance with CDC guidelines as of April 14. One significant result of this change was an increase of more than 3,700 deaths in the New York City count. This change will likely result in increases for death counts elsewhere as well. The AP does not alter the Johns Hopkins source data, so probable deaths are included in this dataset as well.
  • April 29, 2020

    • The AP is now providing timeseries data for counts of COVID-19 cases and deaths. The raw counts are provided here unaltered, along with a population column with Census ACS-5 estimates and calculated daily case and death rates per 100,000 people. Please read the updated caveats section for more information.
  • September 1st, 2020

    • Johns Hopkins is now providing counts for the five New York City counties individually.
  • February 12, 2021

    • The Ohio Department of Health recently announced that as many as 4,000 COVID-19 deaths may have been underreported through the state’s reporting system, and that the "daily reported death counts will be high for a two to three-day period."
    • Because deaths data will be anomalous for consecutive days, we have chosen to freeze Ohio's rolling average for daily deaths at the last valid measure until Johns Hopkins is able to back-distribute the data. The raw daily death counts, as reported by Johns Hopkins and including the backlogged death data, will still be present in the new_deaths column.
  • February 16, 2021

    - Johns Hopkins has reconciled Ohio's historical deaths data with the state.

    Overview

The AP is using data collected by the Johns Hopkins University Center for Systems Science and Engineering as our source for outbreak caseloads and death counts for the United States and globally.

The Hopkins data is available at the county level in the United States. The AP has paired this data with population figures and county rural/urban designations, and has calculated caseload and death rates per 100,000 people. Be aware that caseloads may reflect the availability of tests -- and the ability to turn around test results quickly -- rather than actual disease spread or true infection rates.

This data is from the Hopkins dashboard that is updated regularly throughout the day. Like all organizations dealing with data, Hopkins is constantly refining and cleaning up their feed, so there may be brief moments where data does not appear correctly. At this link, you’ll find the Hopkins daily data reports, and a clean version of their feed.

The AP is updating this dataset hourly at 45 minutes past the hour.

To learn more about AP's data journalism capabilities for publishers, corporations and financial institutions, go here or email kromano@ap.org.

Queries

Use AP's queries to filter the data or to join to other datasets we've made available to help cover the coronavirus pandemic

Interactive

The AP has designed an interactive map to track COVID-19 cases reported by Johns Hopkins.

@(https://datawrapper.dwcdn.net/nRyaf/15/)

Interactive Embed Code

<iframe title="USA counties (2018) choropleth map Mapping COVID-19 cases by county" aria-describedby="" id="datawrapper-chart-nRyaf" src="https://datawrapper.dwcdn.net/nRyaf/10/" scrolling="no" frameborder="0" style="width: 0; min-width: 100% !important;" height="400"></iframe><script type="text/javascript">(function() {'use strict';window.addEventListener('message', function(event) {if (typeof event.data['datawrapper-height'] !== 'undefined') {for (var chartId in event.data['datawrapper-height']) {var iframe = document.getElementById('datawrapper-chart-' + chartId) || document.querySelector("iframe[src*='" + chartId + "']");if (!iframe) {continue;}iframe.style.height = event.data['datawrapper-height'][chartId] + 'px';}}});})();</script>

Caveats

  • This data represents the number of cases and deaths reported by each state and has been collected by Johns Hopkins from a number of sources cited on their website.
  • In some cases, deaths or cases of people who've crossed state lines -- either to receive treatment or because they became sick and couldn't return home while traveling -- are reported in a state they aren't currently in, because of state reporting rules.
  • In some states, there are a number of cases not assigned to a specific county -- for those cases, the county name is "unassigned to a single county"
  • This data should be credited to Johns Hopkins University's COVID-19 tracking project. The AP is simply making it available here for ease of use for reporters and members.
  • Caseloads may reflect the availability of tests -- and the ability to turn around test results quickly -- rather than actual disease spread or true infection rates.
  • Population estimates at the county level are drawn from 2014-18 5-year estimates from the American Community Survey.
  • The Urban/Rural classification scheme is from the Center for Disease Control and Preventions's National Center for Health Statistics. It puts each county into one of six categories -- from Large Central Metro to Non-Core -- according to population and other characteristics. More details about the classifications can be found here.

Johns Hopkins timeseries data - Johns Hopkins pulls data regularly to update their dashboard. Once a day, around 8pm EDT, Johns Hopkins adds the counts for all areas they cover to the timeseries file. These counts are snapshots of the latest cumulative counts provided by the source on that day. This can lead to inconsistencies if a source updates their historical data for accuracy, either increasing or decreasing the latest cumulative count. - Johns Hopkins periodically edits their historical timeseries data for accuracy. They provide a file documenting all errors in their timeseries files that they have identified and fixed here

Attribution

This data should be credited to Johns Hopkins University COVID-19 tracking project

Search
Clear search
Close search
Google apps
Main menu