10 datasets found
  1. P

    CHiME-5 Dataset

    • paperswithcode.com
    Updated Nov 16, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jon Barker; Shinji Watanabe; Emmanuel Vincent; Jan Trmal (2021). CHiME-5 Dataset [Dataset]. https://paperswithcode.com/dataset/chime-5
    Explore at:
    Dataset updated
    Nov 16, 2021
    Authors
    Jon Barker; Shinji Watanabe; Emmanuel Vincent; Jan Trmal
    Description

    The CHiME challenge series aims to advance robust automatic speech recognition (ASR) technology by promoting research at the interface of speech and language processing, signal processing , and machine learning.

  2. CHiME-7 UDASE evaluation data

    • zenodo.org
    • data.niaid.nih.gov
    zip
    Updated Sep 30, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Simon Leglaive; Simon Leglaive; Matthieu Fraticelli; Hend ElGhazaly; Léonie Borne; Mostafa Sadeghi; Mostafa Sadeghi; Scott Wisdom; Manuel Pariente; John Hershey; Daniel Pressnitzer; Daniel Pressnitzer; Jon Barker; Matthieu Fraticelli; Hend ElGhazaly; Léonie Borne; Scott Wisdom; Manuel Pariente; John Hershey; Jon Barker (2024). CHiME-7 UDASE evaluation data [Dataset]. http://doi.org/10.5281/zenodo.10418311
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 30, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Simon Leglaive; Simon Leglaive; Matthieu Fraticelli; Hend ElGhazaly; Léonie Borne; Mostafa Sadeghi; Mostafa Sadeghi; Scott Wisdom; Manuel Pariente; John Hershey; Daniel Pressnitzer; Daniel Pressnitzer; Jon Barker; Matthieu Fraticelli; Hend ElGhazaly; Léonie Borne; Scott Wisdom; Manuel Pariente; John Hershey; Jon Barker
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Description

    This repository contains audio and CSV files used in the evaluation of the UDASE task of the 7th CHiME challenge. In particular, it contains the audio samples and subjective ratings of an ITU-T P.835 listening test.
    If you use this material in your research, please cite the following paper:
    Simon Leglaive, Matthieu Fraticelli, Hend ElGhazaly, Léonie Borne, Mostafa Sadeghi, Scott Wisdom, Manuel Pariente, John R. Hershey, Daniel Pressnitzer, Jon P. Barker
    Computer Speech & Language, vol. 89, 2025

    General information

    The folder 'listening_test' contains:
    • the audio files used for the ITU-T P.835 listening test ('data' subfolder);
    • the subjective evaluation results:
      • 'raw_results_listening_test.csv' contains the individual ratings of each participant (identified by the file name in the 'csv' column);
      • 'MOS_results_listening_test.csv' files contains the mean opinion scores (MOS) computed from the participants' ratings.

    The folder 'objective_evaluation' contains:
    • a subset of the output audio files of the baseline and submitted speech enhancement methods ('data' subfolder);
    • the objective evaluation results ('results_objective_evaluation.csv' file).

    Additional details

    • 'listening_test/data/ref' contains truncated versions of the ITU-T P.501 (2017) test signals for use in telephonometry. We downloaded the files from Microsoft's P.808 Toolkit and we modified them to keep only the first utterance of each audio file and to normalize the loudness.
    • 'listening_test/data/C0' contains audio segments extracted from the binaural recordings of the CHiME-5 dataset ('eval' set). It corresponds to the unprocessed noisy speech condition.
    • 'listening_test/data/{C1, C2, C3, C4}' contains denoised versions of the audio files in 'listening_test/data/C0', where
      • condition 'C1' corresponds to the 'CMGAN-FT' system;
      • condition 'C2' corresponds to the 'ISDS1' system;
      • condition 'C3' corresponds to the 'N&B' system;
      • condition 'C4' corresponds to the 'RemixIT-VAD' system.
    • 'objective_evaluation/data/
    • We do not share the denoised mixtures of the reverberant LibriCHiME-5 dataset (which should be in 'objective_evaluation/data/

    Licences

    The files shared in this repository are licensed under a CC BY-SA 4.0 license. They were derived from the following datasets:

  3. h

    chime

    • huggingface.co
    Updated May 11, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Niklas Hoepner (2025). chime [Dataset]. https://huggingface.co/datasets/nhop/chime
    Explore at:
    Dataset updated
    May 11, 2025
    Authors
    Niklas Hoepner
    License

    Apache License, v2.0https://www.apache.org/licenses/LICENSE-2.0
    License information was derived automatically

    Description

    CHIME Dataset

    A unified version of the CHIME dataset created from the following resources:

    Paper Code 🤗 parent-child relations 🤗 siblings info 🤗 claim-category relations

      Overview
    

    The CHIME Dataset is a unified collection of data specifically designed for evaluating the capabilities of Large Language Models (LLMs) in generating literature reviews. It includes various relationships between academic papers, such as parent-child relations, sibling information, and… See the full description on the dataset page: https://huggingface.co/datasets/nhop/chime.

  4. P

    Kinect-WSJ Dataset

    • paperswithcode.com
    Updated Feb 21, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2021). Kinect-WSJ Dataset [Dataset]. https://paperswithcode.com/dataset/kinect-wsj
    Explore at:
    Dataset updated
    Feb 21, 2021
    Description

    Kinect-WSJ is a multichannel, multispeaker, reverberated, noisy dataset which extends the WSJ0-2mix singlechannel, non-reverberated, noiseless dataset to the strong reverberation and noise conditions and the Kinect-like microphone array geometry used in CHiME-5.

  5. ChannelSet: a composite dataset of diverse acoustic environments

    • zenodo.org
    zip
    Updated Jul 21, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Benjamin Skerritt-Davis; Benjamin Skerritt-Davis; Mattson Ogg; Mattson Ogg (2021). ChannelSet: a composite dataset of diverse acoustic environments [Dataset]. http://doi.org/10.5281/zenodo.5117366
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jul 21, 2021
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Benjamin Skerritt-Davis; Benjamin Skerritt-Davis; Mattson Ogg; Mattson Ogg
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    We introduce ChannelSet, a dataset which provides a launchpad for exploring the extraneous acoustic information typically suppressed or ignored in audio tasks such as automatic speech recognition. We combined components of existing publicly available datasets to encompass broad variability in recording equipment, microphone position, room or surrounding acoustics, event density (i.e., how many audio events are present), and proportion of foreground and background sounds. Source datasets include: the CHiME-3 background dataset, CHiME-5 evaluation dataset, AMI meeting corpus, Freefield1010, and Vystadial2016.

    ChannelSet includes 13 classes spanning various acoustic environments: Indoor_Commercial_Bus, Indoor_Commercial_Cafe, Indoor_Domestic, Indoor_Meeting_Room1, Indoor_Meeting_Room2, Indoor_Meeting_Room3, Outdoor_City_Pedestrian, Outdoor_City_Traffic, Outdoor_Nature_Birds, Outdoor_Nature_Water, Outdoor_Nature_Weather, Telephony_CZ, and Telephony_EN. Each sample is between 1 and 10 seconds in duration. Each class contains 100 minutes of audio, for a total of 21.6 hours, split into separate test (20%) and train (80%) partitions.

    Download includes scripts, metadata, and instructions for producing ChannelSet from source datasets.

  6. ArcGIS Pro COVID-19 Modeling Toolbox (Version 5 - Updated 11 MAY 2020)

    • prep-response-portal.napsgfoundation.org
    • coronavirus-disasterresponse.hub.arcgis.com
    Updated Apr 3, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Esri’s Disaster Response Program (2020). ArcGIS Pro COVID-19 Modeling Toolbox (Version 5 - Updated 11 MAY 2020) [Dataset]. https://prep-response-portal.napsgfoundation.org/content/37ad6eb0d1034cd58844314a9b305de2
    Explore at:
    Dataset updated
    Apr 3, 2020
    Dataset provided by
    Esrihttp://esri.com/
    Authors
    Esri’s Disaster Response Program
    Description

    Please note, the updated version of this toolbox is now available for download on this page. The COVID-19-Modeling-v1.zip file contains version 5 of the toolbox with updated documentation. Version 5 of the toolbox updates the CHIME Model v1.1.5 tool. The COVID-19Surge (CDC) model is unchanged in this version.More information about the toolbox can be found in the toolbox document. More information about the CHIME Model v1.1.5 tool, including the change log, can be found in the tool documentation and this video.More information about the COVID-19Surge (CDC) tool is included in the tool documentation and this video. CHIME Model v1.1.5 ToolVersion 4 - Updated 11 MAY 2020An implementation of Penn Medicine’s COVID-19 Hospital Impact Model for Epidemics (CHIME) for use in ArcGIS Pro 2.3 or later. This tool leverages SIR (Susceptible, Infected, Recovered) modeling to assist hospitals, cities, and regions with capacity planning around COVID-19 by providing estimates of daily new admissions and current inpatient hospitalizations (census), ICU admissions, and patients requiring ventilation. Version 4 of this tool is based on CHIME v1.1.5 (2020-05-07). Learn more about how CHIME works.Version 4 contains the following updates:Updated the CHIME tool from CHIME v1.1.2 to CHIME v1.1.5.Added a new parameter called Date of Social Distancing Measures Effect to specify the date when social distancing measures started showing their effects.Added a new parameter called Recovery to specify the number of recovered cases at the start of the model.COVID-19Surge (CDC) ToolVersion 1 - Released 04 MAY 2020An implementation of Centers for Disease Control and Prevention’s (CDC) COVID-19Surge for use in ArcGIS Pro 2.3 or later. This tool leverages SIICR (Susceptible, Infected, Infectious, Convalescing, Recovered) modeling to assist hospitals, cities, and regions with capacity planning around COVID-19 by providing estimates of daily new admissions and current inpatient hospitalizations (census), ICU admissions, and patients requiring ventilation based on the extent to which mitigation strategies such as social distancing or shelter-in-place recommendations are implemented. This tool is based on COVID-19Surge. Learn more about how COVID-19Surge works.Potential ApplicationsThe illustration above depicts the outputs of the COVID-19Surge (CDC) tool of the COVID-19 Modeling toolbox.A hospital systems administrator needs a simple model to project the number of patients the hospitals in the network will need to accommodate in the next 90 days due to COVID-19. You know the population served by each hospital, the date and level of current social distancing, the number of people who have recovered, and the number of patients that are currently hospitalized with COVID-19 in each facility. Using your hospital point layer, you run the CHIME Model v1.1.5 tool.An aid agency wants to estimate where and when resources will be required in the counties you serve. You know the population and number of COVID-19 cases today and 14 days ago in each county. You run the COVID-19Surge (CDC) tool using your county polygon data, introducing an Intervention Policy and New Infections Per Case (R0) driven by fields to account for differences in anticipated social distancing policies and effectiveness between counties.A county wants to understand how the lessening or removal of interventions may impact hospital bed availability within the county. You run the CHIME Model v1.1.5 and COVID-19Surge (CDC) tool, checking Add Additional Web App Fields in Summary in both tools. You display the published results from each tool in the Capacity Analysis configurable app so estimates can be compared between models.This toolbox requires any license of ArcGIS Pro 2.3 or higher in order to run. Steps for upgrading ArcGIS Pro can be found here.For questions, comments and support, please visit our COVID-19 GeoNet community.

  7. A

    Second DIHARD Challenge Evaluation - Eleven Sources

    • abacus.library.ubc.ca
    iso, txt
    Updated Jul 7, 2022
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abacus Data Network (2022). Second DIHARD Challenge Evaluation - Eleven Sources [Dataset]. https://abacus.library.ubc.ca/dataset.xhtml?persistentId=hdl:11272.1/AB2/ML7KD5
    Explore at:
    iso(1283545088), txt(1308)Available download formats
    Dataset updated
    Jul 7, 2022
    Dataset provided by
    Abacus Data Network
    Description

    AbstractIntroduction Second DIHARD Challenge Evaluation - Eleven Sources was developed by the Linguistic Data Consortium (LDC) and contains approximately 20 hours of English and Chinese speech data along with corresponding annotations used in support of the Second DIHARD Challenge. The DIHARD Challenges are a set of shared tasks on diarization focusing on "hard" diarization; that is, speech diarization for challenging corpora where there was an expectation that existing state-of-the-art systems would fare poorly. As with the first challenge, the second development and evaluation sets were drawn from a diverse sampling of sources including monologues, map task dialogues, broadcast interviews, sociolinguistic interviews, meeting speech, speech in restaurants, clinical recordings, extended child language acquisition recordings, and YouTube videos. Data This release, when combined with Second DIHARD Challenge Evaluation - SEEDLingS (forthcoming from LDC), contains the evaluation set audio data and annotation, except for CHiME-5 audio files, which must be obtained from the University of Sheffield. Data sources in this release are as follows (all sources are in English unless otherwise indicated): Autism Diagnosis Observation Schedule (ADOS) interviews Conversations in Restaurants DCIEM/HCRC map task (LDC96S38) Audiobook recordings from LibriVox Meeting speech from 2004 Spring NIST Rich Transcription (RT-04S) Development (LDC2007S11) and Evaluation (LDC2007S12) releases Meeting speech collected by LDC in 2001 for the ROAR project (see, e.g., ISL Meeting Speech Part 1 (LDC2004S05)) 2001 U.S. Supreme Court oral arguments Sociolinguistic interviews from SLX Corpus of Classic Sociolinguistic Interviews (LDC2003T15) Mixer 6 Speech (LDC2013S03) English and Chinese video collected by LDC as part of the Video Annotation for Speech Technologies (VAST) project YouthPoint radio interviews All audio is provided in the form of 16 kHz, 16-bit, mono-channel FLAC files. The diarization for each recording is stored as a NIST Rich Transcription Time Marked (RTTM) file. RTTM files are space-separated text files containing one turn per line. Segmentation files are stored as HTK label files. Each of these files contains one speech segment per line. Scoring regions for each recording are specific by un-partitioned evaluation map (UEM) files. All annotation file types are encoded as UTF-8. More information about file formats, data sources and domains is contained in the included documentation.

  8. c

    Seth Thomas Northbury 1W Wind Dome Mahogany Beehive Mantel Clock 5 Chime

    • clock.collectionhero.com
    html
    Updated Jun 28, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2025). Seth Thomas Northbury 1W Wind Dome Mahogany Beehive Mantel Clock 5 Chime [Dataset]. https://clock.collectionhero.com/view_item.php?id=65425
    Explore at:
    htmlAvailable download formats
    Dataset updated
    Jun 28, 2025
    Time period covered
    Aug 9, 2022
    Description

    Seth Thomas Northbury 1W Wind Dome Mahogany Beehive Mantel Clock 5 Chime - Sold on eBay Aug 09, 2022 for $420.00 - Historical sales data for collectible reference.

  9. A

    Second DIHARD Challenge Development - SEEDLingS

    • abacus.library.ubc.ca
    iso, txt
    Updated Aug 17, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abacus Data Network (2023). Second DIHARD Challenge Development - SEEDLingS [Dataset]. https://abacus.library.ubc.ca/dataset.xhtml;jsessionid=0acf4fc6bb41bcd687ca921ab796?persistentId=hdl%3A11272.1%2FAB2%2FPKMDCL&version=&q=&fileTypeGroupFacet=%22Archive%22&fileAccess=
    Explore at:
    iso(125206528), txt(1308)Available download formats
    Dataset updated
    Aug 17, 2023
    Dataset provided by
    Abacus Data Network
    Description

    AbstractIntroduction Second DIHARD Challenge Development - SEEDLinGS was developed by Duke University and LDC and contains approximately two hours of English child language recordings along with corresponding annotations used in support of the Second DIHARD Challenge. This release, when combined with Second DIHARD Challenge Development - Eleven Sources (LDC2021S10), contains the development set audio data and annotation, except for CHiME-5 audio files, which must be obtained from the University of Sheffield. The DIHARD Challenges are a set of shared tasks on diarization focusing on "hard" diarization; that is, speech diarization for challenging corpora where there was an expectation that existing state-of-the-art systems would fare poorly. As with the first challenge, the second development and evaluation sets were drawn from a diverse sampling of sources including monologues, map task dialogues, broadcast interviews, sociolinguistic interviews, meeting speech, speech in restaurants, clinical recordings, extended child language acquisition recordings, and YouTube videos. Data Source data is from the SEEDLingS (The Study of Environmental Effects on Developing Linguistic Skills) corpus, designed to investigate how infants' early linguistic and environmental input plays a role in their learning. Recordings were generated in the home environment of infants in the Rochester, New York area. A subset of that data was annotated by LDC for use in the First and Second DIHARD Challenges. The data in this release consists of files provided in the Second DIHARD Challenge as well as subsequently updated annotated files not provided to second challenge participants. All audio is provided in the form of 16 kHz, 16-bit, mono-channel FLAC files. The diarization for each recording is stored as a NIST Rich Transcription Time Marked (RTTM) file. RTTM files are space-separated text files containing one turn per line. Segmentation files are stored as HTK label files. Each of these files contains one speech segment per line. Scoring regions for each recording are specific by un-partitioned evaluation map (UEM) files. All annotation file types are encoded as UTF-8. More information about the file formats and data sources and domains are in the included documentation.

  10. A

    Second DIHARD Challenge Evaluation - SEEDLingS

    • abacus.library.ubc.ca
    iso, txt
    Updated Jul 24, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Abacus Data Network (2023). Second DIHARD Challenge Evaluation - SEEDLingS [Dataset]. https://abacus.library.ubc.ca/dataset.xhtml;jsessionid=f090785ff3348a5967cceb6098ba?persistentId=hdl%3A11272.1%2FAB2%2FCXOTQ3&version=&q=&fileTypeGroupFacet=%22Text%22&fileAccess=
    Explore at:
    txt(1308), iso(125566976)Available download formats
    Dataset updated
    Jul 24, 2023
    Dataset provided by
    Abacus Data Network
    Description

    AbstractIntroduction Second DIHARD Challenge Evaluation - SEEDLingS was developed by Duke University and the Linguistic Data Consortium (LDC) and contains approximately two hours of English child language recordings along with corresponding annotations used in support of the Second DIHARD Challenge. The DIHARD Challenges are a set of shared tasks on diarization focusing on "hard" diarization; that is, speech diarization for challenging corpora where there was an expectation that existing state-of-the-art systems would fare poorly. As with the first challenge, the second development and evaluation sets were drawn from a diverse sampling of sources including monologues, map task dialogues, broadcast interviews, sociolinguistic interviews, meeting speech, speech in restaurants, clinical recordings, extended child language acquisition recordings, and web videos. Data Source data is from the SEEDLingS (The Study of Environmental Effects on Developing Linguistic Skills) corpus, designed to investigate how infants' early linguistic and environmental input plays a role in their learning. Recordings were generated in the home environment of infants in the Rochester, New York area. A subset of that data was annotated by LDC for use in the First and Second DIHARD Challenges. This release, when combined with Second DIHARD Challenge Evaluation - Eleven Sources (LDC2022S06), contains the evaluation set audio data and annotation, except for CHiME-5 audio files which must be obtained from the University of Sheffield. All audio is provided in the form of 16 kHz, 16-bit, mono-channel FLAC files. The diarization for each recording is stored as a NIST Rich Transcription Time Marked (RTTM) file. RTTM files are space-separated text files containing one turn per line. Segmentation files are stored as HTK label files. Each of these files contains one speech segment per line. Scoring regions for each recording are specific by un-partitioned evaluation map (UEM) files. All annotation file types are encoded as UTF-8. More information about file formats, data sources and domains is contained in the included documentation.

  11. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Jon Barker; Shinji Watanabe; Emmanuel Vincent; Jan Trmal (2021). CHiME-5 Dataset [Dataset]. https://paperswithcode.com/dataset/chime-5

CHiME-5 Dataset

CHiME Speech Separation and Recognition Challenge

Explore at:
Dataset updated
Nov 16, 2021
Authors
Jon Barker; Shinji Watanabe; Emmanuel Vincent; Jan Trmal
Description

The CHiME challenge series aims to advance robust automatic speech recognition (ASR) technology by promoting research at the interface of speech and language processing, signal processing , and machine learning.

Search
Clear search
Close search
Google apps
Main menu