4 datasets found
  1. f

    Synthetic vowels of speakers with Parkinson’s disease and Parkinsonism

    • figshare.com
    zip
    Updated Jun 1, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jan Hlavnička; Roman Čmejla; Jiří Klempíř; Evžen Růžička; Jan Rusz (2023). Synthetic vowels of speakers with Parkinson’s disease and Parkinsonism [Dataset]. http://doi.org/10.6084/m9.figshare.7628819.v1
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jun 1, 2023
    Dataset provided by
    figshare
    Authors
    Jan Hlavnička; Roman Čmejla; Jiří Klempíř; Evžen Růžička; Jan Rusz
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The dataset contains synthesized replicas of sustained vowels /A/ and /I/ performed by healthy controls, patients with Parkinson’s disease, multiple system atrophy and progressive supranuclear palsy. The dataset can be used as a reference for evaluation of pitch detectors, detectors of modal fundamental frequency, and detectors of subharmonics.Coding system

    Each recording is named by a unique alphanumeric code in the format Uvxy, where U means abbreviation of the group (HC = healthy control, PD = Parkinson’s disease, MSA = multiple system atrophy, PSP = progressive supranuclear palsy) in upper case characters, v is the numeric identifier of the subject within the group, x denotes type of vowel (a = vowel /A/, i = vowel /I/), and y is the number of repetition. The part of the code of U and v uniquely determine each speaker, whereas x and y determine speaker’s recordings.

    Recordings

    All recordings are briefly described in table dataset.csv. All files of each record (see records.zip) are identified by the corresponding code and suffix. Suffix describe type of the file and is separated from the code by underscore. Naming of the files is illustrated on the record HC8a1. The code describes first repetition of the vowel /A/ performed by healthy speaker HC8. The record HC8a1 consists of following files:HC8a1.wav = waveform of the synthesized replica. This is the reference signal used for the evaluation. Parameters of jitter, shimmer and harmonic to noise ratio (HNR) can be found in dataset.csv.HC8a1_clean.wav = waveform of the synthesized replica without added noise. We provide this signal to make the model more versatile. Authors may add a different kind of noise to this signal or manipulate with HNR. Note that that both signals required normalization prior to writing into wav-file. Original scaling factor between HC8a1 and HCa1_clean can be determined from total power of signals and reference HNR value.HC8a1_LF.wav = sample of the glottal pulse used for the synthesis.HC8a2_impulses.csv = list of impulses’ locations in seconds and corresponding amplitudes. The position of pulses was corrected to match with the beginning of glottal pulse, i.e., first sample of the signal HC8a1_LF.wav begins at each of these positions. The jitter and shimmer listed in dataset.csv were are median values. Jitter and shimmer by other definitions can be calculated from positions and amplitudes of pulses provided by this file. HC8a1_subharmonics.csv = list of subharmonic intervals described by the start time in seconds and end time in seconds. Corresponding index of amplitude modulation expressed as SHR in percent can be found in the table dataset.csv. When no subharmonic was determined by the supervised parameterization, no file was included for the speaker and SHR in the dataset.csv was set to zero [1].

    [1] Note that the supervised detection had lower sensitivity due to senzitivity of pitch trace in PRAAT, so the occurrence of subharmonics in synthesized data is much lower than in the original dataset analyzed by automated segmentation. This is not a problem because subharmonics were synthesized only at the given interval - this illustrates why it is important to detect subharmonics in other way than by pitch.

  2. i

    Data from: Italian Parkinson's Voice and Speech

    • ieee-dataport.org
    Updated Oct 17, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Giovanni Dimauro (2024). Italian Parkinson's Voice and Speech [Dataset]. https://ieee-dataport.org/open-access/italian-parkinsons-voice-and-speech
    Explore at:
    Dataset updated
    Oct 17, 2024
    Authors
    Giovanni Dimauro
    Description

    I would be grateful if you cite my two following papers:

  3. i

    Code for Intra-subject Enveloped Deep Sample Fuzzy Ensemble Learning...

    • ieee-dataport.org
    Updated Sep 13, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Yongming Li (2022). Code for Intra-subject Enveloped Deep Sample Fuzzy Ensemble Learning Algorithm of Speech Data of Parkinson's Disease [Dataset]. https://ieee-dataport.org/documents/code-intra-subject-enveloped-deep-sample-fuzzy-ensemble-learning-algorithm-speech-data
    Explore at:
    Dataset updated
    Sep 13, 2022
    Authors
    Yongming Li
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    a self-collected Parkinson's speech dataset

  4. P

    NeuroVoz Dataset

    • paperswithcode.com
    Updated Mar 3, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). NeuroVoz Dataset [Dataset]. https://paperswithcode.com/dataset/neurovoz
    Explore at:
    Dataset updated
    Mar 3, 2024
    Description

    The NeuroVoz dataset emerges as a pioneering resource in the field of computational linguistics and biomedical research, specifically designed to enhance the diagnosis and understanding of Parkinson's Disease (PD) through speech analysis. This dataset is distinguished as the first of its kind to be made publicly available in Castilian Spanish, addressing a critical gap in the availability of linguistic and dialectical diversity within PD research.

    Compiled from a cohort of 112 participants, including 54 individuals diagnosed with PD and 58 healthy controls, the NeuroVoz dataset offers a rich compilation of speech recordings. All PD participants were recorded under medication (ON state), ensuring consistency and reliability in the speech samples collected. The dataset is meticulously curated to include a variety of speech tasks—ranging from sustained vowel phonations and diadochokinetic (DDK) tests to 16 structured listen-and-repeat utterances and spontaneous monologues. The inclusion of both manually transcribed listen-and-repeat tasks and Whisper-automated transcriptions for monologues underscores our commitment to data accuracy and usability.

    Encompassing 2,977 audio files, the NeuroVoz dataset provides an extensive repository, averaging 26.88 +- 3.35 recordings per participant, making it an invaluable asset for researchers seeking to explore the nuances of PD-affected speech. The dataset's structure and composition facilitate a multifaceted analysis of speech impairments associated with PD, offering insights into phonatory, articulatory, and prosodic changes.

    In contributing to the body of knowledge with the NeuroVoz dataset, we invite the scientific community to engage with this dataset, explore the specific speech characteristics of PD in Castilian Spanish speakers, and advance the field of PD diagnosis through innovative speech analysis techniques.

    Paper explaining the database: https://arxiv.org/abs/2403.02371
    Zenodo repository with the database: https://zenodo.org/records/13647600
    Github repo on how to use the dataset: https://github.com/BYO-UPM/Neurovoz_Dababase

  5. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Jan Hlavnička; Roman Čmejla; Jiří Klempíř; Evžen Růžička; Jan Rusz (2023). Synthetic vowels of speakers with Parkinson’s disease and Parkinsonism [Dataset]. http://doi.org/10.6084/m9.figshare.7628819.v1

Synthetic vowels of speakers with Parkinson’s disease and Parkinsonism

Explore at:
zipAvailable download formats
Dataset updated
Jun 1, 2023
Dataset provided by
figshare
Authors
Jan Hlavnička; Roman Čmejla; Jiří Klempíř; Evžen Růžička; Jan Rusz
License

Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically

Description

The dataset contains synthesized replicas of sustained vowels /A/ and /I/ performed by healthy controls, patients with Parkinson’s disease, multiple system atrophy and progressive supranuclear palsy. The dataset can be used as a reference for evaluation of pitch detectors, detectors of modal fundamental frequency, and detectors of subharmonics.Coding system

Each recording is named by a unique alphanumeric code in the format Uvxy, where U means abbreviation of the group (HC = healthy control, PD = Parkinson’s disease, MSA = multiple system atrophy, PSP = progressive supranuclear palsy) in upper case characters, v is the numeric identifier of the subject within the group, x denotes type of vowel (a = vowel /A/, i = vowel /I/), and y is the number of repetition. The part of the code of U and v uniquely determine each speaker, whereas x and y determine speaker’s recordings.

Recordings

All recordings are briefly described in table dataset.csv. All files of each record (see records.zip) are identified by the corresponding code and suffix. Suffix describe type of the file and is separated from the code by underscore. Naming of the files is illustrated on the record HC8a1. The code describes first repetition of the vowel /A/ performed by healthy speaker HC8. The record HC8a1 consists of following files:HC8a1.wav = waveform of the synthesized replica. This is the reference signal used for the evaluation. Parameters of jitter, shimmer and harmonic to noise ratio (HNR) can be found in dataset.csv.HC8a1_clean.wav = waveform of the synthesized replica without added noise. We provide this signal to make the model more versatile. Authors may add a different kind of noise to this signal or manipulate with HNR. Note that that both signals required normalization prior to writing into wav-file. Original scaling factor between HC8a1 and HCa1_clean can be determined from total power of signals and reference HNR value.HC8a1_LF.wav = sample of the glottal pulse used for the synthesis.HC8a2_impulses.csv = list of impulses’ locations in seconds and corresponding amplitudes. The position of pulses was corrected to match with the beginning of glottal pulse, i.e., first sample of the signal HC8a1_LF.wav begins at each of these positions. The jitter and shimmer listed in dataset.csv were are median values. Jitter and shimmer by other definitions can be calculated from positions and amplitudes of pulses provided by this file. HC8a1_subharmonics.csv = list of subharmonic intervals described by the start time in seconds and end time in seconds. Corresponding index of amplitude modulation expressed as SHR in percent can be found in the table dataset.csv. When no subharmonic was determined by the supervised parameterization, no file was included for the speaker and SHR in the dataset.csv was set to zero [1].

[1] Note that the supervised detection had lower sensitivity due to senzitivity of pitch trace in PRAAT, so the occurrence of subharmonics in synthesized data is much lower than in the original dataset analyzed by automated segmentation. This is not a problem because subharmonics were synthesized only at the given interval - this illustrates why it is important to detect subharmonics in other way than by pitch.

Search
Clear search
Close search
Google apps
Main menu