5 datasets found
  1. Bimodal dataset on Inner speech

    • openneuro.org
    Updated May 23, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson (2023). Bimodal dataset on Inner speech [Dataset]. http://doi.org/10.18112/openneuro.ds004196.v2.0.0
    Explore at:
    Dataset updated
    May 23, 2023
    Dataset provided by
    OpenNeurohttps://openneuro.org/
    Authors
    Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Bimodal dataset on Inner Speech

    Code available: https://github.com/LTU-Machine-Learning/Inner_Speech_EEG_FMRI

    Publication available: https://www.biorxiv.org/content/10.1101/2022.05.24.492109v3

    Abstract: The recognition of inner speech, which could give a `voice' to patients that have no ability to speak or move, is a challenge for brain-computer interfaces (BCIs). A shortcoming of the available datasets is that they do not combine modalities to increase the performance of inner speech recognition. Multimodal datasets of brain data enable the fusion of neuroimaging modalities with complimentary properties, such as the high spatial resolution of functional magnetic resonance imaging (fMRI) and the temporal resolution of electroencephalography (EEG), and therefore are promising for decoding inner speech. This paper presents the first publicly available bimodal dataset containing EEG and fMRI data acquired nonsimultaneously during inner-speech production. Data were obtained from four healthy, right-handed participants during an inner-speech task with words in either a social or numerical category. Each of the 8-word stimuli were assessed with 40 trials, resulting in 320 trials in each modality for each participant.
    The aim of this work is to provide a publicly available bimodal dataset on inner speech, contributing towards speech prostheses.

    Short Dataset description: The dataset consists of 1280 trials in each modality (EEG, FMRI). The stimuli contain 8 words, selected from 2 different categories (social, numeric): Social: child, daughter, father, wife Numeric: four, three, ten, six

    There are 4 subjects in total: sub-01, sub-02, sub-03, sub-05. Initially, there were 5 participants, however, sub-04 data was rejected due to high fluctuations. Details of valid data are available in the file participants.tsv.

    For questions please contact: foteini.liwicki@ltu.se

  2. Metadata record for: Thinking out loud: an open-access EEG-based BCI dataset...

    • springernature.figshare.com
    txt
    Updated May 31, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Scientific Data Curation Team (2023). Metadata record for: Thinking out loud: an open-access EEG-based BCI dataset for inner speech recognition [Dataset]. http://doi.org/10.6084/m9.figshare.16783987.v1
    Explore at:
    txtAvailable download formats
    Dataset updated
    May 31, 2023
    Dataset provided by
    Figsharehttp://figshare.com/
    Authors
    Scientific Data Curation Team
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    This dataset contains key characteristics about the data described in the Data Descriptor Thinking out loud: an open-access EEG-based BCI dataset for inner speech recognition. Contents:

        1. human readable metadata summary table in CSV format
    
    
        2. machine readable metadata file in JSON format
    
  3. FMRI-EEG Prestudy

    • openneuro.org
    Updated Aug 18, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson (2022). FMRI-EEG Prestudy [Dataset]. http://doi.org/10.18112/openneuro.ds004196.v1.0.0
    Explore at:
    Dataset updated
    Aug 18, 2022
    Dataset provided by
    OpenNeurohttps://openneuro.org/
    Authors
    Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Original Paper: Foteini Simistira Liwicki, Vibha Gupta, Rajkumar Saini, Kanjar De, Nosheen Abid, Sumit Rakesh, Scott Wellington, Holly Wilson, Marcus Liwicki, Johan Eriksson (2022), Bimodal pilot study on inner speech decoding reveals the potential of combining EEG and fMRI, Url: https://www.biorxiv.org/content/10.1101/2022.05.24.492109v1

    Overview: The dataset involves a study for inner speech detection using bimodal electroencephalography (EEG) and Functional Magnetic Resonance Imaging (FMRI). The dataset consits of 1280 trials from 4 adult healthy subjects (2 male + 2 female) for each modality. The dataset consits of 1280 trials in each modality (EEG + FMRI). The stimulus consists of 4 words from 2 categories: 1. Social: child, daughter, father, wife 2. Numeric: four, three, ten, six

    There are four subjects sub01, sub02, sub03, sub05 files. Intitially there were 5 participants, however sub04 data was rejected due to high gluctuations. Details of valid data are available at participants.tsv file.

    Experimental paradigm: To Do

    EEG DATA: RAW EEG Data Device: BioSemi Active2 EEG data was sampled

    FMRI DATA:

    All the data was acquired via SIEMENS MAGNETOM Prisma All DICOM files were converted to NFTII using MRIcroGL 1.2.20211006 x86-64 FPC which uses dcm2niix for the conversion. FMRI task event file for each subject session conatins the following columns Innerword - Can be any of the 8 words child, daughter, father, wife, four, three, ten, six Condition - Can be Social or Numeric Stimulus Onset time in Seconds Stimulus Offset time in Seconds

    Task Events are available in subX_seesionY.tsv where X stands for subject number 1,2,3 and 5 and Y stands for session number 1 or 2.

    For FMRI BOLD files the following parameters were used. Slices:68 TR:2160 ms TE:30.00 ms

    For FMRI field maps GRE Field maps: TR: 662.0 ms TE1: 4.92ms TE2: 7.38ms

  4. Inner Speech

    • openneuro.org
    Updated Apr 17, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Nieto Nicolás; Victoria Peterson; Hugo Rufiner; Juan Kamienkowski; Ruben Spies (2021). Inner Speech [Dataset]. http://doi.org/10.18112/openneuro.ds003626.v1.0.0
    Explore at:
    Dataset updated
    Apr 17, 2021
    Dataset provided by
    OpenNeurohttps://openneuro.org/
    Authors
    Nieto Nicolás; Victoria Peterson; Hugo Rufiner; Juan Kamienkowski; Ruben Spies
    License

    CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
    License information was derived automatically

    Description

    Inner Speech Dataset.

    Author: Nieto Nicolás

    Code available at https://github.com/N-Nieto/Inner_Speech_Dataset

    Abstract: Surface electroencephalography is a standard and noninvasive way to measure electrical brain activity. Recent advances in artificial intelligence led to significant improvements in the automatic detection of brain patterns, allowing increasingly faster, more reliable and accessible Brain-Computer Interfaces. Different paradigms have been used to enable the human-machine interaction and the last few years have broad a mark increase in the interest for interpreting and characterizing the "inner voice" phenomenon. This paradigm, called inner speech, raises the possibility of executing an order just by thinking about it, allowing a “natural” way of controlling external devices. Unfortunately, the lack of publicly available electroencephalography datasets, restricts the development of new techniques for inner speech recognition. A ten-subjects dataset acquired under this and two others related paradigms, obtain with an acquisition systems of 136 channels, is presented. The main purpose of this work is to provide the scientific community with an open-access multiclass electroencephalography database of inner speech commands that could be used for better understanding of the related brain mechanisms.

    Conditions = Inner Speech, Pronounced Speech, Visualized Condition

    Classes = "Arriba/Up", "Abajo/Down", "Derecha/Right", "Izquierda/Left"

    Total Trials = 5640

    Please contact us at this e-mail address if you have any doubts: nnieto@sinc.unl.edu.ar

  5. S

    An open dataset of multidimensional signals based on different speech...

    • scidb.cn
    Updated May 9, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Ran Zhao; Yanru Bai; Shuming Zhang; Jianing Zhu; Hongxing Liu; Guangjian Ni (2025). An open dataset of multidimensional signals based on different speech patterns in pragmatic Mandarin [Dataset]. http://doi.org/10.57760/sciencedb.24416
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    May 9, 2025
    Dataset provided by
    Science Data Bank
    Authors
    Ran Zhao; Yanru Bai; Shuming Zhang; Jianing Zhu; Hongxing Liu; Guangjian Ni
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Speech is essential for human communication, but millions of people lose the ability to speak due to conditions such as amyotrophic lateral sclerosis (ALS) or stroke. Assistive technologies like brain-computer interfaces (BCIs), which can convert brain signals into speech, offer hope for these patients. However, these technologies still face challenges in decoding accuracy. This issue is especially challenging for tonal languages like Mandarin Chinese. Their processing requires phoneme encoding and precise tonal information handling, which complicates the decoding of brain signals. Furthermore, most existing speech datasets are based on Indo-European languages, which hinders our understanding of how tonal information is encoded in the brain. To address this, we introduce a comprehensive open dataset, which includes multimodal signals from 30 subjects using Mandarin Chinese across overt, silent, and imagined speech modes, covering electroencephalogram (EEG), surface electromyogram (sEMG), and speech recordings. Unlike many datasets that focus on a single speech mode, this one integrates three speech modes, providing a more comprehensive view of speech-related activity. Incorporating Mandarin facilitates an in-depth examination of the inner mechanisms that encode tonal variations and their interaction with motor and auditory speech representations. This is crucial for enhancing tonal language decoding in BCIs. Beyond BCI applications, this dataset lays a valuable groundwork for exploring the neural encoding of tonal languages, investigating tone-related brain dynamics, and improving assistive communication strategies. It supports cross-linguistic speech processing research and contributes to data-driven neural speech decoding technology innovations.

  6. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson (2023). Bimodal dataset on Inner speech [Dataset]. http://doi.org/10.18112/openneuro.ds004196.v2.0.0
Organization logo

Bimodal dataset on Inner speech

Explore at:
2 scholarly articles cite this dataset (View in Google Scholar)
Dataset updated
May 23, 2023
Dataset provided by
OpenNeurohttps://openneuro.org/
Authors
Foteini Liwicki; Vibha Gupta; Rajkumar Saini; Kanjar De; Nosheen Abid; Sumit Rakesh; Scott Wellington; Holly Wilson; Marcus Liwicki; Johan Eriksson
License

CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically

Description

Bimodal dataset on Inner Speech

Code available: https://github.com/LTU-Machine-Learning/Inner_Speech_EEG_FMRI

Publication available: https://www.biorxiv.org/content/10.1101/2022.05.24.492109v3

Abstract: The recognition of inner speech, which could give a `voice' to patients that have no ability to speak or move, is a challenge for brain-computer interfaces (BCIs). A shortcoming of the available datasets is that they do not combine modalities to increase the performance of inner speech recognition. Multimodal datasets of brain data enable the fusion of neuroimaging modalities with complimentary properties, such as the high spatial resolution of functional magnetic resonance imaging (fMRI) and the temporal resolution of electroencephalography (EEG), and therefore are promising for decoding inner speech. This paper presents the first publicly available bimodal dataset containing EEG and fMRI data acquired nonsimultaneously during inner-speech production. Data were obtained from four healthy, right-handed participants during an inner-speech task with words in either a social or numerical category. Each of the 8-word stimuli were assessed with 40 trials, resulting in 320 trials in each modality for each participant.
The aim of this work is to provide a publicly available bimodal dataset on inner speech, contributing towards speech prostheses.

Short Dataset description: The dataset consists of 1280 trials in each modality (EEG, FMRI). The stimuli contain 8 words, selected from 2 different categories (social, numeric): Social: child, daughter, father, wife Numeric: four, three, ten, six

There are 4 subjects in total: sub-01, sub-02, sub-03, sub-05. Initially, there were 5 participants, however, sub-04 data was rejected due to high fluctuations. Details of valid data are available in the file participants.tsv.

For questions please contact: foteini.liwicki@ltu.se

Search
Clear search
Close search
Google apps
Main menu