2 datasets found
  1. Data Corpus for the IEEE-AASP Challenge on Acoustic Source Localization and...

    • zenodo.org
    • live.european-language-grid.eu
    • +1more
    zip
    Updated Sep 23, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Christine Evers; Christine Evers; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss; Patrick A. Naylor; Patrick A. Naylor; Walter Kellermann; Walter Kellermann; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss (2020). Data Corpus for the IEEE-AASP Challenge on Acoustic Source Localization and Tracking (LOCATA) [Dataset]. http://doi.org/10.5281/zenodo.3630471
    Explore at:
    zipAvailable download formats
    Dataset updated
    Sep 23, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Christine Evers; Christine Evers; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss; Patrick A. Naylor; Patrick A. Naylor; Walter Kellermann; Walter Kellermann; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss
    License

    Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
    License information was derived automatically

    Description

    This repository contains the final release of the development and evaluation datasets for the LOCATA Challenge.

    The challenge of sound source localization in realistic environments has attracted widespread attention in the Audio and Acoustic Signal Processing (AASP) community in recent years. Source localization approaches in the literature address the estimation of positional information about acoustic sources using a pair of microphones, microphone arrays, or networks with distributed acoustic sensors. The IEEE AASP Challenge on acoustic source LOCalization And TrAcking (LOCATA) aimed at providing researchers in source localization and tracking with a framework to objectively benchmark results against competing algorithms using a common, publicly released data corpus that encompasses a range of realistic scenarios in an enclosed acoustic environment.

    Four different microphone arrays were used for the recordings, namely:

    • Planar array with 15 channels (DICIT array) containing uniform linear sub-arrays
    • Spherical array with 32 channels (Eigenmike)
    • Pseudo-spherical array with 12-channels (robot head)
    • Hearing aid dummies on a dummy head (2-channel per hearing aid).

    An optical tracking system (OptiTrack) was used to record the positions and orientations of talker, loudspeakers and microphone arrays. Moreover, the emitted source signals were recorded to determine voice activity periods in the recorded signals for each source separately. The ground truth values are compared to the estimated values submitted by the participants using several criteria to evaluate the accuracy of the estimated directions of arrival and track-to-source association.

    The datasets encompass the following six, increasingly challenging, scenarios:

    • Task 1: Localization of a single, static loudspeaker using static microphones arrays
    • Task 2: Multi-source localization of static loudspeakers using static microphone arrays
    • Task 3: Localization of a single, moving talker using static microphone arrays
    • Task 4: Localization of multiple, moving talkers using static microphone arrays
    • Task 5: Localization of a single, moving talker using moving microphone arrays
    • Task 6: Multi-source localization of moving talkers using moving microphone arrays.

    The development and evaluation datasets in this repository contain the following data:

    • Close-talking speech signals for human talkers, recorded use DPA microphones
    • Distant-talking recordings using four microphone arrays:
      • Spherical Eigenmike (32 channels)
      • Pseudo-spherical prototype NAO robot (12 channels)
      • Planar DICIT array (15 channels)
      • Hearing aids installed in a head-torso simulator (4 channels)
    • Ground-truth annotations of all source and microphone positions, obtained using an OptiTrack system of infrared cameras. The ground-truth positions are provided at the frame rate of the optical tracking system

    The following software is provided with the data:

    For further information, see:

  2. P

    Data from: LOCATA Dataset

    • paperswithcode.com
    • opendatalab.com
    Updated Jun 15, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Heinrich W. Löllmann; Christine Evers; Alexander Schmidt; Heinrich Mellmann; Hendrik Barfuss; Patrick A. Naylor; Walter Kellermann (2020). LOCATA Dataset [Dataset]. https://paperswithcode.com/dataset/locata
    Explore at:
    Dataset updated
    Jun 15, 2020
    Authors
    Heinrich W. Löllmann; Christine Evers; Alexander Schmidt; Heinrich Mellmann; Hendrik Barfuss; Patrick A. Naylor; Walter Kellermann
    Description

    The LOCATA dataset is a dataset for acoustic source localization. It consists of real-world ambisonic speech recordings with optically tracked azimuth-elevation labels.

  3. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Christine Evers; Christine Evers; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss; Patrick A. Naylor; Patrick A. Naylor; Walter Kellermann; Walter Kellermann; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss (2020). Data Corpus for the IEEE-AASP Challenge on Acoustic Source Localization and Tracking (LOCATA) [Dataset]. http://doi.org/10.5281/zenodo.3630471
Organization logo

Data Corpus for the IEEE-AASP Challenge on Acoustic Source Localization and Tracking (LOCATA)

Explore at:
2 scholarly articles cite this dataset (View in Google Scholar)
zipAvailable download formats
Dataset updated
Sep 23, 2020
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Christine Evers; Christine Evers; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss; Patrick A. Naylor; Patrick A. Naylor; Walter Kellermann; Walter Kellermann; Heinrich Loellmann; Heinrich Mellmann; Alexander Schmidt; Hendrik Barfuss
License

Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
License information was derived automatically

Description

This repository contains the final release of the development and evaluation datasets for the LOCATA Challenge.

The challenge of sound source localization in realistic environments has attracted widespread attention in the Audio and Acoustic Signal Processing (AASP) community in recent years. Source localization approaches in the literature address the estimation of positional information about acoustic sources using a pair of microphones, microphone arrays, or networks with distributed acoustic sensors. The IEEE AASP Challenge on acoustic source LOCalization And TrAcking (LOCATA) aimed at providing researchers in source localization and tracking with a framework to objectively benchmark results against competing algorithms using a common, publicly released data corpus that encompasses a range of realistic scenarios in an enclosed acoustic environment.

Four different microphone arrays were used for the recordings, namely:

  • Planar array with 15 channels (DICIT array) containing uniform linear sub-arrays
  • Spherical array with 32 channels (Eigenmike)
  • Pseudo-spherical array with 12-channels (robot head)
  • Hearing aid dummies on a dummy head (2-channel per hearing aid).

An optical tracking system (OptiTrack) was used to record the positions and orientations of talker, loudspeakers and microphone arrays. Moreover, the emitted source signals were recorded to determine voice activity periods in the recorded signals for each source separately. The ground truth values are compared to the estimated values submitted by the participants using several criteria to evaluate the accuracy of the estimated directions of arrival and track-to-source association.

The datasets encompass the following six, increasingly challenging, scenarios:

  • Task 1: Localization of a single, static loudspeaker using static microphones arrays
  • Task 2: Multi-source localization of static loudspeakers using static microphone arrays
  • Task 3: Localization of a single, moving talker using static microphone arrays
  • Task 4: Localization of multiple, moving talkers using static microphone arrays
  • Task 5: Localization of a single, moving talker using moving microphone arrays
  • Task 6: Multi-source localization of moving talkers using moving microphone arrays.

The development and evaluation datasets in this repository contain the following data:

  • Close-talking speech signals for human talkers, recorded use DPA microphones
  • Distant-talking recordings using four microphone arrays:
    • Spherical Eigenmike (32 channels)
    • Pseudo-spherical prototype NAO robot (12 channels)
    • Planar DICIT array (15 channels)
    • Hearing aids installed in a head-torso simulator (4 channels)
  • Ground-truth annotations of all source and microphone positions, obtained using an OptiTrack system of infrared cameras. The ground-truth positions are provided at the frame rate of the optical tracking system

The following software is provided with the data:

For further information, see:

Search
Clear search
Close search
Google apps
Main menu