Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
General description:
This dataset was created in the context of the Pablo project, partially funded by KORG Inc. It contains monophonic recordings of two kind of exercises: single notes and scales. The dataset was reported in the following article:
Romaní Picas O., Parra Rodriguez H., Dabiri D., Tokuda H., Hariya W., Oishi K., & Serra X."A real-time system for measuring sound goodness in instrumental sounds", 138th Audio Engineering Society Convention (2015).
The recordings were made in the Universitat Pompeu Fabra / Phonos recording studio by 15 different professional musicians, all of them holding a music degree and having some expertise in teaching. 12 different instruments were recorded using one or up to 4 different microphones (depending on the recording session). For all the instruments the whole set of playable semitones in the instrument is recorded several times with different tonal characteristics. Each note is recorded into a separate mono .flac audio file of 48kHz and 32 bits. The tonal characteristics are explained both in the the following section and the related publication.
The audio files are organised in one directory for each recording session. In addition to the files, one SQLite database file is included. The structure of the database is related in the following section.
Database description:
The database is meant for organizing the sounds in a handy way. It is organised in four different tables: sounds, takes, packs and ratings.
Sounds
The table containing the sounds annotations.
Takes
A sound can have several takes as some of them were recorded using different microphones at the same time. Each take has an associated audio file.
Packs
A pack is a group of sounds from the same recording session. The audio files are organised in the *sound_files* directory in subfolders with the pack name to which they belong.
Ratings
Some musicians rated some sounds in a 0-10 goodness scale for the user evaluatio of the first project prototype. Please read the paper for more detailed information.
License:
This work is licensed under the Creative Commons Attribution-NonCommercial 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
The Good-sounds dataset includes monophonic audio recordings of two types of musical exercises: single notes and scales. These recordings, performed by 15 professional musicians on flute, clarinet, trumpet, violin, and cello, were made in studio conditions at the Universitat Pompeu Fabra. The dataset supports research in sound quality, music education, and audio classification. It includes audio files in .wav format and metadata in .csv format detailing the exercise type, note, pitch, instrument, and performer, among others.
This dataset contains the ground truth data used to evaluate the musical pitch, tempo and key estimation algorithms developed during the AudioCommons H2020 EU project and which are part of the Audio Commons Audio Extractor tool. It also includes ground truth information for the single-eventness audio descriptor also developed for the same tool. This ground truth data has been used to generate the following documents: Deliverable D4.4: Evaluation report on the first prototype tool for the automatic semantic description of music samples Deliverable D4.10: Evaluation report on the second prototype tool for the automatic semantic description of music samples Deliverable D4.12: Release of tool for the automatic semantic description of music samples All these documents are available in the materials section of the AudioCommons website. All ground truth data in this repository is provided in the form of CSV files. Each CSV file corresponds to one of the individual datasets used in one or more evaluation tasks of the aforementioned deliverables. This repository does not include the audio files of each individual dataset, but includes references to the audio files. The following paragraphs describe the structure of the CSV files and give some notes about how to obtain the audio files in case these would be needed. Structure of the CSV files All CSV files in this repository (with the sole exception of SINGLE EVENT - Ground Truth.csv) feature the following 5 columns: Audio reference: reference to the corresponding audio file. This will either be a string withe the filename, or the Freesound ID (for one dataset based on Freesound content). See below for details about how to obtain those files. Audio reference type: will be one of Filename or Freesound ID, and specifies how the previous column should be interpreted. Key annotation: tonality information as a string with the form "RootNote minor/major". Audio files with no ground truth annotation for tonality are left blank. Ground truth annotations are parsed from the original data source as described in the text of deliverables D4.4 and D4.10. Tempo annotation: tempo information as an integer representing beats per minute. Audio files with no ground truth annotation for tempo are left blank. Ground truth annotations are parsed from the original data source as described in the text of deliverables D4.4 and D4.10. Note that integer values are used here because we only have tempo annotations for music loops which typically only feature integer tempo values. Pitch annotation: pitch information as an integer representing the MIDI note number corresponding to annotated pitch's frequency. Audio files with no ground truth pitch for tempo are left blank. Ground truth annotations are parsed from the original data source as described in the text of deliverables D4.4 and D4.10. The remaining CSV file, SINGLE EVENT - Ground Truth.csv, has only the following 2 columns: Freesound ID: sound ID used in Freesound to identify the audio clip. Single Event: boolean indicating whether the corresponding sound is considered to be a single event or not. Single event annotations were collected by the authors of the deliverables as described in deliverable D4.10. How to get the audio data In this section we provide some notes about how to obtain the audio files corresponding to the ground truth annotations provided here. Note that due to licensing restrictions we are not allowed to re-distribute the audio data corresponding to most of these ground truth annotations. Apple Loops (APPL): This dataset includes some of the music loops included in Apple's music software such as Logic or GarageBand. Access to these loops requires owning a license for the software. Detailed instructions about how to set up this dataset are provided here. Carlos Vaquero Instruments Dataset (CVAQ): This dataset includes single instrument recordings carried out by Carlos Vaquero as part of this master thesis. Sounds are available as Freesound packs and can be downloaded at this page: https://freesound.org/people/Carlos_Vaquero/packs Freesound Loops 4k (FSL4): This dataset set includes a selection of music loops taken from Freesound. Detailed instructions about how to set up this dataset are provided here. Giant Steps Key Dataset (GSKY): This dataset includes a selection of previews from Beatport annotated by key. Audio and original annotations available here. Good-sounds Dataset (GSND): This dataset contains monophonic recordings of instrument samples. Full description, original annotations and audio are available here. University of IOWA Musical Instrument Samples (IOWA): This dataset was created by the Electronic Music Studios of the University of IOWA and contains recordings of instrument samples. The dataset is available upon request by visiting this website. Mixcraft Loops (MIXL): This dataset includes some of the music loops included in Acoustica's Mixcraft music software. Access to thes...
http://opendatacommons.org/licenses/dbcl/1.0/http://opendatacommons.org/licenses/dbcl/1.0/
The dataset is a subset of the Task-2 of DCASE 2020 Challenge. The Challenge is to identify anomaly of a machine using the audio data. There are three different parts of the dataset, namely, training, validation and testing which have been combined into a single dataset.
Training- https://zenodo.org/record/3678171
Validation- https://zenodo.org/record/3727685
Testing- https://zenodo.org/record/3841772
Sound velocity profiles were collected using sound velocimeter in the Great Lakes and Galveston Bay from NOAA NAVIGATION RESPONSE TEAM 4 from 11 April 2006 to 04 August 2006. Data were submitted by Ruby L. Becker of National Ocean Service in Silver Spring, Maryland.
Attribution-NonCommercial-ShareAlike 4.0 (CC BY-NC-SA 4.0)https://creativecommons.org/licenses/by-nc-sa/4.0/
License information was derived automatically
Description
This dataset is a sound dataset for malfunctioning industrial machine investigation and inspection with domain shifts due to changes in operational and environmental conditions (MIMII DUE). The dataset consists of normal and abnormal operating sounds of five different types of industrial machines, i.e., fans, gearboxes, pumps, slide rails, and valves. The data for each machine type includes six subsets called ``sections'', and each section roughly corresponds to a single product. Each section consists of data from two domains, called the source domain and the target domain, with different conditions such as operating speed and environmental noise. This dataset is a subset of the dataset for DCASE 2021 Challenge Task 2, so the dataset is entirely the same as data included in the development dataset and additional training dataset. For more information, please see this paper and the pages of the development dataset and the task description for DCASE 2021 Challenge Task 2.
Baseline system
Two simple baseline systems are available on the Github repositories [URL] and [URL]. The baseline systems provide a simple entry-level approach that gives a reasonable performance in the dataset. They are good starting points, especially for entry-level researchers who want to get familiar with the anomalous-sound-detection task.
Conditions of use
This dataset was made by Hitachi, Ltd. and is available under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International (CC BY-NC-SA 4.0) license.
Publication
If you use this dataset, please cite the following paper:
Ryo Tanabe, Harsh Purohit, Kota Dohi, Takashi Endo, Yuki Nikaido, Toshiki Nakamura, and Yohei Kawaguchi, "MIMII DUE: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection with Domain Shifts due to Changes in Operational and Environmental Conditions," arXiv preprint arXiv: 2105.02702, 2021. [URL]
Feedback
If there is any problem, please contact us:
https://www.gnu.org/licenses/agpl.txthttps://www.gnu.org/licenses/agpl.txt
vibrato_analysis
Vibrato Analysis Dataset,Detection and Parameterization
Introduction
This dataset was created as an outcome of a summer internship with Good-sounds project. Sound recordings were used for vibrato analysis. The dataset contains monophonic recordings of single notes and melodies for four instruments with annotations of four different recording sets regarding some vibrato parameters.
It is organized according to the owner of the recordings. Sounds with vibrato and no vibrato are presented within the folders with their annotations in cvs format. All annotations except than the alto saxophone recordings include derivative analysis of pitch parameters.
One of the recordings for violin was made in the Universitat Pompeu Fabra recording studio by a violist from MTG. Single notes of two octaves and four different melodies starting from different pitch are recorded in wav format sampled at 44.1 kHz. Each sound separated by semitones is recorded four times for no vibrato,slow, standard and fast rates of vibrato. Melodies were played for no vibrato and vibrato at a standard rate to have the attenuation at the end of the note.
One other recording set was taken from Good-sounds project. The only brass instrument within the whole dataset is this one for alto saxophone. It was again recorded in the Universitat Pompeu Fabra / Phonos recording studio. This dataset for alto saxophone not only contains vibrato and non vibrato sounds but also some different tonal characteristics.
Three other recordings are downloaded from a user named Carlos_Vaquero in Freesound. Violin, violoncello and transverse flute recordings were downloaded and used for non commercial purposes under creative commons license. Within the Carlos_Vaquero folder, audio files are separated according to the type of instruments.
Dataset Description
The database is meant for organizing the sounds in a handy way. It is organized according to the creator. In each three datasets, annotations and analysis parameters are available within the csv files and each has 11 field descriptor.
Carlos_Vaquero
Transverse_flute
Violin
Violoncello
Good-sounds-Alto sax
Alto-sax
MTG - Violin
Violin
Except than alto-sax recordings, each contains following parameters:
Peak_Percentage: Percent wise proportion of peaks in first derivation of the pitch trajectory.
Mean_Difference: Mean value of the differences of two consecutive peaks.
Max_Difference: Maximum separation of side by side peaks.
Index_Max (first one): Index value of the maximum peak in the derivative array.
Location_Max (%): Location of the maximum peak in the array, percent wise.
Start_End_Time (seconds): Starting and ending time instants of vibrato in the recording.
Duration: Duration of the vibrato part of the recording.
Documentation
Related report can be found in the GitHub repository as "Vibrato Analysis Internship Report".
License
All the software is distributed with the Affero GPL v3 license.
This dataset provides information about the number of properties, residents, and average property values for Great River Drive cross streets in Sound Beach, NY.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Prevalence of mosquito-borne diseases reported in great sound, bermuda.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
Monthly forecast of mosquito activity risk levels in great sound, bermuda.
These data provide an accurate high-resolution shoreline compiled from imagery of Back Bays, Great Sound to Great Egg Harbor Bay, NJ . This vector shoreline data is based on an office interpretation of imagery that may be suitable as a geographic information system (GIS) data layer. This metadata describes information for both the line and point shapefiles. The NGS attribution scheme 'Coastal...
Impulsive or transient sounds of short duration and of high intensity constitute one of the criteria for defining good ecological status for descriptor 11 relating to the pressure of noise generated by human activities within the framework of the MSFD (D11C1). Good ecological status for sound energy is achieved when the risks of acoustic disturbance, excess mortality from sound exposure and reduction in communication distances of mysticetes are cumulatively low or moderate. For this, the spatial distribution, the temporal extent and the acoustic levels of the sources of anthropogenic impulsive sound must not exceed the levels harmful to the populations of marine animals. These criteria are evaluated at the scale of the marine sub-region for the “English Channel and North Sea”, “Celtic Seas” and “Western Mediterranean” regions; and at the scale of the "North" and "South" subdivisions in the "Bay of Biscay" sub-region. One of the indicators selected for the evaluation of criterion D11C1 is the distribution of the acoustic levels of impulsive emissions (D11C1.3). The pressures considered for the evaluation of the criterion are: acoustic emissions from air guns ; acoustic emissions from impulsive sources other than air guns ; emissions from non-impulsive (transient) sources ; underwater explosions ; emissions due to pile driving. The data used for the calculation of this indicator are declarative emissions data traced by the operators of the activities generating impulsive noise.
https://www.gnu.org/licenses/agpl.txthttps://www.gnu.org/licenses/agpl.txt
Vibrato Dataset is the extended version of the previous dataset. It consists of vibrato sounds for five different instruments providing different pitch ranges of notes, melodies, articulations, scales and musical techniques as staccato and legato. It includes 295 note tracks with vibratos and corresponding non vibrato tones and 49 melody and articulation tracks.
Data Collection
It is organized according to the source of the recordings. Sounds with vibrato and no vibrato are presented within the folders with their annotations in cvs format. All annotations except than the alto saxophone recordings include derivative analysis of pitch parameters.
The reason of choosing following subsets is that all subsets are created in MTG and they are open source.
The MTG violin recordings includes single notes of two octaves and four different melodies starting from different pitch are recorded in wav format sampled at 44.1 kHz. Each sound separated by semitones is recorded four times for no vibrato,slow, standard and fast rates of vibrato. Melodies were played for no vibrato and vibrato at a standard rate to have the attenuation at the end of the note. Good-sounds alto-sax recordings main feature is that it offers the quality of the vibrato sound and provides vibrato tracks of articulation, scales and staccato, legato techniques. The remaining subset offers a good variety of instruments and mostly two octave range of vibrato-non vibrato pairs.
Dataset Description
The database is meant for organizing the sounds in a handy way. It is organized according to the creator. In each three datasets, annotations and analysis parameters are available within the csv files and each has 11 field descriptor.
Except than alto-sax recordings, each annotation files contain following parameters:
License
All the software is distributed with the Affero GPL v3 license.
CC0 1.0 Universal Public Domain Dedicationhttps://creativecommons.org/publicdomain/zero/1.0/
License information was derived automatically
Background: The effects of shared clinical notes on patients, care partners, and clinicians ("open notes") were first studied as a demonstration project in 2010. Since then, multiple studies have shown clinicians agree shared progress notes are beneficial to patients, and patients and care partners report benefits from reading notes. To determine if implementing open notes at a hematology/oncology practice changed providers' documentation style, we assessed the length and readability of clinicians' notes before and after open notes implementation at an academic medical center in Boston, MA.
Methods: We analyzed 143,888 notes from 60 hematology/oncology clinicians before and after the open notes debut at Beth Israel Deaconess Medical Center, from January 1, 2012, to September 1, 2016. We measured the providers' (medical doctor/nurse practitioner) documentation styles by analyzing character length, the number of addenda, note entry mode (dictated vs. typed) and note readability. Measurements used five different readability formulas and were assessed on notes written before and after the introduction of open notes on November 25, 2013.
Results: After the introduction of open notes, the mean length of progress notes increased from 6,174 characters to 6,648 characters (P<0.001), and the mean character length of the "assessment and plan" (A&P) increased from 1,435 characters to 1,597 characters (P<0.001). The Average Grade Level Readability of progress notes decreased from 11.50 to 11.33, and overall readability improved by 0.17 (P=0.01). There were no statistically significant changes in the length or readability of "Initial Notes" or Letters, inter-doctor communication, nor in the modality of the recording of any kind of note.
Conclusions: After the implementation of open notes, progress notes and A&P sections became both longer and easier to read. This suggests clinician documenters may be responding to the perceived pressures of a transparent medical records environment.
Not seeing a result you expected?
Learn how you can add new datasets to our index.
Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
License information was derived automatically
General description:
This dataset was created in the context of the Pablo project, partially funded by KORG Inc. It contains monophonic recordings of two kind of exercises: single notes and scales. The dataset was reported in the following article:
Romaní Picas O., Parra Rodriguez H., Dabiri D., Tokuda H., Hariya W., Oishi K., & Serra X."A real-time system for measuring sound goodness in instrumental sounds", 138th Audio Engineering Society Convention (2015).
The recordings were made in the Universitat Pompeu Fabra / Phonos recording studio by 15 different professional musicians, all of them holding a music degree and having some expertise in teaching. 12 different instruments were recorded using one or up to 4 different microphones (depending on the recording session). For all the instruments the whole set of playable semitones in the instrument is recorded several times with different tonal characteristics. Each note is recorded into a separate mono .flac audio file of 48kHz and 32 bits. The tonal characteristics are explained both in the the following section and the related publication.
The audio files are organised in one directory for each recording session. In addition to the files, one SQLite database file is included. The structure of the database is related in the following section.
Database description:
The database is meant for organizing the sounds in a handy way. It is organised in four different tables: sounds, takes, packs and ratings.
Sounds
The table containing the sounds annotations.
Takes
A sound can have several takes as some of them were recorded using different microphones at the same time. Each take has an associated audio file.
Packs
A pack is a group of sounds from the same recording session. The audio files are organised in the *sound_files* directory in subfolders with the pack name to which they belong.
Ratings
Some musicians rated some sounds in a 0-10 goodness scale for the user evaluatio of the first project prototype. Please read the paper for more detailed information.
License:
This work is licensed under the Creative Commons Attribution-NonCommercial 4.0 International License. To view a copy of this license, visit http://creativecommons.org/licenses/by-nc/4.0/ or send a letter to Creative Commons, PO Box 1866, Mountain View, CA 94042, USA.