52 datasets found
  1. MusicCaps

    • opendatalab.com
    zip
    Updated Jan 7, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Google Research (2024). MusicCaps [Dataset]. https://opendatalab.com/OpenDataLab/MusicCaps
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 7, 2024
    Dataset provided by
    谷歌http://google.com/
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    The MusicCaps dataset contains 5,521 music examples, each of which is labeled with an English aspect list and a free text caption written by musicians. An aspect list is for example "pop, tinny wide hi hats, mellow piano melody, high pitched female vocal melody, sustained pulsating synth lead", while the caption consists of multiple sentences about the music, e.g., "A low sounding male voice is rapping over a fast paced drums playing a reggaeton beat along with a bass. Something like a guitar is playing the melody along. This recording is of poor audio-quality. In the background a laughter can be noticed. This song may be playing in a bar." The text is solely focused on describing how the music sounds, not the metadata like the artist name. The labeled examples are 10s music clips from the AudioSet dataset (2,858 from the eval and 2,663 from the train split).

  2. h

    lp-music-caps-magnatagatune-3k

    • huggingface.co
    Updated Nov 8, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    mulab-mir (2024). lp-music-caps-magnatagatune-3k [Dataset]. https://huggingface.co/datasets/mulab-mir/lp-music-caps-magnatagatune-3k
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Nov 8, 2024
    Dataset authored and provided by
    mulab-mir
    Description

    mulab-mir/lp-music-caps-magnatagatune-3k dataset hosted on Hugging Face and contributed by the HF Datasets community

  3. h

    LP-MusicCaps-MSD

    • huggingface.co
    Updated Aug 1, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    seungheon.doh (2023). LP-MusicCaps-MSD [Dataset]. https://huggingface.co/datasets/seungheondoh/LP-MusicCaps-MSD
    Explore at:
    Dataset updated
    Aug 1, 2023
    Authors
    seungheon.doh
    Description

    ======================================

    !important: Be careful when using caption_attribute_prediction (We don't recommend to use)!

      Dataset Card for LP-MusicCaps-MSD
    
    
    
    
    
      Dataset Summary
    

    LP-MusicCaps is a Large Language Model based Pseudo Music Caption dataset for text-to-music and music-to-text tasks. We construct the music-to-caption pairs with tag-to-caption generation (using three existing multi-label tag datasets and four task
 See the full description on the dataset page: https://huggingface.co/datasets/seungheondoh/LP-MusicCaps-MSD.

  4. M

    Tencent Music Entertainment Group Market Cap 2018-2025 | TME

    • macrotrends.net
    csv
    Updated Jun 30, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    MACROTRENDS (2025). Tencent Music Entertainment Group Market Cap 2018-2025 | TME [Dataset]. https://www.macrotrends.net/stocks/charts/TME/tencent-music-entertainment-group/market-cap
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jun 30, 2025
    Dataset authored and provided by
    MACROTRENDS
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    2010 - 2025
    Area covered
    United States
    Description

    Tencent Music Entertainment Group market cap as of June 26, 2025 is $21.09B. Tencent Music Entertainment Group market cap history and chart from 2018 to 2025. Market capitalization (or market value) is the most commonly used method of measuring the size of a publicly traded company and is calculated by multiplying the current stock price by the number of shares outstanding.

  5. Top global music companies in Q2 2024, by market capitalization share

    • statista.com
    Updated Jun 23, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Top global music companies in Q2 2024, by market capitalization share [Dataset]. https://www.statista.com/statistics/1498658/biggest-music-companies-by-market-cap-share-worldwide/
    Explore at:
    Dataset updated
    Jun 23, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Area covered
    Worldwide
    Description

    In the second quarter of 2024, Spotify became the biggest music company globally in terms of market capitalization share, reaching **** percent. According to the source, the Swedish music giant managed to overtake its competitor Universal Music Group, whose share stood at **** percent.

  6. o

    Harmonized Cultural Access & Participation Dataset for Music

    • explore.openaire.eu
    • data.niaid.nih.gov
    Updated Jan 29, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daniel Antal (2022). Harmonized Cultural Access & Participation Dataset for Music [Dataset]. http://doi.org/10.5281/zenodo.5917741
    Explore at:
    Dataset updated
    Jan 29, 2022
    Authors
    Daniel Antal
    Description

    Changes since the last version: in the .csv export there was a naming problem. - visit_concert: This is a standard CAP variables about visiting frequencies, in numeric form. - fct_visit_concert: This is a standard CAP variables about visiting frequencies, in categorical form. - is_visit_concert: binary variable, 0 if the person had not visited concerts in the previous 12 months. - artistic_activity_played_music: A variable of the frequency of playing music as an amateur or professional practice, in some surveys we have only a binary variable (played in the last 12 months or not) in other we have frequencies. We will convert this into a binary variable. - fct_artistic_activity_played_music: The artistic_activity_played_music in categorical representation. - artistic_activity_sung: A variable of the frequency of singing as an amateur or professional practice, like played_muisc. Because of the liturgical use of singing, and the differences of religious practices among countries and gender, this is a significantly different variable from played_music. - fct_artistic_activity_sung: The artistic_activity_sung variable in categorical representation. - age_exact: The respondent’s age as an integer number. - country_code: an ISO country code - geo: an ISO code that separates Germany to the former East and West Germany, and the United Kingdom to Great Britain and Northern Ireland, and Cyprus to Cyprus and the Turiksh Cypriot community.[we may leave Turkish Cyprus out for practical reasons.] - age_education: This is a harmonized education proxy. Because we work with the data of more than 30 countries, education levels are difficult to harmonize, and we use the Eurobarometer standard proxy, age of leaving education. It is a specially coded variable, and we will re-code them into two variables, age_education and is_student. - is_student: is a dummy variable for the special coding in age_education for “still studying”, i.e. the person does not have yet a school leaving age. It would be tempting to impute age in this case to age_education, but we will show why this is not a good strategy. - w, w1: Post-stratification weights for the 15+ years old population of each country. Use w1 for averages of geo entities treating Northern Ireland, Great Britain, the United Kingdom, the former GDR, the former West Germany, and Germany as geographical areas. Use w when treating the United Kingdom and Germany as one territory. - wex: Projected weight variable. For weighted average values, use w, w1, for projections on the population size, i.e., use with sums, use wex. - id: The identifier of the original survey. - rowid`: A new unique identifier that is unique in all harmonized surveys, i.e., remains unique in the harmonized dataset.

  7. M

    Mills Music Trust Market Cap 2013-2025 | MMTRS

    • macrotrends.net
    csv
    Updated Jul 31, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    MACROTRENDS (2025). Mills Music Trust Market Cap 2013-2025 | MMTRS [Dataset]. https://www.macrotrends.net/stocks/charts/MMTRS/mills-music-trust/market-cap
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jul 31, 2025
    Dataset authored and provided by
    MACROTRENDS
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    2010 - 2025
    Area covered
    United States
    Description

    Mills Music Trust market cap as of July 06, 2025 is $0.01B. Mills Music Trust market cap history and chart from 2013 to 2025. Market capitalization (or market value) is the most commonly used method of measuring the size of a publicly traded company and is calculated by multiplying the current stock price by the number of shares outstanding.

  8. Most successful K-pop companies in South Korea 2025, by market cap

    • statista.com
    Updated Jun 24, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Statista (2025). Most successful K-pop companies in South Korea 2025, by market cap [Dataset]. https://www.statista.com/statistics/1305919/south-korea-major-kpop-companies-by-market-cap/
    Explore at:
    Dataset updated
    Jun 24, 2025
    Dataset authored and provided by
    Statistahttp://statista.com/
    Time period covered
    Jun 24, 2025
    Area covered
    South Korea
    Description

    As of March 2025, the K-pop entertainment HYBE Corporation achieved the highest market cap in the South Korean stock exchange market, at around ** trillion South Korean won. HYBE Corporation, formerly known as Big Hit Entertainment, is home to the globally popular K-pop group BTS. Having only entered the stock exchange market in 2020, the company has since stayed ahead of its competition.

  9. h

    GraySpectrogram

    • huggingface.co
    Updated Oct 22, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    ML (2023). GraySpectrogram [Dataset]. https://huggingface.co/datasets/mickylan2367/GraySpectrogram
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Oct 22, 2023
    Authors
    ML
    License

    Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
    License information was derived automatically

    Description

    Google/Music-CapsăźéŸłćŁ°ăƒ‡ăƒŒă‚żă‚’ă‚čăƒšă‚Żăƒˆăƒ­ă‚°ăƒ©ăƒ ćŒ–ă—ăŸăƒ‡ăƒŒă‚żă€‚

    Music CpasăšăŻïŒšhttps://huggingface.co/datasets/google/MusicCaps GrayScaleじゃăȘă„ă»ă†ă‚‚ă‚ă‚‹ă‹ă‚‰èŠ‹ăŠă­(â‹ˆâ—ïŒžâ—ĄïŒœâ—)ă€‚âœ§â™ĄïŒˆă“ă‚ŒïŒ‰

      ćŸșæœŹæƒ…ć ±
    

    sampling_rate: int = 44100 20秒ぼwavăƒ•ă‚Ąă‚€ăƒ« -> 1600×800たpngăƒ•ă‚Ąă‚€ăƒ«ăžć€‰æ› librosaăźèŠæ Œă«ă‚ˆă‚Šă€ç”»ćƒăźçžŠè»žïŒš(0-10000?Hz), ç”»ćƒăźæšȘ軞(0-40秒) è©łă—ăăŻlibrosa.specshow() -> https://librosa.org/doc/main/auto_examples/plot_display.html

      äœżă„æ–č
    
    
    
    
    
      0: ăƒ‡ăƒŒă‚żă‚»ăƒƒăƒˆă‚’ăƒ€ă‚Šăƒłăƒ­ăƒŒăƒ‰
    

    from datasets import load_dataset data = load_dataset("mickylan2367/spectrogram") data
 See the full description on the dataset page: https://huggingface.co/datasets/mickylan2367/GraySpectrogram.

  10. Multimodal datataset of EEG, ECG and audio recordings of individual...

    • zenodo.org
    zip
    Updated Dec 17, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zenodo (2023). Multimodal datataset of EEG, ECG and audio recordings of individual musicians playing emotional music [Dataset]. http://doi.org/10.5281/zenodo.10396364
    Explore at:
    zipAvailable download formats
    Dataset updated
    Dec 17, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Time period covered
    Dec 12, 2023
    Description

    This dataset contains EEG, ECG and audio recordings of 11 individual musicians playing emotional music on their instrument. The dataset consists of two parts:

    1. Experiment: Musicians’ self-reported ratings, audio recordings, and physiological recordings where the 11 expert musicians were asked to play at least 4~2-minute unfamiliar (non-popularly known) musical pieces. Participants were asked to play at least once one of the following emotions: happiness, sadness, relaxation, and anger. For the physiological recordings EEG, ECG, and GSR signals were recorded. Each musican’s data is denoted by MS_ followed by the order in which they were recorded.

    2. Self-report questionnaire: A self-assessment questionnaire and their answers where 11 expert musicians were asked to rate musical pieces recorded based on:

      1. Objective valence & arousal they felt the piece had;

      2. Felt valence & arousal during playing.

    For a more detailed explanation of the dataset, its recording procedure, and its contents, see

    L. Turchet, B. O'Sullivan, R. Ortner & C. Gugher (2024). Emotion Recognition of Playing Musicians from EEG, ECG, and Acoustic Signals. IEEE Transactions on Human-Machine Systems.

    File Listing

    The following files are available (each explained in more detail below):

    Name

    Format

    Contents

    EEG_ECG_data_for_each_musician

    mat

    This folder contains the raw EEG, ECG, & GSR data for all 11 musicians for each piece they played as well as a resting state recording, which was recorded while a neutral audio stimulus was played.

    audio_data_for_each_musician

    wav, JSON, csv

    This folder contains 3 subfolders:
    1) wav_audio_files_original: This is the raw audio data recorded for each musician and includes all 56 pieces included in the paper reported above, please see below regarding rejected trials.

    2) wav_audio_files (original split_into_3_parts): Here, the 56 pieces are appropriately split into 3 separate parts.

    3) analysis_audio_files: This folder contains the acoustic features extracted, 1714 acoustic features were extracted from each split trial. Each result is stored in .JSON, however, the collated results can be seen in all_results.csv.

    self_report_questionnaire

    pdf, xls

    Two files exist in this folder:
    1. The self-reported questionnaire given to the musicians of the questionnaire during the experiment.
    2.



    File Details

    EEG_ECG_data_for_each_musician

    These are the original raw data recordings. EEG data were recorded using a g.GAMMAcap2 by g.tec Medical Engineering, a 64-channel cap with g.SCARABEO active electrodes, with two g.GAMMAsys reference active ear clip electrodes. Two g.GAMMAbox electrode connector boxes were used to connect the active electrodes to two g.USBamp biosignal amplifiers with a sampling frequency of 256 Hz.

    The following 31 EEG channels were used: Fp1, Fp2, AFz, AF3, AF4, AF7, AF8, Fz, F3, F4, F7, F8, Cz, C3, C4, CP3, CP4, CP5, CP6, P1, P2, P3, P4, P5, P6, P7, P8, PO7, PO8, O1, and O2. AFz was used as a ground electrode and Cz was used as a re-reference electrode. The right-side ear clip electrode was used as a reference electrode.


    ECG data were recorded using a single g.GAMMAclip active electrode clip connected directly to the g.GAMMAbox, sharing the same ground electrode with the EEG cap and placed on position V4 of the subjects.

    GSR data was recorded using the g.GSRsensorÂČ box which contains two small dry electrodes placed underneath the participant’s toes (due to the amount of hand movement required for playing an instrument). The g.GSRsensorÂČ was connected directly to a g.GAMMAbox, using jumper cables connected to different ports in the g.USBAMPs to share the same reference and ground electrodes as the EEG cap.

    The locations of the channels and their corresponding number in the raw data is as follows:

    Channel Number

    Channel Name

    1

    Time series

    2

    AF3

    3

    AF4

    4

    AF7

    5

    AF8

    6

    CP3

    7

    CP4

    8

    CP5

    9

    CP6

    10

    P1

    11

    P2

    12

    P5

    13

    P6

    14

    P7

    15

    P8

    16

    O1

    17

    O2

    18

    Cz

    19

    Fp1

    20

    Fp2

    21

    F3

    22

    F4

    23

    F7

    24

    F8

    25

    C3

    26

    C4

    27

    P3

    28

    P4

    29

    PO7

    30

    PO8

    31

    AFz

    32

    GSR

    33

    ECG



    audio_data_for_each_musician/wav_audio_files_original

    This folder contains all of the raw audio files recorded during the experiment. All pieces were recorded using the software Audacity and exported as WAV files encoded with a bit depth of 32-bits and a sampling rate of 44.1 kHz. 56 of the included trials are present in this folder.

    audio_data_for_each_musician/wav_audio_files (original split_into_3_parts)

    This folder contains the above mentioned 56 raw audio pieces separated into 3 appropriately sized recordings.

    audio_data_for_each_musician/analysis_audio_files

    This folder contains the 1714 acoustic features from each of the 3 separated trials from 56 accepted pieces, these are denoted by MS_ followed by the order in which the musicians were recorded and the order in which the trails were split, the individual features are in JSON format. Within this folder, we have collated all results in a csv file (all_results.csv) where the columns show the intended emotion, trial name and number, followed by the names of the acoustic features.

    self_report_questionnaire

    This pdf is the questionnaire which each musician was given following each trial relating to the emotions communicated and felt during each trial. Most* questions in the questionnaire were multiple-choice and speak pretty much for themselves. The answers for which were collated and are described below.

    self_report_questionnaire_anwsers

    This .xls file contains the results of all the musicians self-reported ratings from the above described questionnaire.

    Column name

    Description

    Subject

    The subject code of the musician, denoted by MS_

    recording_filename

    The name of the trial denoted by MS_01_ followed by the trial number.

    intended_emotion

    The intended emotion the musician was instructed to communicate. The emotions are as follows:

    1. Angry

    2. Sad

    3. Relaxed

    4. Happy

    They are intended to be reported by using the valence-arousal space.

    valence_communicated

    The valence rating (integer between 1 and 5), that participants were asked to objectively rate how they thought the music played would be perceived.

    arousal_communicated

    As above, but relating to arousal rather than valence.

    valence_felt

    The valence rating (integer between 1 and 5), that participants were asked to rate how they felt while playing the piece.

    arousal_felt

    As above, but relating to arousal rather than valence.

    instrument_played

    The instrument played for each trial.



  11. f

    Assessment Scores of participants in the research study for "Combined Music...

    • figshare.com
    xlsx
    Updated Mar 1, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Asta Hodgson (2024). Assessment Scores of participants in the research study for "Combined Music and Movement Entrainment Activities Positively Impact Sharing Capacity in Children aged 4-5 years" [Dataset]. http://doi.org/10.6084/m9.figshare.25303747.v1
    Explore at:
    xlsxAvailable download formats
    Dataset updated
    Mar 1, 2024
    Dataset provided by
    figshare
    Authors
    Asta Hodgson
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Child participants in the study were assessed at three points during the study - once prior to entrainment intervention activities, once at the midpoint, and once after the third and final intervention session. The measurement procedures and scripts are below:MEASUREMENT PROCEDURE 1 - Set up in Butterfly Room. - EY teachers and TAs deliver the following message to students “Miss. H will have some music activities during some COOL periods this week, and she says she would like you and a favourite toy to join her. So when you are called to go and play with her, make sure you have your toy ready”. - TAs to prepare children and toys prior to assessment (follow list as provided once consent received by all parents). - TA to show where the next child to collect is. Ensure toy is present. I will take the toy from the child and ask a selection of the following questions “Who is this?”/“Tell me about your friend/car/game etc”/“Why did you choose to bring
?” - Just outside the Butterfly Room there will be an accident and the toy will be dropped. Exclamation “Oh no, poor 
! I hope they are ok! Let’s play some music to make them feel better/fix any damage” - Inside BR will be iPad (my personal device) and one school device (TAs to help monitor battery and ensure recordings happen). - Equipment set in front of child: 1 xylophone, 1 step glockenspiel, 1 set bongo drums, 1 tambour, 2 sleigh bells, 1 duck castanet, 1 shaky egg - Equipment set in front of experimenter: 1 shaky egg, 2 chime bars (G + E) - Invite child to sit down opposite experimenter and place toy in special cradle. “Let’s play healing music together to make x better after the fall. Do you think we can make them feel better?” If the child shares instruments at this point they will score 3 points. If during the next 90 seconds playing time the child recognises that there are different instruments, or identifies that they could share but won’t for some reason, or if they share after the prompt they receive 2 points. No willingness or action to share scores 1 point. The prompt will take place after 40 seconds. The prompt will be the experimenter noticing to the child and saying “you’re not using all your instruments”. In the instance that they are in fact using them all, the experimenter will say “you actually have more instruments to play for insert toy name”. Playing will cease after 60 seconds. Alarm set and the alarm will also be signal for the toy feeling better/being fixed. - Repeat procedure for each child recording score immediately. MEASUREMENT PROCEDURE 2 - Set up in Butterfly Room (recording devices, sweets, party hats, playlist etc). - Experimenters are SG and AH - Children are kept by intervention group near the BR (can be playing a game with TAs for example). Children called in by both teachers to Butterfly Room individually. Wearing party hats, celebration/birthday song playlist and some balloons around to provide festive feel. Each child greeted and invited to sit in their place. A/S alternating each child say “it’s our birthdays these month, and we wanted to celebrate by having a little party with each child! Shall we celebrate with skittles or M&Ms?” According to child’s choice, A/S distributes 1 or each to birthday teachers, and 4 to the child (on party plate). Sing Happy Birthday to one teacher, then the next. “Ithink I might eat my Skittle/M&M now, yummy” - this is the prompt to share. If the child shares beforehand they receive 3 points, after the prompt, 2 points, not at all, 1 point. MEASUREMENT PROCEDURE 3 stickers (not same as usual music housepoint stickers), envelopes (2 for each child, 1 large with name on, 1 smaller and identifiably different e.g.different colour) - Butterfly Room (recording devices set up) - Each child called to the room individually by myself, TAs assist locating and perhaps preparing next child, no procedure particularly necessary to follow for this measurement. - “Thank you for taking part in my extra music and movement sessions. I want to thank everyone by giving them some stickers. So these are for you. But I don’t know if I will have time to get round to everyone, so if you want to, you can put some to share with other children in this envelope in case I don’t get round to all your friends. Or you can keep them all. It’s up to you. And I won’t watch because i just have to reply to this email”. I will then busy myself with my email for 30 seconds as the child decides what to do. “OK that’s done. So if you’ve decided, you can put your stickers in this envelope with your name in it and I’ll give them out at the end of the day”. Child’s stickers and donated stickers recorded. - 3 points - over half of stickers shared; 2 points - some stickers shared, maybe some deliberation or negotiation about decision; 1 point - no stickers shared.

  12. Harmonized Cultural Access & Participation Dataset for Music

    • zenodo.org
    csv
    Updated Jun 3, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Daniel Antal; Daniel Antal (2022). Harmonized Cultural Access & Participation Dataset for Music [Dataset]. http://doi.org/10.5281/zenodo.5917742
    Explore at:
    csvAvailable download formats
    Dataset updated
    Jun 3, 2022
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Daniel Antal; Daniel Antal
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    - `visit_concert`: This is a standard CAP variables about visiting frequencies.
    - `is_visit_concert`: binary variable, 0 if the person had not visited concerts in the previous 12 months.
    - `artistic_activity_played_music`: A variable of the frequency of playing music as an amateur or professional practice, in some surveys we have only a binary variable (played in the last 12 months or not) in other we have frequencies. We will convert this into a binary variable.
    - `artistic_activity_sung`: A variable of the frequency of singing as an amateur or professional practice, like played_muisc. Because of the liturgical use of singing, and the differences of religious practices among countries and gender, this is a significantly different variable from played_music.
    - `age_exact`: The respondent’s age as an integer number.
    - `country_code`: an ISO country code
    - `geo`: an ISO code that separates Germany to the former East and West Germany, and the United Kingdom to Great Britain and Northern Ireland, and Cyprus to Cyprus and the Turiksh Cypriot community.[we may leave Turkish Cyprus out for practical reasons.]
    - `age_education`: This is a harmonized education proxy. Because we work with the data of more than 30 countries, education levels are difficult to harmonize, and we use the Eurobarometer standard proxy, age of leaving education. It is a specially coded variable, and we will re-code them into two variables, `age_education` and `is_student`.
    - `is_student`: is a dummy variable for the special coding in age_education for “still studying”, i.e. the person does not have yet a school leaving age. It would be tempting to impute `age` in this case to `age_education`, but we will show why this is not a good strategy.
    - `w`, `w1`: Post-stratification weights for the 15+ years old population of each country. Use `w1` for averages of `geo` entities treating Northern Ireland, Great Britain, the United Kingdom, the former GDR, the former West Germany, and Germany as geographical areas. Use `w` when treating the United Kingdom and Germany as one territory.
    - `wex`: Projected weight variable. For weighted average values, use `w`, `w1`, for projections on the population size, i.e., use with sums, use `wex`.
    - `id`: The identifier of the original survey.
    - `rowid``: A new unique identifier that is unique in all harmonized surveys, i.e., remains unique in the harmonized dataset.

  13. f

    Table_4_The critic’s voice: On the role and function of criticism of...

    • frontiersin.figshare.com
    docx
    Updated Jun 16, 2023
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Elena Alessandri; Antonio Baldassarre; Victoria Jane Williamson (2023). Table_4_The critic’s voice: On the role and function of criticism of classical music recordings.docx [Dataset]. http://doi.org/10.3389/fpsyg.2022.925394.s004
    Explore at:
    docxAvailable download formats
    Dataset updated
    Jun 16, 2023
    Dataset provided by
    Frontiers
    Authors
    Elena Alessandri; Antonio Baldassarre; Victoria Jane Williamson
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    In the Western classical tradition music criticism represents one of the most complex and influential forms of performance assessment and evaluation. However, in the age of peer opinion sharing and quick communication channels it is not clear what place music critics’ judgments still hold in the classical music market. This article presents expert music critics’ view on their role, function, and influence. It is based on semi-structured interviews with 14 native English- and German-speaking critics who had an average of 32 years professional activity in classical music review. We present the first visual model to summarize music critics’ descriptions of their role and responsibilities, writing processes, and their influences (on the market and on artists). The model distinguishes six roles (hats): consumer adviser, teacher, judge, writer, stakeholder, and artist advocate. It identifies core principles governing critical writing for music as well as challenges that arise from balancing the above six responsibilities whilst remaining true to an implicit code of conduct. Finally, it highlights the factors that inform critics’ writing in terms of the topics they discuss and the discursive tools they employ. We show that music critics self-identify as highly skilled mediators between artists, producers and consumers, and justify their roles as judge and teacher based on a wealth of experience as against the influx of pervasive amateur reviews. Our research approach also offers occupation-based insights into professional music review standards, including the challenges of maintaining objectivity and resisting commercial pressures. This article offers a new viewpoint on music critics’ judgments and recommendations that helps to explain their expectations and reflections.

  14. h

    JamendoMaxCaps

    • huggingface.co
    Updated Jun 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    AMAAI Lab (2025). JamendoMaxCaps [Dataset]. https://huggingface.co/datasets/amaai-lab/JamendoMaxCaps
    Explore at:
    Dataset updated
    Jun 3, 2025
    Dataset authored and provided by
    AMAAI Lab
    License

    Attribution-ShareAlike 3.0 (CC BY-SA 3.0)https://creativecommons.org/licenses/by-sa/3.0/
    License information was derived automatically

    Description

    JamendoMaxCaps Dataset

    JamendoMaxCaps is a large-scale dataset of over 362,000 instrumental tracks sourced from the Jamendo platform. It includes generated music captions and original metadata. Additionally, we introduce a retrieval system that utilizes both musical features and metadata to identify similar songs, which are then used to impute missing metadata via a local large language model (LLLM). This dataset facilitates research in:

    Music-language understanding Music
 See the full description on the dataset page: https://huggingface.co/datasets/amaai-lab/JamendoMaxCaps.

  15. e

    Sieben - Dataset - B2FIND

    • b2find.eudat.eu
    Updated Dec 15, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2023). Sieben - Dataset - B2FIND [Dataset]. https://b2find.eudat.eu/dataset/85cd91ef-b3ab-5fd1-bfc9-067c2835e973
    Explore at:
    Dataset updated
    Dec 15, 2023
    Description

    Abstract: The play Sieben is a musical staging of the seven mortal sins, which are slightly modified from their original meaning: Greed, envy, inertia, anger, gluttony, condescendence, and sexuality are portrayed and processed by the group in songs composed for the play by Berlin-based singer Susanne Betancor. Each actor, some of them with intellectual disabilities, represents one sin. The actual play and behind-the-scenes reflections on the songwriting by the directors and the performers are shown alternately. Details: Another stage has been erected on the stage, which can be entered left and right via two staircases. The seven performers, five women, and two men sit spread out on the stage. The play starts with a musical staging of the song “Über sieben BrĂŒcken muss du gehen”, a GDR pop classic by the group Karat from the 70s. Then, for the first time, a reflection of the directors is shown, which is conveyed in the format of a video message that is cut in between the staging. The two directors, visual director Antje Siebers, and music director BĂ€rbel Schwarz, make comments on Susanne Betancor’s song, which is supposed to portray “Greed”. Lastly, they suggest she communicate via WhatsApp. The staging of Greed follows. The main character repetitively sings “I’m a bargain hunter, I’ve got a Payback card, I don’t pay my Wi-Fi” as well as other such comparable statements. He is accompanied by the other actors. At times, the song takes a more aggressive tone, and the main actor just shouts “Geiz!” (“Greed!”). Then six actors sit down on stage, wearing red pointed hats. One performer plays the electric guitar. Few set pieces from Snow White are read out, especially those that portray the vanity of the queen. Psychedelic music accompanies this. Once again there is a video message. BĂ€rbel praises the new song that Susanne composed for the play. Then Antje plays a bar, and BĂ€rbel questions if this bar is meant to be or if it is a mistake. In the next video message, the performer of Envy thanks Susanne for the song. He says it is good, only the flutes are annoying him. The staging of Envy follows. Flute playing alternates with descriptive chants, which represent different expressions of envy. At times, the actors talk at the same time, interrupting each other with complaints about envy experiences. Then everyone gathers in front of the stage, the Envy actor plays the electric guitar and climbs the stairs. As he does so, everyone joins in a chorus about Envy. The fragment ends with a riff by the Envy performer. Again, a version of “Über sieben BrĂŒcken must du gehen” is sung, with the actors taking turns singing the verses. Afterward, Antje and BĂ€rbel take a stand on the “Indolence song”. They tell Susanne that they have reworked it into a song about “Laziness”. The lead actress of Laziness then asks Susanne, again in a video message, to make her song longer. The stage is shown. The actress portraying Laziness stands on the upper platform while the others below her walk in circles. In a piano piece, idleness, symbolized in bathing, is sung about. The reflection of the music piece on anger is shown. BĂ€rbel and Antje criticize one of Susanne’s songs - in the process, they fight about minor issues. The main actress of Anger complains in another video message to Susanne that another actor should sing along to “her” song. The staging of anger follows. The actors take turns stepping to the edge of the stage and shouting things they are angry at towards the fourth wall. There are overlaps and interruptions. Finally, the lead actress yells, “Shut up, it’s my turn now.” A metal-like piece of music follows, in which it is sung that she can’t stand horses and that she hates them. After that, the production returns to the motif of Snow White. The first meeting with the seven dwarfs, the entry into their dwelling, as well as the divination of the mirror is depicted. All actors wear pointed hats, six sit as if in a cave under the stage, while an actress plays an electric guitar. There follows another musical reflection by the directors, now on the gluttony song. The two seem stressed and drink alcohol. One actress wishes to add electronic elements to her song about gluttony. The staging follows, with the song lyrics consisting of a stringing together of various dishes and foods. The singing turns into loud drumming, and instead of enumerating types of food, the performers shout “devour” (“fressen”) and “shut up” (“halt die Fresse”) at the top of their voices. A video showing the director eating is shown at the same time. It cuts to another video. One of the actresses, the representative of condescendence, is recorded confessing to Susanne, in tears and completely distraught, that she can’t hit a difficult note in her opera-like song. The representation of condescendence follows. The seven performers walk across the stage. As they do so, they look condescendingly at and through the fourth wall, with the leading actress, in particular, making arrogant remarks about the audience’s clothing choices. The scene transitions into an opera, which is sung flawlessly by the leading lady. In the next video, the performer of “Avarice” is shown. He suggests to the songwriter Susanne that if she still doesn't have Wi-Fi after moving into her new place, she to steal the neighbour’s Wi-Fi. The Snow-White theme is taken up again. Once again, everyone wears pointed hats, and one performer, the actress of “Lust”, plays the electric guitar. The staging slowly moves into the final sin, which is lust. The lead actress repetitively sings “I touch you... You touch me” to a mournful melody, and she enumerates body parts, playing the drums. The stage goes dark. The drums get faster and louder. The other performers dance obscenely across the stage to strobe lights. Slowly it gets brighter again - the six actors line up while the drums continue to play. Slowly and insistently, everyone counts to seven - then the play ends on stage. The last video shows Susanne, supposedly in her new apartment, telling the others that she now has Wi-Fi.

  16. Music Landscape Market Analysis | Industry Growth, Size & Forecast Report...

    • mordorintelligence.com
    pdf,excel,csv,ppt
    Updated May 16, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mordor Intelligence (2025). Music Landscape Market Analysis | Industry Growth, Size & Forecast Report 2030 [Dataset]. https://www.mordorintelligence.com/industry-reports/music-market-landscape
    Explore at:
    pdf,excel,csv,pptAvailable download formats
    Dataset updated
    May 16, 2025
    Dataset authored and provided by
    Mordor Intelligence
    License

    https://www.mordorintelligence.com/privacy-policyhttps://www.mordorintelligence.com/privacy-policy

    Time period covered
    2019 - 2030
    Area covered
    Global
    Description

    The Music Market report segments the industry into Revenue Generation Format (Streaming, Digital (Except Streaming), Physical Products, Performance Rights, Synchronization Revenues) and Geography (North America, Europe, Asia-Pacific, Latin America, Middle East and Africa). Get five years of historical data alongside five-year market forecasts.

  17. m

    Online Music Education Market Size, Trends, Growth 2025 – 2030

    • mordorintelligence.com
    pdf,excel,csv,ppt
    Updated Jul 7, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mordor Intelligence (2025). Online Music Education Market Size, Trends, Growth 2025 – 2030 [Dataset]. https://www.mordorintelligence.com/industry-reports/online-music-education-market
    Explore at:
    pdf,excel,csv,pptAvailable download formats
    Dataset updated
    Jul 7, 2025
    Dataset authored and provided by
    Mordor Intelligence
    License

    https://www.mordorintelligence.com/privacy-policyhttps://www.mordorintelligence.com/privacy-policy

    Time period covered
    2019 - 2030
    Area covered
    Global
    Description

    Online Music Education Market is Segmented by Instrument Type (Piano, Guitar, and More), Platform (App-Based, Web-Based, and Hybrid), End-User (Individual Learners, K-12 Institutions, Higher Education, and More), Learning Model (Self-Paced Asynchronous Courses, Live Instructor-Led One-To-One, and More), Session Format (One-To-One, Small Group, and More), and Geography. The Market Forecasts are Provided in Terms of Value (USD).

  18. t

    Music Recording Global Market Report 2025

    • thebusinessresearchcompany.com
    pdf,excel,csv,ppt
    Updated Jan 8, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    The Business Research Company (2025). Music Recording Global Market Report 2025 [Dataset]. https://www.thebusinessresearchcompany.com/report/music-recording-global-market-report
    Explore at:
    pdf,excel,csv,pptAvailable download formats
    Dataset updated
    Jan 8, 2025
    Dataset authored and provided by
    The Business Research Company
    License

    https://www.thebusinessresearchcompany.com/privacy-policyhttps://www.thebusinessresearchcompany.com/privacy-policy

    Description

    Global Music Recording market size is expected to reach $84.73 billion by 2029 at 6.7%, segmented as by type, record production, music publishers, record distribution, sound recording studios

  19. Sales revenue of all time top-selling music artists online in China 2024

    • statista.com
    Updated Dec 20, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Lai Lin Thomala (2023). Sales revenue of all time top-selling music artists online in China 2024 [Dataset]. https://www.statista.com/topics/4680/digital-music-industry-in-china/
    Explore at:
    Dataset updated
    Dec 20, 2023
    Dataset provided by
    Statistahttp://statista.com/
    Authors
    Lai Lin Thomala
    Area covered
    China
    Description

    Jay Chou was the best-selling music artist in China. As of April 2, 2024, the Taiwanese superstar recorded about 321.6 million yuan in sales revenue across all main music platforms in China. The Chinese singer Xiao Zhan, a member of the idol boy band ININE, ranked sixth with only one single released in April 2020. The two big names and their musical mastery Jay Chou and JJ Lin are the highly influential Asian figures in the Chinese music industry. The former has garnered a massive fan base due to his exceptional ability to compose and produce music that seamlessly blends elements of rock, R&B, and pop. Singaporean singer-songwriter JJ Lin, renowned for his powerful vocals and heartfelt ballads, has captivated millions of listeners with his emotional and relatable music. Both artists have two hits on the all-time online top ten best-selling music singles in China. Swiftonomics in China The story-driven songwriting style of Taylor Swift, catchy pop tunes, and relatable lyrics make her the most loved Western music figure in China. The “Queen of Pop” has the highest-selling English music albums in the country. This accomplishment can be attributed to her active presence on social media platforms and celebrity endorsements in China, which allows her to directly connect with her local fans.

  20. C

    Cowboy Hats Report

    • archivemarketresearch.com
    doc, pdf, ppt
    Updated Apr 18, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Archive Market Research (2025). Cowboy Hats Report [Dataset]. https://www.archivemarketresearch.com/reports/cowboy-hats-261284
    Explore at:
    doc, ppt, pdfAvailable download formats
    Dataset updated
    Apr 18, 2025
    Dataset authored and provided by
    Archive Market Research
    License

    https://www.archivemarketresearch.com/privacy-policyhttps://www.archivemarketresearch.com/privacy-policy

    Time period covered
    2025 - 2033
    Area covered
    Global
    Variables measured
    Market Size
    Description

    The global cowboy hat market, encompassing straw and felt varieties sold through online and offline channels, is experiencing steady growth. While precise figures for market size and CAGR are unavailable, a reasonable estimation, based on similar apparel market trends and considering the enduring popularity of cowboy hats, suggests a market size of approximately $500 million in 2025. Assuming a conservative CAGR of 4% (reflective of the steady, rather than explosive growth typical of niche fashion markets), the market is projected to reach approximately $650 million by 2033. This growth is fueled by several key drivers. The resurgence of country music and Western-themed fashion influences younger demographics, expanding the consumer base beyond traditional enthusiasts. The increasing popularity of outdoor activities like rodeos, country festivals, and even casual outdoor wear occasions contribute to heightened demand. Furthermore, the rise of e-commerce platforms allows hat makers and retailers to reach wider audiences, boosting sales. However, market growth is tempered by factors such as fluctuating raw material prices, competition from other headwear styles, and the impact of economic downturns on discretionary spending. Market segmentation reveals online sales are gradually outpacing offline sales, reflecting the growing preference for convenient online shopping experiences. Key players like Stetson, Resistol, and American Hat Company are leveraging their brand heritage and quality to maintain market share, while newer brands are increasingly focusing on innovative designs and sustainable materials to attract a wider customer base. The geographical distribution of the market shows robust demand in North America, particularly the United States, with Europe and Asia-Pacific also exhibiting notable growth potential. The strategic focus of leading brands involves expanding online presence, diversifying product offerings (e.g., incorporating modern designs while retaining classic styles), and tapping into the growing demand for sustainable and ethically sourced materials. Furthermore, targeted marketing campaigns focusing on specific demographics (e.g., younger consumers interested in Western-inspired fashion) and leveraging the influence of social media will be crucial for driving future growth. The competitive landscape is dynamic, with established brands facing challenges from smaller, more agile competitors who are adept at leveraging e-commerce and digital marketing strategies. The forecast period of 2025-2033 presents significant opportunities for market expansion, particularly in regions with a growing appreciation for Western fashion and culture. Successful players will need to adapt to evolving consumer preferences, embrace sustainable practices, and optimize their distribution channels to capitalize on these opportunities.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Google Research (2024). MusicCaps [Dataset]. https://opendatalab.com/OpenDataLab/MusicCaps
Organization logo

MusicCaps

OpenDataLab/MusicCaps

Explore at:
215 scholarly articles cite this dataset (View in Google Scholar)
zipAvailable download formats
Dataset updated
Jan 7, 2024
Dataset provided by
谷歌http://google.com/
License

Attribution-ShareAlike 4.0 (CC BY-SA 4.0)https://creativecommons.org/licenses/by-sa/4.0/
License information was derived automatically

Description

The MusicCaps dataset contains 5,521 music examples, each of which is labeled with an English aspect list and a free text caption written by musicians. An aspect list is for example "pop, tinny wide hi hats, mellow piano melody, high pitched female vocal melody, sustained pulsating synth lead", while the caption consists of multiple sentences about the music, e.g., "A low sounding male voice is rapping over a fast paced drums playing a reggaeton beat along with a bass. Something like a guitar is playing the melody along. This recording is of poor audio-quality. In the background a laughter can be noticed. This song may be playing in a bar." The text is solely focused on describing how the music sounds, not the metadata like the artist name. The labeled examples are 10s music clips from the AudioSet dataset (2,858 from the eval and 2,663 from the train split).

Search
Clear search
Close search
Google apps
Main menu