11 datasets found
  1. ASVspoof 5: Design, Collection and Validation of Resources for Spoofing,...

    • zenodo.org
    application/gzip, tar +1
    Updated Feb 14, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Xin Wang; Xin Wang; Héctor Delgado; Héctor Delgado; Hemlata Tak; Hemlata Tak; Jee-weon Jung; Jee-weon Jung; Hyejin Shim; Hyejin Shim; Massimiliano Todisco; Massimiliano Todisco; Ivan Kukanov; Ivan Kukanov; Xuechen Liu; Xuechen Liu; Md Sahidullah; Md Sahidullah; Tomi Kinnunen; Tomi Kinnunen; Nicholas Evans; Nicholas Evans; Kong Aik Lee; Kong Aik Lee; Junichi Yamagishi; Junichi Yamagishi; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh (2025). ASVspoof 5: Design, Collection and Validation of Resources for Spoofing, Deepfake, and Adversarial Attack Detection Using Crowdsourced Speech [Dataset]. http://doi.org/10.5281/zenodo.14498691
    Explore at:
    tar, application/gzip, txtAvailable download formats
    Dataset updated
    Feb 14, 2025
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Xin Wang; Xin Wang; Héctor Delgado; Héctor Delgado; Hemlata Tak; Hemlata Tak; Jee-weon Jung; Jee-weon Jung; Hyejin Shim; Hyejin Shim; Massimiliano Todisco; Massimiliano Todisco; Ivan Kukanov; Ivan Kukanov; Xuechen Liu; Xuechen Liu; Md Sahidullah; Md Sahidullah; Tomi Kinnunen; Tomi Kinnunen; Nicholas Evans; Nicholas Evans; Kong Aik Lee; Kong Aik Lee; Junichi Yamagishi; Junichi Yamagishi; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh
    License

    Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
    License information was derived automatically

    Time period covered
    Dec 2024
    Description

    This is the Zenodo repository for the ASVspoof 5 database. ASVspoof 5 is the fifth edition in a series of challenges which promote the study of speech spoofing and deepfake attacks, and the design of detection solutions. Compared to previous challenges, the ASVspoof~5 database is built from crowdsourced data collected from around 2,000 speakers in diverse acoustic conditions. More than 20 attacks, also crowdsourced, are generated and optionally tested using surrogate detection models, while seven adversarial attacks are incorporated for the first time.

    • Please check README.txt and LICENSE.txt before downloading the database.
    • Database paper (to be submitted): https://arxiv.org/abs/2502.08857
      • Please consider citing the reference listed at the bottom of this page.
    • It is highly recommended to follow the rules and instructions in the ASVspoof 5 challenge evaluation plan (phase 2, https://www.asvspoof.org/), if you want to produce results comparable with the literature.
    • Latest work using the ASVspoof 5 database can be found in the Automatic Speaker Verification Spoofing Countermeasures Workshop proceeding: https://www.isca-archive.org/asvspoof_2024/index.html
    • If you are interested in creating spoofed data for research purpose using the ASVspoof 5 protocol, please send request to info@asvspoof.org
  2. h

    asvspoof2019

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    LanceaKing, asvspoof2019 [Dataset]. https://huggingface.co/datasets/LanceaKing/asvspoof2019
    Explore at:
    Authors
    LanceaKing
    License

    https://choosealicense.com/licenses/odc-by/https://choosealicense.com/licenses/odc-by/

    Description

    This is a database used for the Third Automatic Speaker Verification Spoofing and Countermeasuers Challenge, for short, ASVspoof 2019 (http://www.asvspoof.org) organized by Junichi Yamagishi, Massimiliano Todisco, Md Sahidullah, Héctor Delgado, Xin Wang, Nicholas Evans, Tomi Kinnunen, Kong Aik Lee, Ville Vestman, and Andreas Nautsch in 2019.

  3. t

    ASVspoof 2021 - Dataset - LDM

    • service.tib.eu
    Updated Dec 16, 2024
    + more versions
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    (2024). ASVspoof 2021 - Dataset - LDM [Dataset]. https://service.tib.eu/ldmservice/dataset/asvspoof-2021
    Explore at:
    Dataset updated
    Dec 16, 2024
    Description

    The ASVspoof 2021 dataset is a large-scale public dataset for speaker verification and spoofing countermeasures. The dataset contains various types of audio files, including real and fake audio, with different levels of noise and background music.

  4. E

    Data from: Automatic Speaker Verification Spoofing and Countermeasures...

    • dtechtive.com
    • find.data.gov.scot
    gz, pdf, txt
    Updated Sep 17, 2015
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    University of Edinburgh. The Centre for Speech Technology Research (CSTR) (2015). Automatic Speaker Verification Spoofing and Countermeasures Challenge (ASVspoof 2015) Database [Dataset]. http://doi.org/10.7488/ds/298
    Explore at:
    txt(0.0054 MB), txt(0.0006 MB), gz(2.067 MB), gz(7724.032 MB), txt(0.0166 MB), gz(7506.944 MB), pdf(0.1251 MB), pdf(0.1037 MB)Available download formats
    Dataset updated
    Sep 17, 2015
    Dataset provided by
    University of Edinburgh. The Centre for Speech Technology Research (CSTR)
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    The database has been used in the first Automatic Speaker Verification Spoofing and Countermeasures Challenge (ASVspoof 2015). Genuine speech is collected from 106 speakers (45 male, 61 female) and with no significant channel or background noise effects. Spoofed speech is generated from the genuine data using a number of different spoofing algorithms. The full dataset is partitioned into three subsets, the first for training, the second for development and the third for evaluation. More details can be found in the evaluation plan in the summary paper.

  5. Speech endpoint annotations and artefact details for ASVspoof 2017 version...

    • zenodo.org
    • explore.openaire.eu
    txt
    Updated Jan 24, 2020
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Bhusan Chettri; Bhusan Chettri; Emmanouil Benetos; Emmanouil Benetos; Bob L. Sturm; Bob L. Sturm (2020). Speech endpoint annotations and artefact details for ASVspoof 2017 version 2.0 dataset [Dataset]. http://doi.org/10.5281/zenodo.3601188
    Explore at:
    txtAvailable download formats
    Dataset updated
    Jan 24, 2020
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Bhusan Chettri; Bhusan Chettri; Emmanouil Benetos; Emmanouil Benetos; Bob L. Sturm; Bob L. Sturm
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    This repository contains speech endpoint annotations and filelists for different artefacts we found during our study on the ASVspoof 2017 v2.0 dataset as part of our work in the paper "Dataset biases in speaker verification systems: a case study on the ASVspoof 2017 benchmark" which is to be submitted to the IEEE Transactions on Biometrics, Behavior, and Identity Science (T-BIOM).

  6. h

    asvspoof5

    • huggingface.co
    Updated Aug 3, 2025
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Jee-weon Jung (2025). asvspoof5 [Dataset]. https://huggingface.co/datasets/jungjee/asvspoof5
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Aug 3, 2025
    Authors
    Jee-weon Jung
    Description

    This is the Huggingface repository for the ASVspoof 5 database. ASVspoof 5 is the fifth edition in a series of challenges which promote the study of speech spoofing and deepfake attacks, and the design of detection solutions. Compared to previous challenges, the ASVspoof~5 database is built from crowdsourced data collected from around 2,000 speakers in diverse acoustic conditions. More than 20 attacks, also crowdsourced, are generated and optionally tested using surrogate detection models… See the full description on the dataset page: https://huggingface.co/datasets/jungjee/asvspoof5.

  7. h

    encoded_LA_2021

    • huggingface.co
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Mohamad Bisher tello, encoded_LA_2021 [Dataset]. https://huggingface.co/datasets/Bisher/encoded_LA_2021
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Authors
    Mohamad Bisher tello
    License

    https://choosealicense.com/licenses/odc-by/https://choosealicense.com/licenses/odc-by/

    Description

    ASVspoof 2021 LA subset

    sorted original dataset from ASVspoof 2021 LA subset credit: ASVspoof 2021 challenge released under The databases are available under an Open Data Commons Attribution Licence and can be downloaded from the Zenodo repository.

  8. ASVspoof 2019 LA Listening Test Data for Partial Rank Similarity MOS...

    • zenodo.org
    application/gzip, txt
    Updated Oct 8, 2023
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Junichi Yamagishi; Massimiliano Todisco; Md Sahidullah; Héctor Delgado; Xin Wang; Nicolas Evans; Tomi Kinnunen; Kong Aik Lee; Ville Vestman; Andreas Nautsch; Erica Cooper; Junichi Yamagishi; Massimiliano Todisco; Md Sahidullah; Héctor Delgado; Xin Wang; Nicolas Evans; Tomi Kinnunen; Kong Aik Lee; Ville Vestman; Andreas Nautsch; Erica Cooper (2023). ASVspoof 2019 LA Listening Test Data for Partial Rank Similarity MOS Prediction [Dataset]. http://doi.org/10.5281/zenodo.8412617
    Explore at:
    application/gzip, txtAvailable download formats
    Dataset updated
    Oct 8, 2023
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Junichi Yamagishi; Massimiliano Todisco; Md Sahidullah; Héctor Delgado; Xin Wang; Nicolas Evans; Tomi Kinnunen; Kong Aik Lee; Ville Vestman; Andreas Nautsch; Erica Cooper; Junichi Yamagishi; Massimiliano Todisco; Md Sahidullah; Héctor Delgado; Xin Wang; Nicolas Evans; Tomi Kinnunen; Kong Aik Lee; Ville Vestman; Andreas Nautsch; Erica Cooper
    License

    Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
    License information was derived automatically

    Description

    This dataset is a derivitave work of the ASVSpoof 2019 LA condition listening test data found here:
    https://datashare.ed.ac.uk/handle/10283/3336
    -> LA.zip

    "ASVspoof 2019: A large-scale public database of synthesized, converted and replayed speech"
    Xin Wang, Junichi Yamagishi, Massimiliano Todisco, Héctor Delgado, Andreas Nautsch, Nicholas Evans, Md Sahidullah, Ville Vestman, Tomi Kinnunen, Kong Aik Lee, Lauri Juvela, Paavo Alku, Yu-Huai Peng, Hsin-Te Hwang, Yu Tsao, Hsin-Min Wang, Sébastien Le Maguer, Markus Becker, Fergus Henderson, Rob Clark, Yu Zhang, Quan Wang, Ye Jia, Kai Onuma, Koji Mushika, Takashi Kaneda, Yuan Jiang, Li-Juan Liu, Yi-Chiao Wu, Wen-Chin Huang, Tomoki Toda, Kou Tanaka, Hirokazu Kameoka, Ingmar Steiner, Driss Matrouf, Jean-François Bonastre, Avashna Govender, Srikanth Ronanki, Jing-Xuan Zhang, Zhen-Hua Ling.
    Computer Speech and Language Colume 64, 2020.

    This form of the data was used for the PRS paper accepted to ASRU 2023:

    "Partial Rank Similarity Minimization Method for Quality MOS Prediction of
    Unseen Speech Synthesis Systems in Zero-shot and Semi-supervised Setting."
    Hemant Yadav, Erica Cooper, Junichi Yamagishi, Sunayana Sitaram, Rajiv Ratn Shah.

    Modifications to the original data include converting audio from flac -> wav, sv56 normalization, conversion of labels from an 0-9 rating scale to a 1-5 scale, and creation of training/development/testing splits.

  9. ASVspoof 2019 tfrecord Dataset

    • kaggle.com
    Updated Jun 23, 2022
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Awsaf (2022). ASVspoof 2019 tfrecord Dataset [Dataset]. https://www.kaggle.com/datasets/awsaf49/asvspoof-2019-tfrecord-dataset
    Explore at:
    CroissantCroissant is a format for machine-learning datasets. Learn more about this at mlcommons.org/croissant.
    Dataset updated
    Jun 23, 2022
    Dataset provided by
    Kagglehttp://kaggle.com/
    Authors
    Awsaf
    Description

    Dataset

    This dataset was created by Awsaf

    Contents

  10. Data from: CpAug: Refining Copy-Paste Augmentation for Speech Anti-Spoofing

    • zenodo.org
    application/gzip
    Updated Apr 18, 2024
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Linjuan Zhang; Kong Aik Lee; Lin Zhang; Longbiao Wang; Baoning Niu; Linjuan Zhang; Kong Aik Lee; Lin Zhang; Longbiao Wang; Baoning Niu (2024). CpAug: Refining Copy-Paste Augmentation for Speech Anti-Spoofing [Dataset]. http://doi.org/10.5281/zenodo.10829246
    Explore at:
    application/gzipAvailable download formats
    Dataset updated
    Apr 18, 2024
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Linjuan Zhang; Kong Aik Lee; Lin Zhang; Longbiao Wang; Baoning Niu; Linjuan Zhang; Kong Aik Lee; Lin Zhang; Longbiao Wang; Baoning Niu
    License

    Attribution 4.0 (CC BY 4.0)https://creativecommons.org/licenses/by/4.0/
    License information was derived automatically

    Description

    Conventional copy-paste augmentations generate new training instances by concatenating existing utterances to increase the amount of data for neural network training. However, the direct application of copy-paste augmentation for anti-spoofing is problematic. This paper refines the copy-paste augmentation for speech anti-spoofing, dubbed CpAug, to generate more training data with rich intra-class diversity. The CpAug employs two policies: concatenation to merge utterances with identical labels, and substitution to replace segments in an anchor utterance. Besides, considering the impacts of speakers and spoofing attack types, we craft four blending strategies for the CpAug. Furthermore, we explore how CpAug complements the Rawboost augmentation method. Experimental results reveal that the proposed CpAug significantly improves the performance of speech anti-spoofing. Particularly, CpAug with substitution policy leads to relative improvements of 43% and 38% on the ASVspoof’ 19LA and 21LA, respectively. Notably, the CpAug and Rawboost synergize effectively, achieving an EER of 2.91% on ASVspoof’ 21LA.

  11. Data from: Voice Conversion Challenge 2020 Listening Test Data

    • zenodo.org
    • explore.openaire.eu
    • +1more
    zip
    Updated Jan 15, 2021
    Share
    FacebookFacebook
    TwitterTwitter
    Email
    Click to copy link
    Link copied
    Close
    Cite
    Zhao Yi; Wen-Chin Huang; Xiaohai Tian; Junichi Yamagishi; Rohan Kumar Das; Tomi Kinnunen; Zhenhua Ling; Tomoki Toda; Zhao Yi; Wen-Chin Huang; Xiaohai Tian; Junichi Yamagishi; Rohan Kumar Das; Tomi Kinnunen; Zhenhua Ling; Tomoki Toda (2021). Voice Conversion Challenge 2020 Listening Test Data [Dataset]. http://doi.org/10.5281/zenodo.4433173
    Explore at:
    zipAvailable download formats
    Dataset updated
    Jan 15, 2021
    Dataset provided by
    Zenodohttp://zenodo.org/
    Authors
    Zhao Yi; Wen-Chin Huang; Xiaohai Tian; Junichi Yamagishi; Rohan Kumar Das; Tomi Kinnunen; Zhenhua Ling; Tomoki Toda; Zhao Yi; Wen-Chin Huang; Xiaohai Tian; Junichi Yamagishi; Rohan Kumar Das; Tomi Kinnunen; Zhenhua Ling; Tomoki Toda
    Description
    Voice conversion (VC) is a technique to transform a speaker identity included in a source speech waveform into a different one while preserving linguistic information of the source speech waveform.
    
    In 2016, we have launched the Voice Conversion Challenge (VCC) 2016 [1][2] at Interspeech 2016. The objective of the 2016 challenge was to better understand different VC techniques built on a freely-available common dataset to look at a common goal, and to share views about unsolved problems and challenges faced by the current VC techniques. The VCC 2016 focused on the most basic VC task, that is, the construction of VC models that automatically transform the voice identity of a source speaker into that of a target speaker using a parallel clean training database where source and target speakers read out the same set of utterances in a professional recording studio. 17 research groups had participated in the 2016 challenge. The challenge was successful and it established new standard evaluation methodology and protocols for bench-marking the performance of VC systems.
    
    In 2018, we have launched the second edition of VCC, the VCC 2018 [3]. In the second edition, we revised three aspects of the challenge. First, we educed the amount of speech data used for the construction of participant's VC systems to half. This is based on feedback from participants in the previous challenge and this is also essential for practical applications. Second, we introduced a more challenging task refereed to a Spoke task in addition to a similar task to the 1st edition, which we call a Hub task. In the Spoke task, participants need to build their VC systems using a non-parallel database in which source and target speakers read out different sets of utterances. We then evaluate both parallel and non-parallel voice conversion systems via the same large-scale crowdsourcing listening test. Third, we also attempted to bridge the gap between the ASV and VC communities. Since new VC systems developed for the VCC 2018 may be strong candidates for enhancing the ASVspoof 2015 database, we also asses spoofing performance of the VC systems based on anti-spoofing scores.
    
    In 2020, we launched the third edition of VCC, the VCC 2020 [4][5]. In this third edition, we constructed and distributed a new database for two tasks, intra-lingual semi-parallel and cross-lingual VC. The dataset for intra-lingual VC consists of a smaller parallel corpus and a larger nonparallel corpus, where both of them are of the same language. The dataset for cross-lingual VC consists of a corpus of the source speakers speaking in the source language and another corpus of the target speakers speaking in the target language. As a more challenging task than the previous ones, we focused on cross-lingual VC, in which the speaker identity is transformed between two speakers uttering different languages, which requires handling completely nonparallel training over different languages.
    
    As for listening test, we subcontracted the crowd-sourced perceptual evaluation with English and Japanese listeners to Lionbridge TechnologiesInc. and Koto Ltd., respectively. Given the extremely large costs required for the perceptual evaluation, we selected 5 utterances (E30001, E30002, E30003,E30004, E30005) only from each speaker of each team. To evaluate the speaker similarity of the cross-lingual task, we used audio in both the English language and in the target speaker’s L2language as reference. For each source-target speaker pair, we selected three English recordings and two L2 language recordings as the natural reference for the converted five utterances. 
    

    This data repository includes the audio files used for the crowd-sourced perceptual evaluation and raw listening test scores.

    [1] Tomoki Toda, Ling-Hui Chen, Daisuke Saito, Fernando Villavicencio, Mirjam Wester, Zhizheng Wu, Junichi Yamagishi "The Voice Conversion Challenge 2016" in Proc. of Interspeech, San Francisco.
    
    [2] Mirjam Wester, Zhizheng Wu, Junichi Yamagishi "Analysis of the Voice Conversion Challenge 2016 Evaluation Results" in Proc. of Interspeech 2016.
    
    [3] Jaime Lorenzo-Trueba, Junichi Yamagishi, Tomoki Toda, Daisuke Saito, Fernando Villavicencio, Tomi Kinnunen, Zhenhua Ling, "The Voice Conversion Challenge 2018: Promoting Development of Parallel and Nonparallel Methods", Proc Speaker Odyssey 2018, June 2018.
    
    [4] Yi Zhao, Wen-Chin Huang, Xiaohai Tian, Junichi Yamagishi, Rohan Kumar Das, Tomi Kinnunen, Zhenhua Ling, and Tomoki Toda. "Voice conversion challenge 2020: Intra-lingual semi-parallel and cross-lingual voice conversion" Proc. Joint Workshop for the Blizzard Challenge and Voice Conversion Challenge 2020, 80-98, DOI: 10.21437/VCC_BC.2020-14.
    
    [5] Rohan Kumar Das, Tomi Kinnunen, Wen-Chin Huang, Zhenhua Ling, Junichi Yamagishi, Yi Zhao, Xiaohai Tian, and Tomoki Toda. "Predictions of subjective ratings and spoofing assessments of voice conversion challenge 2020 submissions." Proc. Joint Workshop for the Blizzard Challenge and Voice Conversion Challenge 2020, 99-120, DOI: 10.21437/VCC_BC.2020-15.
    
  12. Not seeing a result you expected?
    Learn how you can add new datasets to our index.

Share
FacebookFacebook
TwitterTwitter
Email
Click to copy link
Link copied
Close
Cite
Xin Wang; Xin Wang; Héctor Delgado; Héctor Delgado; Hemlata Tak; Hemlata Tak; Jee-weon Jung; Jee-weon Jung; Hyejin Shim; Hyejin Shim; Massimiliano Todisco; Massimiliano Todisco; Ivan Kukanov; Ivan Kukanov; Xuechen Liu; Xuechen Liu; Md Sahidullah; Md Sahidullah; Tomi Kinnunen; Tomi Kinnunen; Nicholas Evans; Nicholas Evans; Kong Aik Lee; Kong Aik Lee; Junichi Yamagishi; Junichi Yamagishi; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh (2025). ASVspoof 5: Design, Collection and Validation of Resources for Spoofing, Deepfake, and Adversarial Attack Detection Using Crowdsourced Speech [Dataset]. http://doi.org/10.5281/zenodo.14498691
Organization logo

ASVspoof 5: Design, Collection and Validation of Resources for Spoofing, Deepfake, and Adversarial Attack Detection Using Crowdsourced Speech

Explore at:
2 scholarly articles cite this dataset (View in Google Scholar)
tar, application/gzip, txtAvailable download formats
Dataset updated
Feb 14, 2025
Dataset provided by
Zenodohttp://zenodo.org/
Authors
Xin Wang; Xin Wang; Héctor Delgado; Héctor Delgado; Hemlata Tak; Hemlata Tak; Jee-weon Jung; Jee-weon Jung; Hyejin Shim; Hyejin Shim; Massimiliano Todisco; Massimiliano Todisco; Ivan Kukanov; Ivan Kukanov; Xuechen Liu; Xuechen Liu; Md Sahidullah; Md Sahidullah; Tomi Kinnunen; Tomi Kinnunen; Nicholas Evans; Nicholas Evans; Kong Aik Lee; Kong Aik Lee; Junichi Yamagishi; Junichi Yamagishi; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh; Myeonghun Jeong; Ge Zhu; Yongyi Zang; Neil Zhang; Soumi Maiti; Florian Lux; Nicolas Muller; Wangyou Zhang; Chengzhe Sun; Shuwei Hou; Siwei Lyu; Sebastien Le Maguer; Cheng Gong; Hanjie Guo; Liping Chen; Vishwanath Singh
License

Open Data Commons Attribution License (ODC-By) v1.0https://www.opendatacommons.org/licenses/by/1.0/
License information was derived automatically

Time period covered
Dec 2024
Description

This is the Zenodo repository for the ASVspoof 5 database. ASVspoof 5 is the fifth edition in a series of challenges which promote the study of speech spoofing and deepfake attacks, and the design of detection solutions. Compared to previous challenges, the ASVspoof~5 database is built from crowdsourced data collected from around 2,000 speakers in diverse acoustic conditions. More than 20 attacks, also crowdsourced, are generated and optionally tested using surrogate detection models, while seven adversarial attacks are incorporated for the first time.

  • Please check README.txt and LICENSE.txt before downloading the database.
  • Database paper (to be submitted): https://arxiv.org/abs/2502.08857
    • Please consider citing the reference listed at the bottom of this page.
  • It is highly recommended to follow the rules and instructions in the ASVspoof 5 challenge evaluation plan (phase 2, https://www.asvspoof.org/), if you want to produce results comparable with the literature.
  • Latest work using the ASVspoof 5 database can be found in the Automatic Speaker Verification Spoofing Countermeasures Workshop proceeding: https://www.isca-archive.org/asvspoof_2024/index.html
  • If you are interested in creating spoofed data for research purpose using the ASVspoof 5 protocol, please send request to info@asvspoof.org
Search
Clear search
Close search
Google apps
Main menu