• The 2nd Automatic Speaker Verification Spoofing and Countermeasures Challenge (ASVspoof 2017) Database, Version 2 

      Kinnunen, Tomi; Sahidullah, Md; Delgado, Héctor; Todisco, Massimiliano; Evans, Nicholas; Yamagishi, Junichi; Lee, Kong Aik
      This is a database used for the Second Automatic Speaker Verification Spoofing and Countermeasuers Challenge, for short, ASVspoof 2017 (http://www.asvspoof.org) organized by Tomi Kinnunen, Md Sahidullah, Héctor Delgado, ...
    • 96kHz version of the CSTR VCTK Corpus 

      Veaux, Christophe; Yamagishi, Junichi
      This dataset includes 96kHz version of the CSTR VCTK Corpus including speech data uttered by 109 native speakers of English with various accents. The main dataset can be found at https://doi.org/10.7488/ds/1994 (containing ...
    • Acted clear speech corpus 

      Mayo, Catherine (LISTA Consortium: (i) Language and Speech Laboratory, Universidad del Pais Vasco, Spain and Ikerbasque, Spain; (ii) Centre for Speech Technology Research, University of Edinburgh, UK; (iii) KTH Royal Institute of Technology, Sweden; (iv) Institute of Computer Science, FORTH, Greece, 2013-09-24)
      Single male native British English talker recorded producing 25 TIMIT sentences in 5 conditions, two natural: (i) quiet, (ii) while the talker listened to high-intensity speech-shaped noise, and three acted: (i) as if to ...
    • Alba speech corpus 

      Valentini-Botinhao, Cassia; Yamagishi, Junichi
      Single speaker read speech corpus of a Scottish accented female native English speaker (Alba). The corpus was recorded in four speaking styles: plain (normal read speech, around 4 hours of recordings), fast (speaking as ...
    • Artificial Personality 

      Wester, Mirjam; Aylett, Matthew; Tomalin, Marcus; Dall, Rasmus
      This dataset is associated with the paper “Artificial Personality and Disfluency” by Mirjam Wester, Matthew Aylett, Marcus Tomalin and Rasmus Dall published at Interspeech 2015, Dresden. The focus of this paper is ...
    • ASVspoof 2019: The 3rd Automatic Speaker Verification Spoofing and Countermeasures Challenge database 

      Yamagishi, Junichi; Todisco, Massimiliano; Sahidullah, Md; Delgado, Héctor; Wang, Xin; Evans, Nicolas; Kinnunen, Tomi; Lee, Kong Aik; Vestman, Ville; Nautsch, Andreas
      This is a database used for the Third Automatic Speaker Verification Spoofing and Countermeasures Challenge, for short, ASVspoof 2019 (http://www.asvspoof.org) organized by Junichi Yamagishi, Massimiliano Todisco, Md ...
    • Automatic Speaker Verification Spoofing and Countermeasures Challenge (ASVspoof 2015) Database 

      Wu, Zhizheng; Kinnunen, Tomi; Evans, Nicholas; Yamagishi, Junichi
      The database has been used in the first Automatic Speaker Verification Spoofing and Countermeasures Challenge (ASVspoof 2015). Genuine speech is collected from 106 speakers (45 male, 61 female) and with no significant channel ...
    • CSTR VCTK Corpus: English Multi-speaker Corpus for CSTR Voice Cloning Toolkit 

      Veaux, Christophe; Yamagishi, Junichi; MacDonald, Kirsten
      This CSTR VCTK Corpus (Centre for Speech Technology Voice Cloning Toolkit) includes speech data uttered by 109 native speakers of English with various accents. 96kHz versions of the recordings are available at https://do ...
    • Device Recorded VCTK (Small subset version) 

      Sarfjoo, Seyyed Saeed; Yamagishi, Junichi
      This dataset is a new variant of the voice cloning toolkit (VCTK) dataset: device-recorded VCTK (DR-VCTK), where the high-quality speech signals recorded in a semi-anechoic chamber using professional audio devices are ...
    • DiapixFL 

      Cooke, Martin; Garcia Lecumberri, Maria Luisa; Wester, Mirjam (LISTA Consortium: (i) Language and Speech Laboratory, Universidad del Pais Vasco, Spain and Ikerbasque, Spain; (ii) Centre for Speech Technology Research, University of Edinburgh, UK; (iii) KTH Royal Institute of Technology, Sweden; (iv) Institute of Computer Science, FORTH, Greece., 2013-10-01)
      DiapixFL consists of speakers whose first language (L1) is either English or Spanish solving a "spot-the-difference" task in both their L1 and their second language (L2, which for native English talkers is Spanish, and for ...
    • Dutch English Lombard Speech Native and Non-Native (DELNN) 

      Marcoux, Katherine; Ernestus, Mirjam; King, Simon
      The DELNN (Dutch English Lombard speech Native and Non-Native) corpus consists of 30 native Dutch speakers reading sentences in a quiet environment and in a noisy environment, to elicit Lombard speech. The Dutch speakers ...
    • Experiment materials for "Disfluencies in change detection in natural, vocoded and synthetic speech." 

      Dall, Rasmus; Wester, Mirjam; Corley, Martin
      The current dataset is associated with the DiSS paper "Disfluencies in change detection in natural, vocoded and synthetic speech." In this paper we investigate the effect of filled pauses, a discourse marker and silent ...
    • Experiment materials for "Testing the consistency assumption: pronunciation variant forced alignment in read and spontaneous speech synthesis" 

      Dall, Rasmus
      The matlab scripts are used to analyse the results files in the results folder. The Test_Wavs are the wavfiles used for the listening test divided by group and the pre-test test files.
    • Experiment materials for "The temporal delay hypothesis: Natural, vocoded and synthetic speech." 

      Corley, Martin; Dall, Rasmus; Wester, Mirjam
      Including disfluencies in synthetic speech is being explored as a way of making synthetic speech sound more natural and conversational. How to measure whether the resulting speech is actually more natural, however, is not ...
    • Human vs Machine Spoofing 

      Wester, Mirjam; Wu, Zhizheng; Yamagishi, Junichi
      Listening test materials for "Human vs Machine Spoofing Detection on Wideband and Narrowband data." They include lists of the speech material selected from the SAS spoofing database and the listeners' responses. The main ...
    • Hurricane natural speech corpus 

      Cooke, Martin; Mayo, Catherine; Valentini-Botinhao, Cassia (LISTA Consortium: (i) Language and Speech Laboratory, Universidad del Pais Vasco, Spain and Ikerbasque, Spain; (ii) Centre for Speech Technology Research, University of Edinburgh, UK; (iii) KTH Royal Institute of Technology, Sweden; (iv) Institute of Computer Science, FORTH, Greece, 2013-10-01)
      Single male native British-English talker recorded producing three speech sets (Harvard sentences, Modified Rhyme Test, news sentences) in quiet and while the talker was listening to speech-shaped noise at 84dB(A). A higher ...
    • Hurricane natural speech corpus - higher quality version 

      Valentini-Botinhao, Cassia; Mayo, Cassie; Cooke, Martin
      Single male native British-English talker recorded producing three speech sets (Harvard sentences, Modified Rhyme Test, news sentences) in quiet and while the talker was listening to speech-shaped noise at 84dB(A). This ...
    • Key files for Spoofing and Anti-Spoofing (SAS) corpus v1.0 

      Wu, Zhizheng; Khodabakhsh, Ali; Demiroglu, Cenk; Yamagishi, Junichi; Saito, Daisuke; Toda, Tomoki; Ling, Zhen-Hua; King, Simon
      These files are complementary to the fileset: Wu et al. (2015). Spoofing and Anti-Spoofing (SAS) corpus v1.0, [dataset]. University of Edinburgh. The Centre for Speech Technology Research (CSTR). https://doi.org/10.7488/ds/252. ...
    • Listening test materials for "A study of speaker adaptation for DNN-based speech synthesis" 

      Wu, Zhizheng
      The dataset contains the testing stimuli and listeners' MUSHRA test responses for the Interspeech 2015 paper, "A study of speaker adaptation for DNN-based speech synthesis". In this paper, we conduct an experimental analysis ...
    • Listening test materials for "A template-based approach for speech synthesis intonation generation using LSTMs" 

      Ronanki, Srikanth; Henter, Gustav Eje; Wu, Zhizheng; King, Simon
      This data release contains listening test materials associated with the paper "A template-based approach for speech synthesis intonation generation using LSTMs", presented at Interspeech 2016 in San Francisco, USA.