Text-dependent Speaker Verification (TdSV) Challenge 2024

Short-duration speaker verification in two different configurations

tdsvc_poster.jpg

Following the success of two previous Short-duration Speaker Verification Challenges, this challenge aims to focus researchers’ attention on the relevance of recent training strategies, such as self-supervised learning, in the context of text-dependent speaker verification (TdSV). The challenge evaluates the TdSV task in the following two practical scenarios:

Task 1: Conventional TdSV Using Predefined Passphrases

Participants are tasked with training a speaker encoder model using a significantly large training dataset sourced from a predefined phrase pool consisting of 10 phrases. Speaker models are created using three repetitions of a specific passphrase from the phrase pool. Speakers are then verified using an utterance containing the passphrase.

Task 2: TdSV Using User-defined Passphrases

Participants are required to train a speaker encoder model using a large text-independent training dataset. Additionally, there are utterances from a predefined pool of 6 phrases for each in-domain training speaker. Speaker models are created using three repetitions of a user-defined passphrase (not included in the phrase pool), along with some free-text utterances. Speakers are verified using an utterance containing the user-defined passphrase.


Challenge Prizes

There will be three cash prizes for each task. Winners will be selected based on the results of the evaluation dataset and other qualitative factors. In addition to the cash prize, each winner will receive a certificate for their achievement. The cash prizes for each task are as follows:

  • Rank 1: 2000 USD
  • Rank 2: 1000 USD
  • Rank 3: 500 USD

Challenge Dataset

The evaluation dataset of the challenge is drawn from the new versions of the multi-purpose DeepMine dataset[1]. The dataset has three parts and among them, Parts 1 and 3 are used in this challenge.

[1] H. Zeinali, L. Burget, J. Cernocky, “A multi-purpose and large scale speech corpus in Persian and English for speaker and speech recognition: the DeepMine database”, in: Proc. ASRU 2019 The 2019 IEEE Automatic Speech Recognition and Understanding Workshop, 2019.


Challenge Evaluation Plan

The full challenge evaluation plan version 1.0 can be found at this link and the high-quality challenge poster is available through this link. If you have any more questions regarding the challenge you can contact organizers via tdsvc.2024[at]gmail.com.


Planned Evaluation Schedule

Apr 10, 2024 Release of Evaluation Plan
Apr 10, 2024 Release of Train, Dev, and Eval Sets
Apr 20, 2024 Evaluation Platform Open
Jun 10, 2024 Challenge Deadline
Jun 17, 2024 System Description Deadline
Jun 20, 2024 SLT Paper Submission Deadline
Dec 02, 2024      TdSV Challenge 2024 Special Session at SLT
 

Sponsors

Amirkabir University of Technology (Tehran Polytechnic)     
Sharif DeepMine Ltd.