Text-dependent Speaker Verification (TdSV) Challenge 2024
Short-duration speaker verification in two different configurations
Following the success of two previous Short-duration Speaker Verification Challenges, this challenge aims to focus researchers’ attention on the relevance of recent training strategies, such as self-supervised learning, in the context of text-dependent speaker verification (TdSV). The challenge evaluates the TdSV task in the following two practical scenarios:
Task 1: Conventional TdSV Using Predefined Passphrases
Participants are tasked with training a speaker encoder model using a significantly large training dataset sourced from a predefined phrase pool consisting of 10 phrases. Speaker models are created using three repetitions of a specific passphrase from the phrase pool. Speakers are then verified using an utterance containing the passphrase.
Task 2: TdSV Using User-defined Passphrases
Participants are required to train a speaker encoder model using a large text-independent training dataset. Additionally, there are utterances from a predefined pool of 6 phrases for each in-domain training speaker. Speaker models are created using three repetitions of a user-defined passphrase (not included in the phrase pool), along with some free-text utterances. Speakers are verified using an utterance containing the user-defined passphrase.
Challenge Prizes
There will be three cash prizes for each task. Winners will be selected based on the results of the evaluation dataset and other qualitative factors. In addition to the cash prize, each winner will receive a certificate for their achievement. The cash prizes for each task are as follows:
- Rank 1: 2000 USD
- Rank 2: 1000 USD
- Rank 3: 500 USD
Challenge Dataset
The evaluation dataset of the challenge is drawn from the new versions of the multi-purpose DeepMine dataset[1]. The dataset has three parts and among them, Parts 1 and 3 are used in this challenge.
[1] H. Zeinali, L. Burget, J. Cernocky, “A multi-purpose and large scale speech corpus in Persian and English for speaker and speech recognition: the DeepMine database”, in: Proc. ASRU 2019 The 2019 IEEE Automatic Speech Recognition and Understanding Workshop, 2019.
Challenge Evaluation Plan
The full challenge evaluation plan version 1.0 can be found at this link and the high-quality challenge poster is available through this link. If you have any more questions regarding the challenge you can contact organizers via tdsvc.2024[at]gmail.com.
Planned Evaluation Schedule
Apr 10, 2024 | Release of Evaluation Plan |
Apr 10, 2024 | Release of Train, Dev, and Eval Sets |
Apr 20, 2024 | Evaluation Platform Open |
Jun 10, 2024 | Challenge Deadline |
Jun 17, 2024 | System Description Deadline |
Jun 20, 2024 | SLT Paper Submission Deadline |
Dec 02, 2024 | TdSV Challenge 2024 Special Session at SLT |
Sponsors
Amirkabir University of Technology (Tehran Polytechnic) | |
Sharif DeepMine Ltd. |