On evaluation trials in speaker verification

Lantian Li, Di Wang, Andrew Abel, Dong Wang*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Evaluation trials are crucial to measure performance of speaker verification systems. However, the design of trials that can faithfully reflect system performance and accurately distinguish between different systems remains an open issue. In this paper, we focus on a particular problem: the impact of trials that are easy to solve for the majority of systems. We show that these ‘easy trials’ not only report over-optimistic absolute performance, but also lead to biased relative performance in system comparisons when they are asymmetrically distributed. This motivated the idea of mining ‘hard trials’, i.e., trials that are regarded to be difficult by current representative techniques. Three approaches to retrieving hard trials will be reported, and the properties of the retrieved hard trials are studied, from the perspectives of both machines and humans. Finally, a novel visualization tool which we name a Config-Performance (C-P) map is proposed. In this map, the value at each location represents the performance with a particular proportion of easy and hard trials, thus offering a global view of the system in various test conditions. The identified hard trials and the code of the C-P map tool have been released at http://lilt.cslt.org/trials/demo/ .

Original languageEnglish
Pages (from-to)113-130
Number of pages18
JournalApplied Intelligence
Volume54
Issue number1
Early online date5 Dec 2023
DOIs
Publication statusPublished - Jan 2024

Keywords

  • Performance evaluation
  • Speaker verification
  • Test trials

Fingerprint

Dive into the research topics of 'On evaluation trials in speaker verification'. Together they form a unique fingerprint.

Cite this