Data-driven Speech Segmentation for Language Identification and Speaker Verification

Dijana Petrovska-Delacrétaz, Marcos Abalo, Asmaa El Hannani, Gérard Chollet

Research output: Contribution to conferencePaperpeer-review

Abstract

The common denominator of many speech processing methods is the set of acoustic units chosen to represent the structure of the data. The majority of current systems use phones (or related units) as an atomic representation of speech. The major problems that arise when phone based systems are being developed is the possible mismatch with the data being used and the lack of transcribed databases. The set of speech units can also be learned from examples, like in data-driven approaches. We have already used data-driven acoustic speech units, denoted as Automatic Language Independent Speech Processing (ALISP) units for segmental speaker verification experiments, based on Multiple Layer Perceptrons, and on Dynamic Time Warping (DTW). In this paper we give an overview of the DTW based speaker verification and we present further developments of the data-driven ALISP speech segmentation for language identification experiments. The results confirm the applicability of the proposed method for these two tasks.

Original languageEnglish
Publication statusPublished - 1 Jan 2003
Externally publishedYes
Event2003 ISCA Tutorial and Research Workshop on Nonlinear Speech Processing, NOLISP 2003 - Le Croisic, France
Duration: 20 May 200323 May 2003

Conference

Conference2003 ISCA Tutorial and Research Workshop on Nonlinear Speech Processing, NOLISP 2003
Country/TerritoryFrance
CityLe Croisic
Period20/05/0323/05/03

Fingerprint

Dive into the research topics of 'Data-driven Speech Segmentation for Language Identification and Speaker Verification'. Together they form a unique fingerprint.

Cite this