University of Surrey

Test tubes in the lab Research in the ATI Dance Research

Two novel visual voice activity detectors based on appearance models and retinal filtering

Aubrey, A, Rivet, B, Hicks, Y, Girin, L, Chambers, J and Jutten, C (2007) Two novel visual voice activity detectors based on appearance models and retinal filtering

Full text not available from this repository.

Abstract

In this paper we present two novel methods for visual voice activity detection (V-VAD) which exploit the bimodality of speech (i.e. the coherence between speaker's lips and the resulting speech). The first method uses appearance parameters of a speaker's lips, obtained from an active appearance model (AAM). An HMM then dynamically models the change in appearance over time. The second method uses a retinal filter on the region of the lips to extract the required parameter. A corpus of a single speaker is applied to each method in turn, where each method is used to classify voice activity as speech or non speech. The efficiency of each method is evaluated individually using receiver operating characteristics and their respective performances are then compared and discussed. Both methods achieve a high correct silence detection rate for a small false detection rate. © 2007 EURASIP.

Item Type: Conference or Workshop Item (UNSPECIFIED)
Authors :
NameEmailORCID
Aubrey, AUNSPECIFIEDUNSPECIFIED
Rivet, BUNSPECIFIEDUNSPECIFIED
Hicks, YUNSPECIFIEDUNSPECIFIED
Girin, LUNSPECIFIEDUNSPECIFIED
Chambers, Jj.a.chambers@surrey.ac.ukUNSPECIFIED
Jutten, CUNSPECIFIEDUNSPECIFIED
Date : 1 December 2007
Depositing User : Symplectic Elements
Date Deposited : 17 May 2017 13:26
Last Modified : 17 May 2017 13:26
URI: http://epubs.surrey.ac.uk/id/eprint/839204

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year


Information about this web site

© The University of Surrey, Guildford, Surrey, GU2 7XH, United Kingdom.
+44 (0)1483 300800