University of Surrey

Test tubes in the lab Research in the ATI Dance Research

Predicting the Perceived Level of Reverberation using Features from Nonlinear Auditory Model

Safavi, Saeid, Wang, Wenwu, Plumbley, Mark, Choobbasti, Ali Janalizadeh and Fazekas, George (2018) Predicting the Perceived Level of Reverberation using Features from Nonlinear Auditory Model In: 23rd FRUCT conference, 13-16 Nov 2018, Bologna, Italy.

[img]
Preview
Text
Saf.pdf - Accepted version Manuscript
Available under License Creative Commons Attribution No Derivatives.

Download (615kB) | Preview
[img] Text
Predicting the Perceived Level of Reverberation using Features from Nonlinear Auditory Model.pdf - Accepted version Manuscript
Restricted to Repository staff only
Available under License Creative Commons Attribution No Derivatives.

Download (358kB)

Abstract

Perceptual measurements have typically been recognized as the most reliable measurements in assessing perceived levels of reverberation. In this paper, a combination of blind RT60 estimation method and a binaural, nonlinear auditory model is employed to derive signal-based measures (features) that are then utilized in predicting the perceived level of reverberation. Such measures lack the excess of effort necessary for calculating perceptual measures; not to mention the variations in either stimuli or assessors that may cause such measures to be statistically insignificant. As a result, the automatic extraction of objective measurements that can be applied to predict the perceived level of reverberation become of vital significance. Consequently, this work is aimed at discovering measurements such as clarity, reverberance, and RT60 which can automatically be derived directly from audio data. These measurements along with labels from human listening tests are then forwarded to a machine learning system seeking to build a model to estimate the perceived level of reverberation, which is labeled by an expert, autonomously. The data has been labeled by an expert human listener for a unilateral set of files from arbitrary audio source types. By examining the results, it can be observed that the automatically extracted features can aid in estimating the perceptual rates.

Item Type: Conference or Workshop Item (Conference Paper)
Divisions : Faculty of Engineering and Physical Sciences > Electronic Engineering
Authors :
NameEmailORCID
Safavi, Saeids.safavi@surrey.ac.uk
Wang, WenwuW.Wang@surrey.ac.uk
Plumbley, Markm.plumbley@surrey.ac.uk
Choobbasti, Ali Janalizadeh
Fazekas, George
Date : 16 November 2018
Funders : H2020; European Commission
Grant Title : Grant Agreement number 688382
Copyright Disclaimer : © 2018 The authors.
Related URLs :
Depositing User : Clive Harris
Date Deposited : 16 Oct 2018 10:21
Last Modified : 11 Dec 2018 11:24
URI: http://epubs.surrey.ac.uk/id/eprint/849679

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year


Information about this web site

© The University of Surrey, Guildford, Surrey, GU2 7XH, United Kingdom.
+44 (0)1483 300800