University of Surrey

Test tubes in the lab Research in the ATI Dance Research

Multimodal Information Fusion

Poh, N and Kittler, J (2010) Multimodal Information Fusion pp. 153-169.

Full text not available from this repository.

Abstract

This chapter gives an overview of multimodal information fusion from the machine-learning perspective. Humans interact with each other using different modalities of communication. These include speech, gestures, documents, etc. It is therefore natural that human-computer interaction (HCI) should facilitate the same multimodal form of communication. To capture this information, one uses different types of sensors, i.e., microphones to capture the audio signal, cameras to capture life video images, 3D sensors to directly capture the surface information in real time. In each of these cases, commercial off-the-shelf (COTS) devices are already available and can be readily deployed for HCI applications. Examples of HCI applications include audio-visual speech recognition, gesture recognition, emotional recognition, and person recognition using biometrics. © 2010 Elsevier Ltd All rights reserved.

Item Type: Article
Authors :
NameEmailORCID
Poh, Nn.poh@surrey.ac.ukUNSPECIFIED
Kittler, JUNSPECIFIEDUNSPECIFIED
Date : 1 December 2010
Identification Number : 10.1016/B978-0-12-374825-6.00017-4
Depositing User : Symplectic Elements
Date Deposited : 17 May 2017 13:01
Last Modified : 17 May 2017 15:07
URI: http://epubs.surrey.ac.uk/id/eprint/837608

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year


Information about this web site

© The University of Surrey, Guildford, Surrey, GU2 7XH, United Kingdom.
+44 (0)1483 300800