University of Surrey

Test tubes in the lab Research in the ATI Dance Research

Multimodal Information Fusion

Poh, N and Kittler, J (2010) Multimodal Information Fusion pp. 153-169.

Full text not available from this repository.


This chapter gives an overview of multimodal information fusion from the machine-learning perspective. Humans interact with each other using different modalities of communication. These include speech, gestures, documents, etc. It is therefore natural that human-computer interaction (HCI) should facilitate the same multimodal form of communication. To capture this information, one uses different types of sensors, i.e., microphones to capture the audio signal, cameras to capture life video images, 3D sensors to directly capture the surface information in real time. In each of these cases, commercial off-the-shelf (COTS) devices are already available and can be readily deployed for HCI applications. Examples of HCI applications include audio-visual speech recognition, gesture recognition, emotional recognition, and person recognition using biometrics. © 2010 Elsevier Ltd All rights reserved.

Item Type: Article
Divisions : Surrey research (other units)
Authors :
Kittler, J
Date : 1 December 2010
DOI : 10.1016/B978-0-12-374825-6.00017-4
Depositing User : Symplectic Elements
Date Deposited : 17 May 2017 13:01
Last Modified : 24 Jan 2020 23:16

Actions (login required)

View Item View Item


Downloads per month over past year

Information about this web site

© The University of Surrey, Guildford, Surrey, GU2 7XH, United Kingdom.
+44 (0)1483 300800