Nothing Special   »   [go: up one dir, main page]

Skip to main content
  • Research Article
  • Published:

Automatic Speechreading with Applications to Human-Computer Interfaces

Abstract

There has been growing interest in introducing speech as a new modality into the human-computer interface (HCI). Motivated by the multimodal nature of speech, the visual component is considered to yield information that is not always present in the acoustic signal and enables improved system performance over acoustic-only methods, especially in noisy environments. In this paper, we investigate the usefulness of visual speech information in HCI related applications. We first introduce a new algorithm for automatically locating the mouth region by using color and motion information and segmenting the lip region by making use of both color and edge information based on Markov random fields. We then derive a relevant set of visual speech parameters and incorporate them into a recognition engine. We present various visual feature performance comparisons to explore their impact on the recognition accuracy, including the lip inner contour and the visibility of the tongue and teeth. By using a common visual feature set, we demonstrate two applications that exploit speechreading in a joint audio-visual speech signal processing task: speech recognition and speaker verification. The experimental results based on two databases demonstrate that the visual information is highly effective for improving recognition performance over a variety of acoustic noise levels.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Xiaozheng Zhang.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Zhang, X., Broun, C.C., Mersereau, R.M. et al. Automatic Speechreading with Applications to Human-Computer Interfaces. EURASIP J. Adv. Signal Process. 2002, 240192 (2002). https://doi.org/10.1155/S1110865702206137

Download citation

  • Received:

  • Revised:

  • Published:

  • DOI: https://doi.org/10.1155/S1110865702206137

Keywords