A New Temporal Filtering Method for Improved Automatic Lipreading


The KIPS Transactions:PartB , Vol. 15, No. 2, pp. 123-130, Apr. 2008
10.3745/KIPSTB.2008.15.2.123,   PDF Download:

Abstract

Automatic lipreading is to recognize speech by observing the movement of a speaker's lips. It has received attention recently as a method of complementing performance degradation of acoustic speech recognition in acoustically noisy environments. One of the important issues in automatic lipreading is to define and extract salient features from the recorded images. In this paper, we propose a feature extraction method by using a new filtering technique for obtaining improved recognition performance. The proposed method eliminates frequency components which are too slow or too fast compared to the relevant speech information by applying a band-pass filter to the temporal trajectory of each pixel in the images containing the lip region and, then, features are extracted by principal component analysis. We show that the proposed method produces improved performance in both clean and visually noisy conditions via speaker-independent recognition experiments.


Statistics
Show / Hide Statistics

Statistics (Cumulative Counts from September 1st, 2017)
Multiple requests among the same browser session are counted as one view.
If you mouse over a chart, the values of data points will be shown.


Cite this article
[IEEE Style]
J. S. Lee and C. H. Park, "A New Temporal Filtering Method for Improved Automatic Lipreading," The KIPS Transactions:PartB , vol. 15, no. 2, pp. 123-130, 2008. DOI: 10.3745/KIPSTB.2008.15.2.123.

[ACM Style]
Jong Seok Lee and Cheol Hoon Park. 2008. A New Temporal Filtering Method for Improved Automatic Lipreading. The KIPS Transactions:PartB , 15, 2, (2008), 123-130. DOI: 10.3745/KIPSTB.2008.15.2.123.