EMG signal processing for audio-emg-based multi-modal speech recognition

Zhipeng Zhang, Hiroyuki Manabe, Tsutomu Horikoshi, Tomoyuki Ohya

Research output: Chapter in Book/Report/Conference proceedingConference contribution

1 Citation (Scopus)

Abstract

This paper proposes robust methods for processing EMG (electromyography) signals in the framework of audio-EMG-based speech recognition. The EMG signals are captured when uttered and used as auxiliary information for recognizing speech. Two robust methods (Cepstral Mean Normalization and Spectral Subtraction) for EMG signal processing are investigated to improve the recognition performance. We also investigate the importance of stream weighting in audio-EMG-based multi-modal speech recognition. Experiments are carried out at various noise conditions and the results show the effectiveness of the proposed methods. A significant improvement in word accuracy over the audio-only recognition scheme is achieved by combining the methods.

Original languageEnglish
Title of host publicationProceedings of the 3rd IASTED International Conference on Biomedical Engineering 2005
EditorsM.H. Hamza
Pages430-433
Number of pages4
Publication statusPublished - 2005
Externally publishedYes
Event3rd IASTED International Conference on Medical Engineering 2005 - Innsbruck, Austria
Duration: 2005 Feb 162005 Feb 18

Publication series

NameProceedings of the 3rd IASTED International Conference on Biomedical Engineering 2005

Conference

Conference3rd IASTED International Conference on Medical Engineering 2005
Country/TerritoryAustria
CityInnsbruck
Period05/2/1605/2/18

Keywords

  • Cepstral Mean Subtraction
  • EMG
  • Multimodal speech recognition
  • Spectral Subtraction

ASJC Scopus subject areas

  • Engineering(all)

Fingerprint

Dive into the research topics of 'EMG signal processing for audio-emg-based multi-modal speech recognition'. Together they form a unique fingerprint.

Cite this