Direction-Aware Joint Adaptation of Neural Speech Enhancement and Recognition in Real Multiparty Conversational Environments
説明
This paper describes noisy speech recognition for an augmented reality headset that helps verbal communication within real multiparty conversational environments. A major approach that has actively been studied in simulated environments is to sequentially perform speech enhancement and automatic speech recognition (ASR) based on deep neural networks (DNNs) trained in a supervised manner. In our task, however, such a pretrained system fails to work due to the mismatch between the training and test conditions and the head movements of the user. To enhance only the utterances of a target speaker, we use beamforming based on a DNN-based speech mask estimator that can adaptively extract the speech components corresponding to a head-relative particular direction. We propose a semi-supervised adaptation method that jointly updates the mask estimator and the ASR model at run-time using clean speech signals with ground-truth transcriptions and noisy speech signals with highly-confident estimated transcriptions. Comparative experiments using the state-of-the-art distant speech recognition system show that the proposed method significantly improves the ASR performance.
INTERSPEECH 2022
収録刊行物
-
- Interspeech 2022
-
Interspeech 2022 2918-2922, 2022-09-18
ISCA
- Tweet
キーワード
- FOS: Computer and information sciences
- Computer Science - Machine Learning
- Sound (cs.SD)
- speech recognition
- Computer Science - Sound
- [INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI]
- Machine Learning (cs.LG)
- Audio and Speech Processing (eess.AS)
- [INFO.INFO-SD]Computer Science [cs]/Sound [cs.SD]
- FOS: Electrical engineering, electronic engineering, information engineering
- speech enhancement
- run-time adaptation
- humancomputer interaction
- Electrical Engineering and Systems Science - Audio and Speech Processing
詳細情報 詳細情報について
-
- CRID
- 1873961342896002560
-
- データソース種別
-
- OpenAIRE