Auditory scene analysis based on time-frequency integration of shared FM and AM

説明

This paper describes a new method for computational auditory scene analysis which is based on (1) waveform operators to extract instantaneous frequency (IF), frequency change (FM), and amplitude change (AM) from subband signals, and (2) the introduction of a voting method into a probability distribution function to extract coherency (shared fundamental frequency, shared FM, and shared AM) involved in them. We introduce non-parametric Kalman filtering for the time-axis integration. A consistent AM operator which is independent of frequency change is newly defined. The sharpness of the resultant probability distribution is examined with relation to the definition of the operators and subband bandwidth. We evaluate the performance of the algorithm by using several speech sounds.

収録刊行物

詳細情報 詳細情報について

問題の指摘

ページトップへ