计算机科学
耳蜗
语音识别
非线性系统
整改
过滤器组
语音处理
频域
信号处理
LTI系统理论
人工神经网络
滤波器(信号处理)
人工智能
计算机视觉
线性系统
工程类
数学
数字信号处理
量子力学
医学
解剖
电气工程
物理
数学分析
电压
计算机硬件
标识
DOI:10.1109/icassp.1982.1171644
摘要
We claim that speech analysis algorithms should be based on computational models of human audition, starting at the ears. While much is known about how hearing works, little of this knowledge has been applied in the speech analysis field. We propose models of the inner ear, or cochlea, which are expressed as time- and place-domain signal processing operations; i.e. the models are computational expressions of the important functions of the cochlea. The main parts of the models concern mechanical filtering effects and the mapping of mechanical vibrations into neural representation. Our model cleanly separates these effects into time-invariant linear filtering based on a simple cascade/parallel filterbank network of second-order sections, plus transduction and compression based on half-wave rectification with a nonlinear coupled automatic gain control network. Compared to other speech analysis techniques, this model does a much better job of preserving important detail in both time and frequency, which is important for robust sound analysis. We discuss the ways in which this model differs from more detailed cochlear models.
科研通智能强力驱动
Strongly Powered by AbleSci AI