Simulation of Human Ear Recognition Sound Direction Based on Convolutional Neural Network

In recent years, more and more people are applying Convolutional Neural Networks to the study of sound signals. The main reason is the translational invariance of convolution in time and space. Thereby the diversity of the sound signal can be overcome. However, in terms of sound direction recognitio...

Descripción completa

Guardado en:
Detalles Bibliográficos
Autores principales: Wang Zhuhe, Li Nan, Wu Tao, Zhang Haoxuan, Feng Tao
Formato: article
Lenguaje:EN
Publicado: De Gruyter 2020
Materias:
gcc
Q
Acceso en línea:https://doaj.org/article/f31548cf575547efbec9a13fa73d2164
Etiquetas: Agregar Etiqueta
Sin Etiquetas, Sea el primero en etiquetar este registro!
Descripción
Sumario:In recent years, more and more people are applying Convolutional Neural Networks to the study of sound signals. The main reason is the translational invariance of convolution in time and space. Thereby the diversity of the sound signal can be overcome. However, in terms of sound direction recognition, there are also problems such as a microphone matrix being too large, and feature selection. This paper proposes a sound direction recognition using a simulated human head with microphones at both ears. Theoretically, the two microphones cannot distinguish the front and rear directions. However, we use the original data of the two channels as the input of the convolutional neural network, and the resolution effect can reach more than 0.9. For comparison, we also chose the delay feature (GCC) for sound direction recognition. Finally, we also conducted experiments that used probability distributions to identify more directions.