AES E-Library

Predicting Personalized Head Related Transfer Functions using Acoustic Scattering Neural Networks

The quality of spatial audio production has become critical to deliver a truly immersive sound experience in the recent past. Binaural audio is one of the most convenient formats to deliver accurate spatial audio over headphones. Personalized Head-related Transfer Functions (HRTFs) are an integral component of binaural audio that determines the quality of the spatial audio experience. In this paper, we describe a novel technique to predict personalized HRTFs based on 2D images or a video capture. The state-of-the-art 3D reconstruction techniques were developed for generic objects and thus do not perform well with complex structures such as an ear. We propose a novel 3D reconstruction algorithm that is modeled taking into account the geometry of the ear. The 3D output is then fed to a Acoustic Scattering Neural Network (ASNN) designed on the principles of Boundary Element Method (BEM) that outputs personalized HRTFs. The personalized HRTFs predicted are then compared both objectively with the measured HRTFs. We discuss the results, limitations, and the caveats necessary for an accurate modeling of personalized HRTFs.

 

Author (s):
Affiliation: (See document for exact affiliation information.)
AES Convention: Paper Number:
Publication Date:
Session subject:
Permalink: https://aes2.org/publications/elibrary-page/?id=21937


(1092KB)


Click to purchase paper as a non-member or login as an AES member. If your company or school subscribes to the E-Library then switch to the institutional version. If you are not an AES member Join the AES. If you need to check your member status, login to the Member Portal.

Type:
E-Libary location:
16938
Choose your country of residence from this list:










Skip to content