Parametric Sound Field Interpolation for Scene-based Navigable Immersive Audio


Olgun O., Deveci İ. E., Hachabiboğlu H.

IEEE Transactions on Audio, Speech and Language Processing, vol.34, pp.1608-1620, 2026 (SCI-Expanded, Scopus) identifier

  • Publication Type: Article / Article
  • Volume: 34
  • Publication Date: 2026
  • Doi Number: 10.1109/taslpro.2026.3671646
  • Journal Name: IEEE Transactions on Audio, Speech and Language Processing
  • Journal Indexes: Science Citation Index Expanded (SCI-EXPANDED), Scopus
  • Page Numbers: pp.1608-1620
  • Keywords: Immersive audio, microphone arrays, plane wave decomposition, six degrees-of-freedom, sound field, spherical harmonics
  • Middle East Technical University Affiliated: Yes

Abstract

Modern eXtended Reality (XR) systems are capable of tracking their users' positions within a navigable volume. This opens up the possibility of more realistic interactions with the content not only via head rotations, but also via locomotion. Such capability necessitates the design of audio systems that can generate a perceptually veridical rendering of the sound field at every position which the user can navigate to. While such interactivity can be provided by object-based audio (OBA) renderers such as in 3D computer games and VR applications, the new generation of volumetric visual content requires volumetric capture and rendering of the entire captured scenes, without explicitly identifying/extracting the audio objects therein. This article presents a parametric sound field interpolation method via multi-point sparse plane wave decomposition (SPWD) in the spherical harmonic domain. The proposed approach relies on higher-order Ambisonics (HOA) recordings of sound fields captured at multiple positions. The sound field is separated into direct and diffuse components that are interpolated parametrically and linearly, respectively. The proposed approach is compared with linear interpolation objectively and subjectively to demonstrate its utility.