Parametric Sound Field Interpolation for Scene-based Navigable Immersive Audio


Olgun O., Deveci İ. E., Hachabiboğlu H.

IEEE Transactions on Audio, Speech and Language Processing, cilt.34, ss.1608-1620, 2026 (SCI-Expanded, Scopus) identifier

  • Yayın Türü: Makale / Tam Makale
  • Cilt numarası: 34
  • Basım Tarihi: 2026
  • Doi Numarası: 10.1109/taslpro.2026.3671646
  • Dergi Adı: IEEE Transactions on Audio, Speech and Language Processing
  • Derginin Tarandığı İndeksler: Science Citation Index Expanded (SCI-EXPANDED), Scopus
  • Sayfa Sayıları: ss.1608-1620
  • Anahtar Kelimeler: Immersive audio, microphone arrays, plane wave decomposition, six degrees-of-freedom, sound field, spherical harmonics
  • Orta Doğu Teknik Üniversitesi Adresli: Evet

Özet

Modern eXtended Reality (XR) systems are capable of tracking their users' positions within a navigable volume. This opens up the possibility of more realistic interactions with the content not only via head rotations, but also via locomotion. Such capability necessitates the design of audio systems that can generate a perceptually veridical rendering of the sound field at every position which the user can navigate to. While such interactivity can be provided by object-based audio (OBA) renderers such as in 3D computer games and VR applications, the new generation of volumetric visual content requires volumetric capture and rendering of the entire captured scenes, without explicitly identifying/extracting the audio objects therein. This article presents a parametric sound field interpolation method via multi-point sparse plane wave decomposition (SPWD) in the spherical harmonic domain. The proposed approach relies on higher-order Ambisonics (HOA) recordings of sound fields captured at multiple positions. The sound field is separated into direct and diffuse components that are interpolated parametrically and linearly, respectively. The proposed approach is compared with linear interpolation objectively and subjectively to demonstrate its utility.