TR2024-169

Spatially-Aware Losses for Enhanced Neural Acoustic Fields


    •  Ick, Christopher, Wichern, Gordon, Masuyama, Yoshiki, Germain, François G, Le Roux, Jonathan, "Spatially-Aware Losses for Enhanced Neural Acoustic Fields", Tech. Rep. TR2024-169, Mitsubishi Electric Research Laboratories, Cambridge, MA, December 2024.
      BibTeX TR2024-169 PDF
      • @techreport{MERL_TR2024-169,
      • author = {Ick, Christopher; Wichern, Gordon; Masuyama, Yoshiki; Germain, François G; Le Roux, Jonathan},
      • title = {Spatially-Aware Losses for Enhanced Neural Acoustic Fields},
      • institution = {MERL - Mitsubishi Electric Research Laboratories},
      • address = {Cambridge, MA 02139},
      • number = {TR2024-169},
      • month = dec,
      • year = 2024,
      • url = {https://www.merl.com/publications/TR2024-169/}
      • }
  • MERL Contacts:
  • Research Areas:

    Artificial Intelligence, Machine Learning, Speech & Audio

Abstract:

For immersive audio experiences, it is essential that sound propagation is accurately modeled from a source to a listener through space. For human listeners, binaural audio characterizes the acoustic environment, as well as the spatial aspects of an acoustic scene. Recent advancements in neural acoustic fields have demonstrated spatially continuous models that are able to accurately reconstruct binaural impulse responses for a given source/listener pair. Despite this, these approaches have not explicitly examined or evaluated the quality of these reconstructions in terms of the inter-aural cues that define spatialization for human listeners. In this work, we propose extending neural acoustic field-based methods with spatially-aware metrics for training and evaluation to better capture spatial acoustic cues. We develop a dataset based on the existing SoundSpaces dataset to better model these features, and we demonstrate performance improvements by utilizing spatially-aware losses.