Physics Collision

AI can now "photograph" a room's sound by treating audio like it's a 3D visual scene.

April 14, 2026

Original Paper

AudioGS: Spectrogram-Based Audio Gaussian Splatting for Sound Field Reconstruction

arXiv · 2604.08967

The Takeaway

It maps out 3D sound fields using only audio data, reconstructing how sound waves bounce off invisible walls without needing a single camera. This allows for high-fidelity binaural audio that is more accurate for virtual reality than methods using actual photos of the room.

From the abstract

Spatial audio is fundamental to immersive virtual experiences, yet synthesizing high-fidelity binaural audio from sparse observations remains a significant challenge. Existing methods typically rely on implicit neural representations conditioned on visual priors, which often struggle to capture fine-grained acoustic structures. Inspired by 3D Gaussian Splatting (3DGS), we introduce AudioGS, a novel visual-free framework that explicitly encodes the sound field as a set of Audio Gaussians based on