SEE-2-SOUND: Zero-Shot Spatial Environment-to-Spatial Sound

1University of Toronto, 2Temerty Centre for Artificial Intelligence Research and Education in Medicine, 3Sunnybrook Research Institute,

SEE-2-SOUND generates spatial audio from images, animated images, and videos.



Our method works great on videos generated by OpenAI's Sora and can be used to generate spatial audio for video generation models.


Our method also works great for images from the web.

Abstract

Generating combined visual and auditory sensory experiences is critical for the consumption of immersive content. Recent advances in neural generative models have enabled the creation of high-resolution content across multiple modalities such as images, text, speech, and videos. Despite these successes, there remains a significant gap in the generation of high-quality spatial audio that complements generated visual content. Furthermore, current audio generation models excel in either generating natural audio or speech or music but fall short in integrating spatial audio cues necessary for immersive experiences. In this work, we introduce SEE-2-SOUND, a zero-shot approach that decomposes the task into (1) identifying visual regions of interest; (2) locating these elements in 3D space; (3) generating mono-audio for each; and (4) integrating them into spatial audio. Using our framework, we demonstrate compelling results for generating spatial audio for high-quality videos, images, and dynamic images from the internet, as well as media generated by learned approaches.

Mono Audio

A mono audio clip.

Spatial Audio

A spatial audio clip generated by SEE-2-SOUND.

How does SEE-2-SOUND work?

Overview of SEE-2-SOUND

SEE-2-SOUND consists of three main components: source estimation, audio generation, and surround sound spatial audio generation.

In the source estimation phase, the model identifies regions of interest in the input media and estimates their 3D positions on a viewing sphere. It also estimates the monocular depth map of the input image to further refine the spatial information.

Next, in the audio generation phase, the model generates mono audio clips for each identified region of interest, leveraging a pre-trained CoDi model. These audio clips are then combined with the spatial information to create a 4D representation for each region.

Finally, the model generates 5.1 surround sound spatial audio by placing sound sources in a virtual room and computing Room Impulse Responses (RIRs) for each source-microphone pair. Microphones are positioned according to the 5.1 channel configuration, ensuring compatibility with prevalent audio systems and enhancing the immersive quality of the audio output.

Generated Examples

Example Image 1
Example Image 2
Example Image 3

Example Image 4
Example Image 5
Example Image 6

Example Image 4
Example Image 5
Example Image 6

Example Image 4
Example Image 5
Example Image 6

Example Image 4
Example Image 5

BibTeX

@misc{dagli2024see2sound,
      title={SEE-2-SOUND: Zero-Shot Spatial Environment-to-Spatial Sound}, 
      author={Rishit Dagli and Shivesh Prakash and Robert Wu and Houman Khosravani},
      year={2024},
      eprint={2406.06612},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}