Abstract
A novel frame-level online Video-to-Audio generation model, SoundReactor, uses a causal transformer and DINOv2 vision encoder to generate high-quality, synchronized audio with low latency from video frames.
Prevailing Video-to-Audio (V2A) generation models operate offline, assuming an entire video sequence or chunks of frames are available beforehand. This critically limits their use in interactive applications such as live content creation and emerging generative world models. To address this gap, we introduce the novel task of frame-level online V2A generation, where a model autoregressively generates audio from video without access to future video frames. Furthermore, we propose SoundReactor, which, to the best of our knowledge, is the first simple yet effective framework explicitly tailored for this task. Our design enforces end-to-end causality and targets low per-frame latency with audio-visual synchronization. Our model's backbone is a decoder-only causal transformer over continuous audio latents. For vision conditioning, it leverages grid (patch) features extracted from the smallest variant of the DINOv2 vision encoder, which are aggregated into a single token per frame to maintain end-to-end causality and efficiency. The model is trained through a diffusion pre-training followed by consistency fine-tuning to accelerate the diffusion head decoding. On a benchmark of diverse gameplay videos from AAA titles, our model successfully generates semantically and temporally aligned, high-quality full-band stereo audio, validated by both objective and human evaluations. Furthermore, our model achieves low per-frame waveform-level latency (26.3ms with the head NFE=1, 31.5ms with NFE=4) on 30FPS, 480p videos using a single H100. Demo samples are available at https://koichi-saito-sony.github.io/soundreactor/.
Community
SoundReactor: Frame-level Online Video-to-Audio Generation
Project page: https://koichi-saito-sony.github.io/soundreactor/
✅ Simple design architecture (vison encoder, full-band stereo audio VAE, multimodal transformer with diffusio head)
✅ Full-band stereo audio generation with audio-visual semantic and temporal syncronization
✅ Remarkably low frame-level latency (26.3ms with diffusion head NFE=1, 31.5ms with NFE=4) on 30FPS, 480p videos on H100.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- SALSA-V: Shortcut-Augmented Long-form Synchronized Audio from Videos (2025)
- HunyuanVideo-Foley: Multimodal Diffusion with Representation Alignment for High-Fidelity Foley Audio Generation (2025)
- Ovi: Twin Backbone Cross-Modal Fusion for Audio-Video Generation (2025)
- X-Streamer: Unified Human World Modeling with Audiovisual Interaction (2025)
- MIDAS: Multimodal Interactive Digital-humAn Synthesis via Real-time Autoregressive Video Generation (2025)
- Efficient Video-to-Audio Generation via Multiple Foundation Models Mapper (2025)
- Taming Text-to-Sounding Video Generation via Advanced Modality Condition and Interaction (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper