Train Any Control Modality Fast With AVControl Framework
Based on research by Matan Ben-Yosef, Tavi Halperin, Naomi Ken Korem, Mohammad Salama, Harel Cain
Generating realistic video and audio content usually demands complex setups where every new control requirement, such as adjusting depth or camera movement, forces costly architectural changes to the underlying model. Researchers have now cracked this code with a lightweight framework that treats each control modality as a separate module, eliminating the need for expensive overhauls whenever a creator wants to add a new feature.
The system relies on extending an existing audio-visual foundation model by adding small, specialized adapters known as LoRAs. While previous attempts at controlling video generation failed when trying to apply image-based logic to structural elements like 3D depth or human pose, this parallel canvas approach solves the problem entirely without changing the core architecture beyond these lightweight additions. The result is a method that is vastly more efficient, requiring only a fraction of the computing power and data usually needed for training monolithic models.
On the VACE Benchmark, this new framework outperformed all previous baselines in tasks involving depth guidance, pose control, and image inpainting, while also delivering competitive results in camera trajectory tracking and audio-visual generation. By enabling independent training for diverse inputs such as edges, sparse motion, and video editing tools, it establishes a flexible standard for future generative media applications.
Source: AVControl: Efficient Framework for Training Audio-Visual Controls by Matan Ben-Yosef et al., https://arxiv.org/abs/2603.24793