VOID AI removes objects from videos along with all physical interactions they induce —
shadows, reflections, and even falling objects. Powered by Netflix's open-source model.
Open-source model by Netflix Research
VOID (Video Object and Interaction Deletion) is a first-of-its-kind AI model that understands physical interactions between objects in video scenes.
Goes beyond visual artifacts — removes physical interactions like objects falling when a person is removed from the scene.
Uses a 4-value mask system to precisely identify primary objects, overlap regions, affected areas, and background.
Two-pass refinement with warped-noise ensures smooth, flickering-free results across video frames.
Built on CogVideoX-Fun-V1.5-5b architecture with 5 billion parameters. Fully open-source and research-ready.
Traditional object removal tools only handle visual effects. VOID understands physics.
Everything that makes VOID the most advanced video object removal model available.
Understands cause-and-effect in video scenes — removes objects and their downstream physical effects.
Precise control over what to remove, what's affected, and what to keep using a novel masking approach.
Base inpainting pass followed by warped-noise temporal consistency refinement for smooth results.
Built on CogVideoX-Fun-V1.5-5b-InP architecture for high-quality video generation and inpainting.
Process long video sequences at 384x672 resolution with consistent quality throughout.
Model weights, code, and training methodology published by Netflix Research. Free to use and extend.
Common questions about VOID AI and the VOID model.
Try the VOID model — open source, interaction-aware, state-of-the-art.