Our Research
Building general-purpose multimodal
simulators of the world.
We believe models that use video as their main input/output modality, when supplemented by other modalities like text and audio, will form the next paradigm of computing.
We're advancing research in AI systems that can understand and simulate the world and its dynamics.
RNA Sessions
An ongoing series of talks about frontier research in
AI and art, hosted by Mova.Learn moreIntroducing Act-One
Foundations for Safe Generative Media
Pioneering New Interfaces in the Age of Generative Media
Introducing Gen-3 Alpha: A New Frontier for Video Generation
Introducing General World Models
More control, fidelity and expressibility
Mitigating stereotypical biases in text to image generative systems
Scale, Speed and Stepping Stones: The path to Gen-2
Gen-2: Generate novel videos with text, images or video clips
Gen-1: The next step forward for generative AI
Towards unified keyframe propagation models
High-Resolution Image Synthesis with Latent Diffusion Models
Soundify: Matching sound effects to video
Distributing Work: Adventures in queuing