Google Research unveiled Lumiere, a text-to-video diffusion model that creates remarkably realistic videos from text or image prompts.
The still images generated by tools like Midjourney or DALL-E are incredible, but text-to-video (TTV) has understandably lagged behind and has been lots less impressive thus far.
TTV models like those from Pika Labs or Stable Video Diffusion have come a great distance within the last 12 months however the realism and continuity of motion are still just a little clunky.
Lumiere represents a giant jump in TTV as a consequence of a novel approach to generating video that’s spatially and temporally coherent. In other words, the goal is that the scenes in each frame stay visually consistent and the movements are smooth.
What can Lumiere do?
Lumiere has a spread of video generation functionality including the next:
- Text-to-video – Enter a text prompt and Lumiere generates a 5-second video clip made up of 80 frames at 16 frames per second.
- Image-to-video – Lumiere takes a picture because the prompt and turns it right into a video.
- Stylized generation – An image may be used as a mode reference. Lumiere uses a text prompt to generate a video within the kind of the reference image.
- Video stylization – Lumiere can edit a source video to match a stylistic text prompt.
- Cinemagraphs – Select a region in a still image and Lumiere will animate that a part of the image.
- Video inpainting – Lumiere can take a masked video scene and inpaint it to finish the video. It may edit source video by removing or replacing elements within the scene.
The video below shows a few of the impressive videos Lumiere can generate.
How does Lumiere do it?
Existing TTV models adopt a cascaded design where a base model generates a subset of keyframes after which they use a temporal super-resolution (TSR) model to generate data to fill the gaps between frames.
This approach is memory efficient but attempting to fill gaps between a sub-sampled set of keyframes leads to a video with temporal inconsistency, or glitchy motion. The low-resolution frames are then upscaled using a spatial super-resolution (SSR) model on non-overlapping windows.
Lumiere takes a unique approach. It uses a Space-Time U-Net (STUNet) architecture that learns to downsample the signal in each space and time and processes all of the frames without delay.
Because it isn’t just passing a subset of sample keyframes to a TSR, Lumiere achieves globally coherent motion. To obtain the high-resolution video, Lumiere applies an SSR model on overlapping windows and uses MultiDiffusion to mix the predictions right into a coherent result.
Google Research did a user study that showed users overwhelmingly preferred Lumiere videos from other TTV models.
User preferences of the standard of text-to-video quality, how well the video aligned with the text prompt, and image-to-video video quality. Source: Google Research
The final result may only be a 5-second clip, however the realism and coherent visuals and movement are higher than the rest currently available. Most other TTV solutions only generate 3-second clips for now.
Lumiere doesn’t handle scene transitions or multi-shot video scenes, but longer multi-scene functionality is nearly actually within the pipeline.
In the Lumiere research paper, Google noted that “there may be a risk of misuse for creating fake or harmful content with our technology.”
Hopefully, they discover a solution to effectively watermark their videos and avoid copyright issues in order that they can release Lumiere for us to place it through its paces.