This repository is the official implementation of Kandinsky Video model
Paper | Project | | Telegram-bot | Habr post |
Kandinsky Video is a text-to-video generation model, which is based on the FusionFrames architecture, consisting of two main stages: keyframe generation and interpolation. Our approach for temporal conditioning allows us to generate videos with high-quality appearance, smoothness and dynamics.
The encoded text prompt enters the U-Net keyframe generation model with temporal layers or blocks, and then the sampled latent keyframes are sent to the latent interpolation model in such a way as to predict three interpolation frames between two keyframes. A temporal MoVQ-GAN decoder is used to get the final video result.
Architecture details
- Text encoder (Flan-UL2) - 8.6B
- Latent Diffusion U-Net3D - 4.0B
- MoVQ encoder/decoder - 256M
Check our jupyter notebooks with examples in ./examples
folder
from video_kandinsky3 import get_T2V_pipeline
t2v_pipe = get_T2V_pipeline('cuda', fp16=True)
pfps = 'medium' # ['low', 'medium', 'high']
video = t2v_pipe(
'a red car is drifting on the mountain road, close view, fast movement',
width=640, height=384, fps=fps
)
- Vladimir Arkhipkin: Github, Google Scholar
- Zein Shaheen: Github, Google Scholar
- Viacheslav Vasilev: Github, Google Scholar
- Igor Pavlov: Github
- Elizaveta Dakhova: Github
- Anastasia Lysenko: Github
- Sergey Markov
- Denis Dimitrov: Github, Google Scholar
- Andrey Kuznetsov: Github, Google Scholar
If you use our work in your research, please cite our publication:
@article{arkhipkin2023fusionframes,
title = {FusionFrames: Efficient Architectural Aspects for Text-to-Video Generation Pipeline},
author = {Arkhipkin, Vladimir and Shaheen, Zein and Vasilev, Viacheslav and Dakhova, Elizaveta and Kuznetsov, Andrey and Dimitrov, Denis},
journal = {arXiv preprint arXiv:2311.13073},
year = {2023},
}