After the successful launch of a text-to-image model, a controversial launch of a text-to-music model, and a largely unnoticed launch of a text generation model, Stability AI has just announced the release of Stable Video Diffusion, a text-to-video tool that aims to carve a chunk out of the nascent generative video space.
“Stable Video Diffusion [is] a latent video diffusion model for high-resolution state-of-the-art text-to-video and image-to-video generation,” Stability AI explains in the model’s research paper, and adds in the official announcement, “Spanning across modalities including image, language, audio, 3D, and code, our portfolio is a testament to Stability AI’s dedication to amplifying human intelligence.”
This adaptability, coupled with open-source technology, paves the way for numerous applications in advertising, education, and entertainment. Stable Video Diffusion, which is now available in a research preview, is able to ”outperform image-based methods at a fraction of their compute budget,” according to researchers.
Go to Source to See Full Article
Author: Jose Antonio Lanz
Tip BTC Newswire with Cryptocurrency