Alibaba Cloud, the digital technology and intelligence backbone of Alibaba Group, recently made its AI models for video generation freely available as part of its latest efforts to contribute to the open-source community.
It is open-sourcing four models of its 14-billion(B)-parameter and 1.3-billion(B)-parameter versions of Wan2.1 series, the latest iteration of its video foundation model Tongyi Wanxiang (Wan).
The four models, including T2V-14B, T2V-1.3B, I2V-14B-720P, and I2V-14B-480P, are designed to generate high-quality images and videos from text and image inputs. They are available for download on Alibaba Cloud’s AI model community, ModelScope, and the collaborative AI platform Hugging Face, accessible to academics, researchers, and commercial institutions worldwide. Within a week of their launch, the combined downloads of the four Wan2.1 open-source models on ModelScope and Hugging Face exceeded 1 million.
Unveiled earlier this year, the Wan2.1 series is the first video generation model to support text effects in both Chinese and English. It excels at generating realistic visuals by accurately handling complex movements, enhancing pixel quality, adhering to physical principles, and optimizing the precision of instruction execution. Its precision in following instructions has propelled Wan2.1 to the top of the VBench leaderboard, a comprehensive benchmark suite for video generative models.
According to VBench, the Wan2.1 series, with an overall score of 86.22%, leads in key dimensions such as dynamic degree, spatial relationships, color, and multi-object interactions.
Training video foundation models requires immense computing resources and vast amounts of high-quality training data. Open access helps lower the barrier for more businesses to leverage AI, enabling them to create high-quality visual content tailored to their needs in a cost-effective way.
The T2V-14B model is better suited for creating high-quality visuals with substantial motion dynamics, while the T2V-1.3B model strikes a balance between generation quality and computational power, making it ideal for a broad range of developers conducting secondary development and academic research. For example, the T2V-1.3B model allows users with standard personal laptops to generate a 5-second-long video at 480p resolution in as little as around 4 minutes.
In addition to supporting text-to-video generation, the I2V-14B-720P and I2V-14B-480P models also offer image-to-video capabilities. Users simply need to input a single image along with a brief text description to generate dynamic video content. The platform supports normal-sized image inputs of any dimensions.
Alibaba Cloud was one of the first major global tech companies to make its self-developed large-scale AI model open-source, releasing its first open-model Qwen (Qwen-7B) in August 2023. Qwen open-models have consistently topped the Hugging Face Open LLM Leaderboards, with performances matching that of leading global AI models across various benchmarks.
As of now, more than 100,000 derivative models based on the Qwen family of models have been developed on Hugging Face, making it one of the largest AI model families worldwide.