VideoPoet
VideoPoet is a simple modeling method that can convert any autoregressive language model or large language model (LLM) into a high-quality video generator. It contains a few simple components. An autoregressive language model learns across video, image, audio, and text modalities to autoregressively predict the next video or audio token in the sequence. A mixture of multimodal generative learning objectives are introduced into the LLM training framework, including text-to-video, text-to-image, image-to-video, video frame continuation, video inpainting and outpainting, video stylization, and video-to-audio. Furthermore, such tasks can be composed together for additional zero-shot capabilities. This simple recipe shows that language models can synthesize and edit videos with a high degree of temporal consistency.
Learn more
Hunyuan-Vision-1.5
HunyuanVision is a cutting-edge vision-language model developed by Tencent’s Hunyuan team. It uses a mamba-transformer hybrid architecture to deliver strong performance and efficient inference in multimodal reasoning tasks. The version Hunyuan-Vision-1.5 is designed for “thinking on images,” meaning it not only understands vision+language content, but can perform deeper reasoning that involves manipulating or reflecting on image inputs, such as cropping, zooming, pointing, box drawing, or drawing on the image to acquire additional knowledge. It supports a variety of vision tasks (image + video recognition, OCR, diagram understanding), visual reasoning, and even 3D spatial comprehension, all in a unified multilingual framework. The model is built to work seamlessly across languages and tasks and is intended to be open sourced (including checkpoints, technical report, inference support) to encourage the community to experiment and adopt.
Learn more
Gen-2
Gen-2: The Next Step Forward for Generative AI.
A multi-modal AI system that can generate novel videos with text, images, or video clips. Realistically and consistently synthesize new videos. Either by applying the composition and style of an image or text prompt to the structure of a source video (Video to Video). Or, using nothing but words (Text to Video). It's like filming something new, without filming anything at all. Based on user studies, results from Gen-2 are preferred over existing methods for image-to-image and video-to-video translation.
Learn more
HunyuanVideo-Avatar
HunyuanVideo‑Avatar supports animating any input avatar images to high‑dynamic, emotion‑controllable videos using simple audio conditions. It is a multimodal diffusion transformer (MM‑DiT)‑based model capable of generating dynamic, emotion‑controllable, multi‑character dialogue videos. It accepts multi‑style avatar inputs, photorealistic, cartoon, 3D‑rendered, anthropomorphic, at arbitrary scales from portrait to full body. Provides a character image injection module that ensures strong character consistency while enabling dynamic motion; an Audio Emotion Module (AEM) that extracts emotional cues from a reference image to enable fine‑grained emotion control over generated video; and a Face‑Aware Audio Adapter (FAA) that isolates audio influence to specific face regions via latent‑level masking, supporting independent audio‑driven animation in multi‑character scenarios.
Learn more