With Hunyuan text-to-video technology, you can now generate videos in just 6 simple steps. Yes, you read that right—only six!
Getting Started: Setting Up the Workflow
Here’s how you can create your own Hunyuan video in just a few steps:
- Download the Workflow
Visit ComfyUI Blog and download the pre-built workflow tailored for Hunyuan video creation. - Load ComfyUI
Launch ComfyUI on your system and load the downloaded workflow. For reference, this tutorial uses a system with 12GB of VRAM. - Gather Required Files
You’ll need the following files:
(Save this in the VAE folder of ComfyUI.)pytorch_model.pt
(Place this in the Diffusion Model folder.)mp_rank_00_model_states.pt
andClip I of Flux
(Save these in the Text Encoder folder.)Llava Llama3 FP8
- Load the Files into ComfyUI
- In the VAE loader, select
.pytorch_model.pt
- In the Diffusion Model loader, select
. If you encounter memory issues, switch the Weight Dtype tomp_rank_00_model_states.pt
.FP8
- For Dual Clip Load, select
andClip I of Flux
, and set the type to Hunyuan Video.Llava Llama3 FP8
- In the VAE loader, select
- Configure the Workflow
- Add your text prompt in the Clip Text Encoder and set Flux Guidance to
and Model Sampling to10
.7
- In the Empty Latent section, set the resolution to
for standard output or848x480
for HD.720x1224
- In the Basic Scheduler, select the Simple Scheduler and set the steps to
for quick results or up to6
for enhanced quality.20
- Add your text prompt in the Clip Text Encoder and set Flux Guidance to
Download Files
- VAE: pytorch_model.pt
- Diffusion Model: mp_rank_00_model_states.pt
- Text Encoder: llava_llama3_fp8_scaled.safetensors