ComfyUI Wan2.1 Fun InP Video Examples
This guide demonstrates how to use Wan2.1 Fun InP in ComfyUI to generate videos with first and last frame control
About Wan2.1-Fun-InP
Wan-Fun InP is an open-source video generation model released by Alibaba, part of the Wan2.1-Fun series, focusing on generating videos from images with first and last frame control.
Key features:
- First and last frame control: Supports inputting both first and last frame images to generate transitional video between them, enhancing video coherence and creative freedom. Compared to earlier community versions, Alibaba’s official model produces more stable and significantly higher quality results.
- Multi-resolution support: Supports generating videos at 512×512, 768×768, 1024×1024 and other resolutions to accommodate different scenario requirements.
Model versions:
- 1.3B Lightweight: Suitable for local deployment and quick inference with lower VRAM requirements
- 14B High-performance: Model size reaches 32GB+, offering better results but requiring higher VRAM
Below are the relevant model weights and code repositories:
- Wan2.1-Fun-1.3B-Input
- Wan2.1-Fun-14B-Input
- Code repository: VideoX-Fun
Currently, ComfyUI natively supports the Wan2.1 Fun InP model. Before starting this tutorial, please update your ComfyUI to ensure your version is after this commit.
Wan2.1 Fun InP Workflow
Download the image below and drag it into ComfyUI to load the workflow:
1. Workflow File Download
2. Manual Model Installation
If automatic model downloading is ineffective, please download the models manually and save them to the corresponding folders.
The following models can be found at Wan_2.1_ComfyUI_repackaged and Wan2.1-Fun
Diffusion models - choose 1.3B or 14B. The 14B version has a larger file size (32GB) and higher VRAM requirements:
- Wan2.1-Fun-1.3B-InP: Rename to
Wan2.1-Fun-1.3B-InP.safetensors
after downloading - Wan2.1-Fun-14B-InP: Rename to
Wan2.1-Fun-14B-InP.safetensors
after downloading
Text encoders - choose one of the following models (fp16 precision has a larger size and higher performance requirements):
VAE
CLIP Vision
File storage location:
3. Complete the Workflow Step by Step
- Ensure the
Load Diffusion Model
node has loadedWan2.1-Fun-1.3B-InP.safetensors
- Ensure the
Load CLIP
node has loadedumt5_xxl_fp8_e4m3fn_scaled.safetensors
- Ensure the
Load VAE
node has loadedwan_2.1_vae.safetensors
- Ensure the
Load CLIP Vision
node has loadedclip_vision_h.safetensors
- Upload the starting frame to the
Load Image
node (renamed toStart_image
) - Upload the ending frame to the second
Load Image
node - (Optional) Modify the prompt (both English and Chinese are supported)
- (Optional) Adjust the video size in
WanFunInpaintToVideo
, avoiding overly large dimensions - Click the
Run
button or use the shortcutCtrl(cmd) + Enter
to execute video generation
4. Workflow Notes
Please make sure to use the correct model, as Wan2.1-Fun-1.3B-InP.safetensors
and Wan2.1-Fun-1.3B-Control.safetensors
are stored in the same folder and have very similar names. Ensure you’re using the right model.
- When using Wan Fun InP, you may need to frequently modify prompts to ensure the accuracy of the corresponding scene transitions.
Other Wan2.1 Fun InP or video-related custom node packages
Was this page helpful?