Wan2.1 I2V GGUF w/ Optimization Nodes + Causvid/FusionX/LightX Workflow

Details

Model description

No more OOM errors here with 16gbvram. Use this workflow as a reference to incorporate these optimization nodes to your own workflow or use it yourself if you really want to.

I modified definitelynotadog's version 1 workflow to optimize for 16gbvram (I'm sure lower can work too) so shoutout to him for making the main thing - /model/1622023?modelVersionId=1835720

I added SageAttention, BlockSwap and TeaCache nodes. Around 3 minutes using the 480p GGUF model. Around 6 minutes using 720p GGUF. Feel free to yoink the optimization nodes I added and add it to his V2 workflow or to your own workflow even.

Firstly, Download a GGUF version of Wan2.1 model of your choice for better performance. higher the Q number = higher quality. I personally went with Q3 bc it is faster than Q8 and the difference in quality was very minimal from what I could tell.

LoRa Accelerators:

The Causvid LORA is replaceable with FusionX or LightX lora because it's built-in apparently so choose one or the other.

SelfForcing - /model/1713337/wan-self-forcing-rank-16-accelerator

FusionX - /model/1678575/wan21fusionx-the-lora

CausVid - /model/1585622/self-forcing-causvid-accvid-lora-massive-speed-up-for-wan21-made-by-kijai

Models:

Here are I2V 720p gguf models - https://huggingface.co/city96/Wan2.1-I2V-14B-720P-gguf/tree/main

I2V 480p gguf models model - https://huggingface.co/city96/Wan2.1-I2V-14B-480P-gguf/tree/main

Make sure to also download the missing VAE, clip vision, and clip

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/vae/wan_2.1_vae.safetensors

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/clip_vision/clip_vision_h.safetensors

https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors

Make sure you have Custom Manager to install Custom Nodes and dependencies but this specific one you might have to install manually: https://github.com/city96/ComfyUI-GGUF

Settings:

Make sure you're running 4-10 steps when using fusionx lora or causvid lora, stick to 16fps, 81 length max. 7 steps is the sweet spot for me, anything below I get a lot of artifacts/glitches. 848 for 480p and 1280 for 720p

OPTIMIZATIONS:

Sage gives ~25% more speed. If you don't have SageAttention installed on your system. Here's a guide for sage on windows, you might bork your comfyui if done incorrectly: https://www.reddit.com/r/StableDiffusion/comments/1h7hunp/how_to_run_hunyuanvideo_on_a_single_24gb_vram_card/

BlockSwap Node, it offloads VRAM to system RAM (NO MORE OOM ERRORS)Set to 40 for 14B model, 30 for 1.3B model. Look at vram usage while generating to find the sweetspot; the higher the blocks, the less vram usage. Keep in mind resolution effects VRAM by a lot.

TeaCache Node gives 2X speed but at the cost of QUALITY. If video is too fuzzy or you are missing limbs, set the thres lower to 0.140 or feel free to bypass/delete this node if you don't like the results. I'm still tweaking it to find the best settings for it.

I'm still new to these nodes so feel free provide any useful info you may know to help others.

Images made by this model

No Images Found.