QWEN-Anime Official Workflow
ファイルをダウンロード
このバージョンについて
# 🎨 Official ComfyUI Workflow for QWEN-Anime-Beta2 (GGUF)
This is the GGUF-optimized workflow designed specifically for QWEN-Anime-Beta2 GGUF versions (Q4/Q6/Q8/F16). This workflow is tailored for users who want to use the GGUF quantized models for better memory efficiency or CPU inference.
## ⚠️ CRITICAL REQUIREMENT
YOU MUST INSTALL ComfyUI-GGUF FIRST!
This workflow will NOT work without the ComfyUI-GGUF custom node:
- GitHub: https://github.com/city96/ComfyUI-GGUF
- Installation: Via ComfyUI Manager or manual git clone
Without this custom node, GGUF models cannot be loaded!
---
## ✨ Features
### GGUF-Specific:
- ✅ Uses UnetLoaderGGUF node for GGUF model loading
- ✅ Correct folder structure unet/ instead of diffusion_models/)
- ✅ Optimized for GGUF quantized models (Q4/Q6/Q8/F16)
- ✅ Great for limited VRAM or CPU inference setups
### General Features:
- 📋 Step-by-step workflow with clear color-coded groups
- 📦 Integrated GGUF model download links and folder structure
- ⚡ Support for Lightning LoRAs (4-step and 8-step) for fast generation
- 🔼 Optional Hires Fix upscaler with side-by-side comparison
- 📐 Preset resolutions from 512×512 to 2K (2048×1152)
- 💾 Works with limited VRAM (GGUF Q4 = 10GB file!)
---
## 📦 What's Included
- Complete GGUF model loading setup with UnetLoaderGGUF node
- Prompt templates with tips for better results
- Resolution presets (SDXLResolutionPresets) for quick size selection
- Base generation + optional upscaling workflow
- Image comparison tool (rgthree) to compare before/after upscaling
- All necessary nodes with helpful titles and descriptions
- GGUF-specific instructions built directly into the workflow
---
## ⚙️ Requirements
### Essential:
1. ComfyUI (latest version recommended)
2. ComfyUI-GGUF custom node ⚠️ MANDATORY
- https://github.com/city96/ComfyUI-GGUF
### Additional Custom Nodes:
3. rgthree-comfy (for image comparison and group bypasser)
- https://github.com/rgthree/rgthree-comfy
4. comfyui-copilot (for resolution presets)
- https://github.com/hylarucoder/comfyui-copilot
### Files to Download:
- QWEN-Anime-Beta2 GGUF (Q4/Q6/Q8/F16) - Choose one
- Text Encoder (qwen_2.5_vl_7b_fp8_scaled.safetensors)
- VAE (qwen_image_vae.safetensors)
- Optional: Lightning LoRAs for speed boost
All download links are included directly in the workflow!
---
## 🚀 Quick Start
### Step 1: Install ComfyUI-GGUF
```
CRITICAL: This MUST be done first!
https://github.com/city96/ComfyUI-GGUF
```
Install via ComfyUI Manager or:
```bash
cd ComfyUI/custom_nodes/
git clone https://github.com/city96/ComfyUI-GGUF
```
Restart ComfyUI after installation.
### Step 2: Download Files
1. Load the workflow in ComfyUI
2. Read the Model Links note in the workflow
3. Download your chosen GGUF version (Q8 recommended for quality)
4. Download Text Encoder and VAE from provided links
5. Optional: Download Lightning LoRA
### Step 3: Place Files Correctly
⚠️ IMPORTANT: GGUF files go in a DIFFERENT folder!
```
📂 ComfyUI/models/
├── 📂 unet/ ⚠️ GGUF goes HERE!
│ └── qwen-anime-beta2-Q8.gguf
├── 📂 loras/Qwen/
│ └── Qwen-Image-Lightning-8steps-V1.0.safetensors
├── 📂 vae/QWEN/
│ └── qwen_image_vae.safetensors
└── 📂 text_encoders/QWEN/
└── qwen_2.5_vl_7b_fp8_scaled.safetensors
```
Common Mistake: Don't put GGUF in diffusion_models/ folder - it won't work!
### Step 4: Generate
1. Write your positive prompt
2. Write your negative prompt
3. Select resolution preset
4. Click Queue Prompt
5. Enjoy your anime art!
---
## 🎯 Which GGUF Version to Choose?
### Q8 (20.23 GB) - RECOMMENDED ⭐
- Best quality among quantized versions
- Minimal quality loss vs full precision
- Good balance of size and quality
- Recommended for most users
### Q6 (15.63 GB) - Balanced
- Good quality with smaller file size
- Slight quality reduction vs Q8
- Good for storage-limited systems
### Q4 (10.72 GB) - Smallest
- Smallest file size
- Fastest loading
- Noticeable but acceptable quality loss
- Great for very limited storage/VRAM
### F16 (38.07 GB) - Full Precision
- Maximum quality
- Same as original full precision
- Very large file size
- Only if you need absolute best quality
---
## 💡 GGUF Benefits
### Why Use GGUF Instead of Regular Safetensors?
✅ Smaller File Sizes - Q4 is only 10GB vs 38GB!
✅ Lower VRAM Usage - Works on lower-end GPUs
✅ CPU Inference - Can run on CPU if needed
✅ Faster Loading - Smaller files load quicker
✅ Good Quality - Q8 has minimal quality loss
### When to Use GGUF:
- Limited disk space
- Limited VRAM (below 8GB)
- Want faster model loading
- Need CPU inference support
- Storage/bandwidth limited
### When to Use Regular Safetensors Instead:
- Have plenty of VRAM (12GB+)
- Want absolute maximum quality
- Storage space not a concern
- Don't want to install extra custom nodes
---
## 📖 Workflow Guide
The workflow is organized into 5 color-coded groups:
### 📦 Step 1: Load Models (Blue)
- Uses UnetLoaderGGUF node (GGUF-specific!)
- Load Text Encoder with CLIPLoader
- Load VAE with VAELoader
- Optional: Add Lightning LoRA for speed
### ✍️ Step 2: Write Prompts (Green)
- Positive Prompt: Describe what you want
- Negative Prompt: Describe what to avoid
- Tips and templates included in workflow
### 📐 Step 3: Choose Size (Cyan)
- Use Resolution Presets for quick selection
- Tested resolutions: 512×512 to 2K
- All major aspect ratios supported
### 🎨 Step 4: Generate (Purple)
- Base KSampler creates initial image
- Automatic VAE decoding
- Preview and save base version
### 🔼 Step 5: Upscale (Orange - Optional)
- Toggle ON/OFF with Fast Groups Bypasser
- Upscales by 1.5x for better detail
- Second pass refines upscaled image
- Side-by-side comparison with original
---
## ⚙️ Recommended Settings
### Fast Mode (with 8-step LoRA):
- Steps: 8
- CFG: 1.0
- Resolution: 512×512 or 1024×1024
- Upscaler: OFF
- Time: ~30-80 seconds
### Quality Mode (no LoRA):
- Steps: 24-32
- CFG: 3.6-4.0
- Resolution: 1024×1024 or higher
- Upscaler: ON
- Time: ~180-400 seconds
### Ultra Quality (with upscaler):
- Steps: 32 base + 16 upscale
- CFG: 3.6-4.2
- Resolution: Start at 1024×1024, upscale to 1536×1536
- Upscaler: ON with denoise 0.55-0.65
- Time: ~300-600 seconds
---
## 🎨 Prompting Tips
### Good Prompt Structure:
```
anime, [quality tags], [subject], [details], [style], [lighting]
```
### Example Positive Prompt:
```
anime, masterpiece, best quality, 1girl, nekomimi, fluffy cat ears,
long silver hair, golden eyes, playful smirk, witch hat,
casting magic, cozy library, warm lighting, detailed, vibrant colors
```
### Example Negative Prompt:
```
low quality, blurry, distorted, bad anatomy, bad hands,
text, watermark, signature, ugly, mutation, extra limbs
```
---
## 🐛 Troubleshooting
### "UnetLoaderGGUF node not found"
Solution: Install ComfyUI-GGUF custom node!
https://github.com/city96/ComfyUI-GGUF
### "Cannot find GGUF model"
Solution: Make sure file is in models/unet/ folder, NOT diffusion_models/
### "Out of memory" error
Solution:
- Use smaller GGUF version (Q6 or Q4)
- Lower resolution (try 512×512)
- Disable upscaler
- Close other applications
### Images look worse than expected
Solution:
- Try higher GGUF version (Q8 instead of Q4)
- Increase steps (20-32)
- Adjust CFG scale (try 3.0-4.0)
- Add quality tags to prompt
### Workflow loads but generates errors
Solution:
- Verify all custom nodes are installed
- Check all files are in correct folders
- Restart ComfyUI after installing nodes
- Check ComfyUI console for specific errors
---
## 💬 What Makes This Different?
### vs Regular Safetensors Workflow:
- ✅ Uses UnetLoaderGGUF instead of UNETLoader
- ✅ Files go in unet/ folder instead of diffusion_models/
- ✅ Requires ComfyUI-GGUF custom node
- ✅ Smaller file sizes (10-20GB vs 38GB)
- ✅ Better for limited VRAM setups
### vs Basic GGUF Setup:
- ✅ Complete workflow with all steps
- ✅ Built-in upscaler and comparison
- ✅ Resolution presets included
- ✅ All download links provided
- ✅ Instructions built into workflow
---
## 📋 Checklist Before Starting
Before you start generating, make sure:
- [ ] ComfyUI-GGUF custom node is installed
- [ ] rgthree-comfy custom node is installed
- [ ] comfyui-copilot custom node is installed
- [ ] GGUF model file is in models/unet/ folder
- [ ] Text Encoder is in models/text_encoders/QWEN/
- [ ] VAE is in models/vae/QWEN/
- [ ] (Optional) Lightning LoRA in models/loras/Qwen/
- [ ] ComfyUI has been restarted after installing nodes
- [ ] Workflow loads without errors
If all boxes are checked, you're ready to generate! 🎨
---
## 🎯 Perfect For
- Beginners - Step-by-step color-coded workflow
- Limited Storage - Smaller file sizes (Q4 = 10GB!)
- Limited VRAM - Works on lower-end GPUs
- CPU Users - GGUF supports CPU inference
- Bandwidth Limited - Smaller downloads
- Advanced Users - Full control over all settings
---
## 📞 Support
### If you need help:
1. Check the troubleshooting section above
2. Verify all custom nodes are installed
3. Check ComfyUI console for error messages
4. Make sure files are in correct folders
5. Visit the CivitAI model page for updates
### Common Questions:
Q: Do I need the regular safetensors version too?
A: No! This workflow is complete and self-contained for GGUF.
Q: Can I use regular safetensors with this workflow?
A: No, this workflow is GGUF-specific. Use the regular workflow for safetensors.
Q: Which GGUF version is best?
A: Q8 for best quality, Q4 for smallest size. See "Which GGUF Version" section above.
Q: Why can't I find my GGUF file in ComfyUI?
A: Make sure it's in models/unet/ folder, NOT diffusion_models/!
Q: Do I still need VAE and Text Encoder?
A: Yes! GGUF is just the main model. You still need VAE and Text Encoder separately.
---
## 🙏 Credits
Workflow created for: QWEN-Anime-Beta2 GGUF versions
Base Model: qwen-image-edit 2509
Custom Dataset: Hand-created by model author
ComfyUI-GGUF: city96 (https://github.com/city96/ComfyUI-GGUF)
rgthree-comfy: rgthree
comfyui-copilot: hylarucoder
---
## 📝 Version Info
Workflow Version: 1.0
Compatible With: QWEN-Anime-Beta2 GGUF (all versions)
Created: November 2025
Last Updated: November 2025
---
Download the workflow, install ComfyUI-GGUF, and start creating amazing anime art! 🎨
Perfect for users who want smaller file sizes and better memory efficiency while maintaining great quality!
---
## 🔗 Useful Links
- QWEN-Anime-Beta2 Model: (CivitAI page)
- ComfyUI-GGUF: https://github.com/city96/ComfyUI-GGUF
- rgthree-comfy: https://github.com/rgthree/rgthree-comfy
- comfyui-copilot: https://github.com/hylarucoder/comfyui-copilot
- Text Encoder: https://huggingface.co/Comfy-Org/Qwen-Image_ComfyUI/resolve/main/split_files/text_encoders/qwen_2.5_vl_7b_fp8_scaled.safetensors
- Lightning LoRAs: /model/2048940 & /model/2046883
モデル説明
🎨 QWEN-Anime-Beta1 用 公式 ComfyUI ワークフロー
これは QWEN-Anime-Beta1 チェックポイント用に最適化されたワークフローです。高品質なアニメ画像を素早く生成するために必要なすべてが含まれています。
✨ 特徴:
- 明確な色分けされたグループで構成されたステップバイステップワークフロー
- モデルダウンロードリンクとフォルダ構成ガイドを統合
- Lightning LoRAs(4ステップおよび8ステップ)対応で高速生成可能
- オプションで Hires Fix アップスケーラーを搭載し、横並び比較可能
- 512×512 から 2K(2048×1152)までのプリセット解像度
- 8GB VRAM でも動作
📦 含まれる内容:
- 完全なモデル読み込み設定
- ヒント付きプロンプトテンプレート
- 解像度プリセット(SDXLResolutionPresets)
- ベース生成 + オプションのアップスケーリング
- 画像比較ツール(rgthree)
- すべての必要なノードに役立つタイトルと説明を付属
⚙️ 必要条件:
- ComfyUI
- カスタムノード:rgthree-comfy、comfyui-copilot(ワークフロー内にインストールリンクを含む)
- QWEN-Anime-Beta1 チェックポイント + VAE + Text Encoder(ワークフロー内にダウンロードリンクを含む)
- オプション:速度向上用 Lightning LoRAs
🚀 クイックスタート:
ComfyUI でワークフローを読み込む
モデルリンクの注意書きに従って必要なファイルをダウンロード
ファイルを正しいフォルダに配置(ワークフローに構成を表示)
プロンプトを入力
解像度を選択
生成!
初心者から上級者まで幅広く利用可能。すべてのモデルリンク、インストール手順、ヒントがワークフローに組み込まれています。


