--- tags: - text-to-video - diffusion - merged-model - video-generation - wan2.1 widget: - text: >- Prompt: Tight close-up of her smiling lips and sparkling eyes, catching golden hour sunlight. She wears a white sundress with floral prints and a wide-brimmed straw hat. Camera pulls back in a dolly motion, revealing her twirling under a cherry blossom tree. Petals flutter in the air, casting playful shadows. Soft lens flares enhance the euphoric, dreamlike vibe. *(πŸ”Έ Before vs After β€” Left: Wan2.1 | Right: Merged model Wan14BT2V_MasterModel)* output: url: videos/AnimateDiff_00001.mp4 base_model: - Wan-AI/Wan2.1-T2V-14B license: apache-2.0 --- # πŸŒ€ Wan2.1_14B_FusionX **High-Performance Merged Text-to-Video Model** Built on WAN 2.1 and fused with research-grade components for cinematic motion, detail, and speed β€” optimized for ComfyUI and rapid iteration in as few as 6 steps. Merged models for faster, richer motion & detail β€” high performance even at just 8 steps. > πŸ“Œ Important: Please read the full description. Small setting changes can drastically affect results. I've tested and documented better settings below β€” don't skip it! --- ## πŸ“‚ Workflows & Model Downloads - πŸ’‘ **ComfyUI workflows** can be found here: πŸ‘‰ [Workflow Collection (WIP)](https://civitai.com/models/1663553) - πŸ“¦ **Model files (T2V, I2V, Phantom, VACE)**: πŸ‘‰ [Main Hugging Face Repo](https://huggingface.co/vrgamedevgirl84/Wan14BT2VFusioniX/tree/main) ### 🧠 GGUF Variants: - πŸ–ΌοΈ [FusionX Image-to-Video (GGUF)](https://huggingface.co/QuantStack/Wan2.1_I2V_14B_FusionX-GGUF/tree/main) - πŸŽ₯ [FusionX Text-to-Video (GGUF)](https://huggingface.co/QuantStack/Wan2.1_T2V_14B_FusionX-GGUF/tree/main) - 🎞️ [FusionX T2V VACE (for native)](https://huggingface.co/QuantStack/Wan2.1_T2V_14B_FusionX_VACE-GGUF/tree/main) - πŸ‘» [FusionX Phantom](https://huggingface.co/QuantStack/Phantom_Wan_14B_FusionX-GGUF/tree/main) --- ## 🎬 Example Videos Want to see what FusionX can do? Check out these real outputs generated using the latest workflows and settings: - **Text-to-Video** πŸ‘‰ [Watch Examples](https://civitai.com/posts/17874424) - **Image-to-Video** πŸ‘‰ [Watch Examples](https://civitai.com/posts/18029174) - **Phantom Mode** πŸ‘‰ [Watch Examples](https://civitai.com/posts/17986906) - **VACE Integration** πŸ‘‰ [Watch Examples](https://civitai.com/posts/18080876) --- ## πŸš€ Overview A powerful text-to-video model built on top of **WAN 2.1 14B**, merged with several research-grade models to boost: - Motion quality - Scene consistency - Visual detail Comparable with closed-source solutions, but open and optimized for **ComfyUI** workflows. --- ## πŸ’‘ Inside the Fusion This model includes the following merged components: - **CausVid** – Causal motion modeling for better flow and dynamics - **AccVideo** – Better temporal alignment and speed boost - **MoviiGen1.1** – Cinematic smoothness and lighting - **MPS Reward LoRA** – Tuned for motion and detail - **Custom LoRAs** – For texture, clarity, and facial enhancements All merged models use permissive open licenses (Apache 2.0 / MIT). --- ## πŸ”§ Usage Details ### Text-to-Video - **CGF**: Must be set to `1` - **Shift**: - `1024x576`: Start at `1` - `1080x720`: Start at `2` - For realism β†’ lower values - For stylized β†’ test `3–9` - **Scheduler**: - Recommended: `uni_pc` - Alternative: `flowmatch_causvid` (better for some details) ### Image-to-Video - **CGF**: `1` - **Shift**: `2` works best in most cases - **Scheduler**: - Recommended: `dmp++_sde/beta` - To boost motion and reduce slow-mo effect: - Frame count: `121` - FPS: `24` --- ## πŸ›  Technical Notes - Works in as few as **6 steps** - Best quality at **8–10 steps** - Drop-in replacement for `Wan2.1-T2V-14B` - Up to **50% faster rendering**, especially with **SageAttn** - Works natively and with **Kaji Wan Wrapper** [Wrapper GitHub](https://github.com/kijai/ComfyUI-WanVideoWrapper) - Do **not** re-add merged LoRAs (CausVid, AccVideo, MPS) - Feel free to add **other LoRAs** for style/variation - Native WAN workflows also supported (slightly slower) --- ## πŸ§ͺ Performance Tips - RTX 5090 β†’ ~138 sec/video at 1024x576 / 81 frames - If VRAM is limited: - Enable block swapping - Start with `5` blocks and adjust as needed - Use **SageAttn** for ~30% speedup (wrapper only) - Do **not** use `teacache` - "Enhance a video" (tested): Adds vibrance (try values 2–4) - "SLG" not tested β€” feel free to explore --- ## 🧠 Prompt Help Want better cinematic prompts? Try the **WAN Cinematic Video Prompt Generator GPT** β€” it adds visual richness and makes a big difference in quality. [Download Here](https://chatgpt.com/g/g-67c3a6d6d19c81919b3247d2bfd01d0b-wan-cinematic-video-prompt-generator) --- ## πŸ“£ Join The Community We’re building a friendly space to chat, share outputs, and get help. - Motion LoRAs coming soon - Tips, updates, and support from other users πŸ‘‰ [Join the Discord](https://discord.com/invite/hxPmmXmRW3) --- ## βš–οΈ License Merged under permissive licenses: - Apache 2.0 / MIT - You **can** use, modify, and redistribute - You **must** retain original license info - Outputs are not necessarily licensed β€” do your due diligence This model is for **research, education, and personal use** only. Commercial use is your own responsibility. Please consult a legal advisor before monetizing outputs. --- ## πŸ™ Credits - WAN Team (base model) - aejion (AccVideo) - Tianwei Yin (CausVid) - ZuluVision (MoviiGen) - Alibaba PAI (MPS LoRA) - Kijai (ComfyUI Wrapper) And thanks to the open-source community! ---