Skip to content

vipshop/cache-dit

Repository files navigation

A PyTorch-native and Flexible Inference Engine with
Hybrid Cache Acceleration and Parallelism for 🤗DiTs
Featured|HelloGitHub

Baseline SCM S S* SCM F D* SCM U D* +TS +compile +FP8*
24.85s 15.4s 11.4s 8.2s 8.2s 🎉7.1s 🎉4.5s

Scheme: DBCache + SCM(steps_computation_mask) + TS(TaylorSeer) + FP8*, L20x1, S*: static cache,
D*: dynamic cache, S: Slow, F: Fast, U: Ultra Fast, TS: TaylorSeer, FP8*: FP8 DQ + Sage, FLUX.1-Dev

SGLang Diffusion x Cache-DiT News vLLM Omni x Cache-DiT News

🔥Hightlight

We are excited to announce that the 🎉v1.1.0 version of cache-dit has finally been released! It brings 🔥Context Parallelism and 🔥Tensor Parallelism to cache-dit, thus making it a PyTorch-native and Flexible Inference Engine for 🤗DiTs. Key features: Unified Cache APIs, Forward Pattern Matching, Block Adapter, DBCache, DBPrune, Cache CFG, TaylorSeer, SCM, Context Parallelism (w/ UAA), Tensor Parallelism and 🎉SOTA performance.

pip3 install -U cache-dit # Also, pip3 install git+https://github.com/huggingface/diffusers.git (latest)

You can install the stable release of cache-dit from PyPI, or the latest development version from GitHub. Then try ♥️ Cache Acceleration with just one line of code ~ ♥️

>>> import cache_dit
>>> from diffusers import DiffusionPipeline
>>> pipe = DiffusionPipeline.from_pretrained("Qwen/Qwen-Image") # Can be any diffusion pipeline
>>> cache_dit.enable_cache(pipe) # One-line code with default cache options.
>>> output = pipe(...) # Just call the pipe as normal.
>>> stats = cache_dit.summary(pipe) # Then, get the summary of cache acceleration stats.
>>> cache_dit.disable_cache(pipe) # Disable cache and run original pipe.

📚Core Features

🔥Supported DiTs

Tip

✅: supported; ✖️: not supported now; Q: nunchaku; C-P: Context Parallelism; T-P: Tensor Parallelism; TE-P: Text Encoder Parallelism; CN-P: ControlNet Parallelism; VE-P: VAE Parallelism.

📚Model Cache C-P T-P TE-P CN-P VE-P
🔥LongCat-Image ✖️ ✖️
🔥LongCat-Image-Edit ✖️ ✖️
🔥Z-Image-Turbo ✖️ ✖️
🔥Z-Image-ControlNet ✖️
🔥Ovis-Image ✖️ ✖️
🔥HuyuanVideo-1.5 ✖️ ✖️ ✖️ ✖️
🔥FLUX.2-dev ✖️ ✖️
🎉FLUX.1-dev ✖️ ✖️
🎉FLUX.1-Fill-dev ✖️ ✖️
🎉Qwen-Image ✖️ ✖️
🎉Qwen-Image-Edit ✖️ ✖️
🎉Qwen-Image-ControlNet ✖️ ✖️
🎉Qwen-Image-Lightning ✖️ ✖️
🎉Qwen-Image-Edit-Lightning ✖️ ✖️
🎉Wan-2.2 T2V/ITV ✖️ ✖️
🎉Wan-2.2 VACE ✖️ ✖️
🎉Wan-2.1 T2V/ITV ✖️ ✖️
🎉Wan-2.1 VACE ✖️ ✖️
🎉HunyuanImage ✖️ ✖️
🎉HunyuanVideo ✖️ ✖️
🎉FLUX.1-dev Q ✖️ ✖️ ✖️
🎉FLUX.1-Fill-dev Q ✖️ ✖️ ✖️
🎉Qwen-Image Q ✖️ ✖️ ✖️
🎉Qwen-Image-Edit Q ✖️ ✖️ ✖️
🎉Qwen-Image-Lightning Q ✖️ ✖️ ✖️
🎉Qwen-Image-Edit-Lightning Q ✖️ ✖️ ✖️
🎉SkyReelsV2 ✖️ ✖️
🎉LongCatVideo ✖️ ✖️ ✖️ ✖️
🎉ChronoEdit ✖️ ✖️
🎉Kandinsky-5 ✅️ ✅️ ✖️ ✖️
🎉PRX-T2I ✖️ ✖️ ✖️ ✖️
🎉LTXVideo ✖️ ✖️
🎉CogVideoX ✖️ ✖️
🎉CogVideoX-1.5 ✖️ ✖️
🎉CogView4 ✖️ ✖️
🎉CogView3Plus ✖️ ✖️
🎉PixArt Sigma ✖️ ✖️
🎉PixArt Alpha ✖️ ✖️
🎉Chroma-HD ️✅ ✖️ ✖️
🎉VisualCloze ✖️ ✖️
🎉ConsisID ✖️ ✖️
🎉Mochi ✖️ ✖️ ✖️
🎉Lumina 1/2 ✖️ ✖️ ✖️
🎉HiDream ✖️ ✖️ ✖️ ✖️
🎉HunyuanDiT ✖️ ✖️ ✖️
🎉Sana ✖️ ✖️ ✖️ ✖️
🎉Bria ✖️ ✖️ ✖️ ✖️
🎉DiT-XL ✖️ ✖️ ✖️
🎉Allegro ✖️ ✖️ ✖️ ✖️
🎉Cosmos ✖️ ✖️ ✖️ ✖️
🎉OmniGen ✖️ ✖️ ✖️ ✖️
🎉EasyAnimate ✖️ ✖️ ✖️ ✖️
🎉StableDiffusion-3 ✖️ ✖️ ✖️ ✖️
🎉Amused ✖️ ✖️ ✖️ ✖️
🎉AuraFlow ✖️ ✖️ ✖️ ✖️
🔥Click here to show many Image/Video cases🔥

🎉Now, cache-dit covers almost All Diffusers' DiT Pipelines🎉
🔥Qwen-Image | Qwen-Image-Edit | Qwen-Image-Edit-Plus 🔥
🔥FLUX.1 | Qwen-Image-Lightning 4/8 Steps | Wan 2.1 | Wan 2.2 🔥
🔥HunyuanImage-2.1 | HunyuanVideo | HunyuanDiT | HiDream | AuraFlow🔥
🔥CogView3Plus | CogView4 | LTXVideo | CogVideoX | CogVideoX 1.5 | ConsisID🔥
🔥Cosmos | SkyReelsV2 | VisualCloze | OmniGen 1/2 | Lumina 1/2 | PixArt🔥
🔥Chroma | Sana | Allegro | Mochi | SD 3/3.5 | Amused | ... | DiT-XL🔥

🔥Wan2.2 MoE | +cache-dit:2.0x↑🎉 | HunyuanVideo | +cache-dit:2.1x↑🎉

🔥Qwen-Image | +cache-dit:1.8x↑🎉 | FLUX.1-dev | +cache-dit:2.1x↑🎉

🔥Qwen...Lightning | +cache-dit:1.14x↑🎉 | HunyuanImage | +cache-dit:1.7x↑🎉

🔥Qwen-Image-Edit | Input w/o Edit | Baseline | +cache-dit:1.6x↑🎉 | 1.9x↑🎉

🔥FLUX-Kontext-dev | Baseline | +cache-dit:1.3x↑🎉 | 1.7x↑🎉 | 2.0x↑ 🎉

🔥HiDream-I1 | +cache-dit:1.9x↑🎉 | CogView4 | +cache-dit:1.4x↑🎉 | 1.7x↑🎉

🔥CogView3 | +cache-dit:1.5x↑🎉 | 2.0x↑🎉| Chroma1-HD | +cache-dit:1.9x↑🎉

🔥Mochi-1-preview | +cache-dit:1.8x↑🎉 | SkyReelsV2 | +cache-dit:1.6x↑🎉

🔥VisualCloze-512 | Model | Cloth | Baseline | +cache-dit:1.4x↑🎉 | 1.7x↑🎉

🔥LTX-Video-0.9.7 | +cache-dit:1.7x↑🎉 | CogVideoX1.5 | +cache-dit:2.0x↑🎉

🔥OmniGen-v1 | +cache-dit:1.5x↑🎉 | 3.3x↑🎉 | Lumina2 | +cache-dit:1.9x↑🎉

🔥Allegro | +cache-dit:1.36x↑🎉 | AuraFlow-v0.3 | +cache-dit:2.27x↑🎉

🔥Sana | +cache-dit:1.3x↑🎉 | 1.6x↑🎉| PixArt-Sigma | +cache-dit:2.3x↑🎉

🔥PixArt-Alpha | +cache-dit:1.6x↑🎉 | 1.8x↑🎉| SD 3.5 | +cache-dit:2.5x↑🎉

🔥Asumed | +cache-dit:1.1x↑🎉 | 1.2x↑🎉 | DiT-XL-256 | +cache-dit:1.8x↑🎉
♥️ Please consider to leave a ⭐️ Star to support us ~ ♥️

📖Table of Contents

🚀Quick Links

  • 📊Examples - The easiest way to enable hybrid cache acceleration and parallelism for DiTs with cache-dit is to start with our examples for popular models: FLUX, Z-Image, Qwen-Image, Wan, etc.
  • 🌐HTTP Serving - Deploy cache-dit models with HTTP API for text-to-image, image editing, multi-image editing, and text/image-to-video generation.
  • 🎉User Guide - For more advanced features, please refer to the 🎉User_Guide.md for details.
  • ❓FAQ - Frequently asked questions including attention backend configuration, troubleshooting, and optimization tips.

📚Documentation

👋Contribute

How to contribute? Star ⭐️ this repo to support us or check CONTRIBUTE.md.

🎉Projects Using CacheDiT

Here is a curated list of open-source projects integrating CacheDiT, including popular repositories like jetson-containers, flux-fast, sdnext, 🔥stable-diffusion.cpp, 🔥vLLM-Omni, and 🔥SGLang Diffusion. 🎉CacheDiT has been recommended by many famous opensource projects: 🔥Z-Image, 🔥Wan 2.2, 🔥Qwen-Image, 🔥LongCat-Video, Qwen-Image-Lightning, Kandinsky-5, LeMiCa, 🤗diffusers, HelloGitHub and GiantPandaLLM.

©️Acknowledgements

Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and production-level deployment of this project. We learned the design and reused code from the following projects: 🤗diffusers, SGLang, ParaAttention, xDiT, TaylorSeer and LeMiCa.

©️Citations

@misc{cache-dit@2025,
  title={cache-dit: A PyTorch-native and Flexible Inference Engine with Hybrid Cache Acceleration and Parallelism for DiTs.},
  url={https://github.com/vipshop/cache-dit.git},
  note={Open-source software available at https://github.com/vipshop/cache-dit.git},
  author={DefTruth, vipshop.com},
  year={2025}
}

About

🤗 A PyTorch-native Inference Engine with Hybrid Cache Acceleration and Parallelism for DiTs.

Topics

Resources

License

Stars

Watchers

Forks

Packages

No packages published