cache-dit 1.3.5


pip install cache-dit

  Latest version

Released: Mar 30, 2026


Meta
Author: DefTruth, vipshop.com, etc.
Maintainer: DefTruth, vipshop.com, etc
Requires Python: >=3.10

Classifiers

🤗🎉A PyTorch-native Inference Engine with Hybrid
Cache Acceleration and Massive Parallelism for DiTs
Featured|HelloGitHub

🤗Why Cache-DiT❓❓Cache-DiT is built on top of the 🤗Diffusers library and now supports nearly ALL DiTs from Diffusers. It provides hybrid cache acceleration (DBCache, TaylorSeer, SCM, etc.) and comprehensive parallelism optimizations, including Context Parallelism, Tensor Parallelism, hybrid 2D or 3D parallelism, and dedicated extra parallelism support for Text Encoder, VAE, and ControlNet.

Cache-DiT is compatible with compilation, CPU Offloading, and quantization, fully integrates with SGLang Diffusion, vLLM-Omni, ComfyUI, and runs natively on NVIDIA GPUs, Ascend NPUs and AMD GPUs. Cache-DiT is fast, easy to use, and flexible for various DiTs (online docs at 📘readthedocs.io).

(Let's go beyond Cache-DiT) 👉 SGLang Diffusion | vLLM-Omni | Diffusers

🚀Quick Start

You can install the cache-dit from PyPI or from source:

pip3 install -U cache-dit # or, pip3 install git+https://github.com/vipshop/cache-dit.git

Then accelerate your DiTs with just ♥️one line♥️ of code ~

>>> import cache_dit
>>> from diffusers import DiffusionPipeline
>>> pipe = DiffusionPipeline.from_pretrained("Qwen/Qwen-Image")
>>> cache_dit.enable_cache(pipe) # Cache Acceleration with One-line code.
>>> from cache_dit import DBCacheConfig, ParallelismConfig
>>> cache_dit.enable_cache( # Or, Hybrid Cache Acceleration + Parallelism.
...   pipe, cache_config=DBCacheConfig(), # w/ default
...   parallelism_config=ParallelismConfig(ulysses_size=2))
>>> from cache_dit import DBCacheConfig, ParallelismConfig, QuantizeConfig
>>> cache_dit.enable_cache( # Or, Hybrid Cache + Parallelism + Quantization.
...   pipe, cache_config=DBCacheConfig(), # w/ default
...   parallelism_config=ParallelismConfig(ulysses_size=2),
...   quantize_config=QuantizeConfig(quant_type="..."))
>>> output = pipe(...) # Then, just call the pipe as normal.

For more advanced features, please refer to our online documentation at 📘readthedocs.io.

🌐Community Integration

©️Acknowledgements

Special thanks to vipshop's Computer Vision AI Team for supporting document, testing and deployment of this project. We learned the design and reused code from the following projects: Diffusers, SGLang, vLLM, vLLM-Omni, ParaAttention, xDiT and TaylorSeer.

©️Citations

@misc{cache-dit@2025,
  title={Cache-DiT: A PyTorch-native Inference Engine with Hybrid Cache Acceleration and Massive Parallelism for DiTs.},
  url={https://github.com/vipshop/cache-dit.git},
  note={Open-source software available at https://github.com/vipshop/cache-dit.git},
  author={DefTruth, vipshop.com, etc.},
  year={2025}
}
1.3.5 Mar 30, 2026
1.3.4 Mar 27, 2026
1.3.3 Mar 26, 2026
1.3.2 Mar 26, 2026
1.3.1 Mar 25, 2026
1.3.0 Mar 11, 2026
1.2.3 Feb 26, 2026
1.2.2 Feb 10, 2026
1.2.1 Feb 02, 2026
1.2.0 Jan 16, 2026
1.1.10 Dec 31, 2025
1.1.9 Dec 22, 2025
1.1.8 Dec 10, 2025
1.1.7 Dec 06, 2025
1.1.6 Dec 05, 2025
1.1.5 Dec 05, 2025
1.1.4 Nov 28, 2025
1.1.3 Nov 28, 2025
1.1.2 Nov 24, 2025
1.1.1 Nov 19, 2025
1.1.0 Nov 18, 2025
1.0.16 Nov 17, 2025
1.0.15 Nov 13, 2025
1.0.14 Nov 11, 2025
1.0.13 Nov 07, 2025
1.0.12 Nov 07, 2025
1.0.11 Nov 05, 2025
1.0.10 Oct 30, 2025
1.0.9 Oct 24, 2025
1.0.8 Oct 22, 2025
1.0.7 Oct 22, 2025
1.0.6 Oct 20, 2025
1.0.5 Oct 15, 2025
1.0.4 Oct 14, 2025
1.0.3 Oct 12, 2025
1.0.2 Oct 10, 2025
1.0.1 Sep 26, 2025
1.0.0 Sep 25, 2025
0.3.3 Sep 23, 2025
0.3.2 Sep 22, 2025
0.3.1 Sep 19, 2025
0.3.0 Sep 17, 2025
0.2.37 Sep 17, 2025
0.2.36 Sep 16, 2025
0.2.34 Sep 12, 2025
0.2.33 Sep 10, 2025
0.2.32 Sep 08, 2025
0.2.31 Sep 08, 2025
0.2.30 Sep 05, 2025
0.2.29 Sep 04, 2025
0.2.28 Sep 03, 2025
0.2.27 Sep 01, 2025
0.2.26 Aug 29, 2025
0.2.25 Aug 28, 2025
0.2.24 Aug 26, 2025
0.2.23 Aug 25, 2025
0.2.22 Aug 25, 2025
0.2.21 Aug 22, 2025
0.2.20 Aug 21, 2025
0.2.19 Aug 20, 2025
0.2.18 Aug 20, 2025
0.2.17 Aug 19, 2025
0.2.16 Aug 15, 2025
0.2.15 Aug 11, 2025
0.2.14 Aug 05, 2025
0.2.13 Jul 30, 2025
0.2.12 Jul 24, 2025
0.2.11 Jul 21, 2025
0.2.10 Jul 17, 2025
0.2.9 Jul 13, 2025
0.2.8 Jul 11, 2025
0.2.7 Jul 10, 2025
0.2.6 Jul 09, 2025
0.2.5 Jul 09, 2025
0.2.4 Jul 03, 2025
0.2.3 Jul 01, 2025
0.2.2 Jun 30, 2025
0.2.1 Jun 22, 2025
0.2.0 Jun 20, 2025
0.1.8 Jun 20, 2025
0.1.7 Jun 18, 2025
0.1.6 Jun 18, 2025
0.1.5 Jun 18, 2025
0.1.3 Jun 17, 2025
0.1.2 Jun 17, 2025
0.1.1 Jun 17, 2025
0.1.1.dev2 Jun 16, 2025
0.1.0 Jun 17, 2025

Wheel compatibility matrix

Platform Python 3
any

Files in release

Extras:
Dependencies:
pyyaml
torch (>=2.7.1)
diffusers (>=0.36.0)
transformers (>=4.55.2)