SenCache: Accelerating Diffusion Model Inference
via Sensitivity-Aware Caching

Yasaman Haghighi Alexandre Alahi
Ecole Polytechnique Fédérale de Lausanne (EPFL)

Abstract

Diffusion models achieve state-of-the-art video generation quality, but their inference remains expensive due to the large number of sequential denoising steps. This has motivated a growing line of research on accelerating diffusion inference. Among training-free acceleration methods, caching reduces computation by reusing previously computed model outputs across timesteps. Existing caching methods rely on heuristic criteria to choose cache/reuse timesteps and require extensive tuning. We address this limitation with a principled sensitivity-aware caching framework. Specifically, we formalize the caching error through an analysis of the model output sensitivity to perturbations in the denoising inputs, i.e., the noisy latent and the timestep, and show that this sensitivity is a key predictor of caching error. Based on this analysis, we propose Sensitivity-Aware Caching (SenCache), a dynamic caching policy that adaptively selects caching timesteps on a per-sample basis. Our framework provides a theoretical basis for adaptive caching, explains why prior empirical heuristics can be partially effective, and extends them to a dynamic, sample-specific approach. Experiments on Wan 2.1, CogVideoX, and LTX-Video show that SenCache achieves better visual quality than existing caching methods under similar computational budgets.

SenCache caching criterion overview
SenCache uses sensitivity as a caching criterion. At each denoising step, if the changes in the noisy latent xt and the sampling step t are sufficiently small such that the sensitivity score falls below ε, we reuse the cached denoiser output; otherwise, we refresh the cache at the current state. By skipping expensive denoiser evaluations when the output is expected to change minimally, SenCache accelerates diffusion-model inference.

Qualitative Comparisons

Side-by-side comparison of videos generated with the Baseline, MagCache, and SenCache (Ours). All methods use the same prompt and share the same noise seed.

Difference: the dog and the vase

Baseline (CogVideoX)

MagCache

SenCache (Ours)

Difference: the character and the background

Baseline (LTX-Video)

MagCache

SenCache (Ours)

Difference: the bird has a third wing in the middle video

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the treasurer

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the building

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the mountaineer

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the shoes

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the roof of the chalet

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the airplane in the background

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the parrot

Baseline (LTX-Video)

MagCache

SenCache (Ours)

Difference: the eyes of the pinguin

Baseline (CogVideoX)

MagCache

SenCache (Ours)

Difference: the color of the woman's top

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Difference: the brand of the machine

Baseline (WAN 2.1)

MagCache

SenCache (Ours)

Citation

If you find SenCache useful, please cite our work.

@article{haghighi2026sencache,
  title={SenCache: Accelerating Diffusion Model Inference via Sensitivity-Aware Caching},
  author={Haghighi, Yasaman and Alahi, Alexandre},
  journal={arXiv preprint arXiv:2602.24208},
  year={2026}
}