Debugging¶
This page collects practical debugging steps for FastVideo inference issues.
Collect Environment Info¶
From the repository root, run:
Attach the output when filing a GitHub issue.
Increase Logging¶
FastVideo logging level is controlled by environment variables:
Useful variables:
FASTVIDEO_LOGGING_LEVEL:DEBUG,INFO,WARNING,ERRORFASTVIDEO_STAGE_LOGGING: print per-stage timings during pipeline executionFASTVIDEO_ATTENTION_BACKEND: force an attention backend (for exampleTORCH_SDPAorFLASH_ATTN)
Common Failure Modes¶
Out-of-memory¶
Try, in order:
- Reduce
height,width,num_frames, ornum_inference_steps. - Enable offloading flags such as
dit_layerwise_offload(single GPU) oruse_fsdp_inference(multi-GPU). - Enable
vae_cpu_offload,image_encoder_cpu_offload, andtext_encoder_cpu_offload.
See Inference Offloading for recommended combinations.
Attention backend import errors¶
If forcing a backend fails, verify optional dependencies are installed:
FLASH_ATTN:flash-attnVIDEO_SPARSE_ATTN:fastvideo-kernelSLIDING_TILE_ATTN: STA legacy workflow insta_do_not_delete+fastvideo-kernelSAGE_ATTN/SAGE_ATTN_THREE: SageAttention packages
As a fallback, use:
Configuration parsing errors¶
When using --config, keep keys aligned with CLI argument names (underscores or
hyphens are both accepted). For nested config values, use nested objects
(vae_config, dit_config) rather than dotted keys.
Issue Template¶
When opening an issue, include:
- exact command or Python snippet,
- model ID/path,
- full traceback,
collect_env.pyoutput,- whether the problem reproduces with
FASTVIDEO_ATTENTION_BACKEND=TORCH_SDPA.