sparsity
Modules
Kernel integrations for sparse attention: Triton FA and diffusers/LTX backends. |
|
Sparsity GEMM kernels (placeholder for future implementations). |
Sparsity kernels: attention (Triton skip-softmax backends) and gemm (placeholder).
Modules
Kernel integrations for sparse attention: Triton FA and diffusers/LTX backends. |
|
Sparsity GEMM kernels (placeholder for future implementations). |
Sparsity kernels: attention (Triton skip-softmax backends) and gemm (placeholder).