Found 75 repositories(showing 30)
thu-ml
TurboDiffusion: 100–200× Acceleration for Video Diffusion Models
thu-ml
[ICLR2025, ICML2025, NeurIPS2025 Spotlight] Quantized Attention achieves speedup of 2-5x compared to FlashAttention, without losing end-to-end metrics across language, image, and video models.
thu-ml
[ICML2025] SpargeAttention: A training-free sparse attention that accelerates any model inference.
wildminder
Pre-compiled Python whl for Flash-attention, SageAttention, NATTEN, xFormer etc
Cross-platform installer for Triton and SageAttention on ComfyUI. Simplifies GPU-accelerated inference setup for Windows users with automated dependency management and RTX 5090 support.
Grey3016
A batch file to install the latest ComfyUI, make a venv within it and automatically install Triton and SageAttention for Hunyaun etc workflows.
Powershell-Script to install Sageattention in Comfyui for windows portable edition
BradPita
🪟 为 Windows AI 开发者提供预编译 wheel 文件的集中仓库 | 自动抓取并整理 PyTorch、Flash Attention、xformers、SageAttention 等常用库的最新版本 | 免编译,开箱即用 | 特别适合 ComfyUI 和 Stable Diffusion 用户
Ph0rk0z
Sage attention for turning.
Rogala
Pre-compiled acceleration packages (xFormers, SageAttention, Flash Attention) for ComfyUI on Windows — NVIDIA RTX 5000 Series (Blackwell, SM120).
Automatically benchmark and optimize attention in diffusion models. 1.5-2x speedup on RTX 4090.
eddyhhlure1Eddy
No description available
wallen0322
An experimental node
freyandere
Triton + SageAttention 2++(2.2) install
eddyhhlure1Eddy
https://github.com/thu-ml/SageAttention
No description available
mobcat40
Prebuilt SageAttention 2.2.0 wheel for RTX 5090/5080/5070 (Blackwell sm_120) with PyTorch 2.11 nightly and CUDA 12.8. ~35% faster diffusion sampling. Includes build instructions for the PyTorch header patch fix.
Light-x02
Automatic installation of ComfyUI with Python 3.12.9, Torch 2.9.0.dev20250716+cu128, Triton 3.3.1.post19, SageAttention enabled, and fp16 accumulation.
Cleaned up parameterized script for running dashtoon's Hunyuan Keyframe Control lora for start/end frames I2V, with ffmpeg, batching, and other options/fixes (including cpu offloading and sageattention/flash).
loscrossos
Helper Project with Nvidia 50 Series support
LinkSoulsAI
Provides a robust, two-stage Docker setup for deploying ComfyUI with high-performance custom nodes like SageAttention and Nunchaku.
ecarmen16
ComfyUI + SageAttention for NVIDIA DGX Spark (Blackwell GB10, ARM64, sm_121)
typemovie
TypeMovie-ParaAttention is an enhanced version of ParaAttention, designed to accelerate Diffusion Transformer (DiT) model inference with context parallelism, dynamic caching, and a new high-performance SageAttention backend
ScalierBullet63
The easiest way to install SageAttention on ComfyUI on Windows!
nrgumn-ai-creation
One-click Windows installer for ComfyUI with PyTorch 2.8, Triton and SageAttention
snw35
Wheel builder for Sageattention
ZhiqiJiang
对https://github.com/thu-ml/SageAttention进行算子融合优化
BlackMixture
A guide to get Triton and Sage Attention working on Windows
bxone-git
Dance SCAIL BPlan - SCAIL dance video generation on RunPod Serverless (RTX 5090 + SageAttention)
matheohan
An all-in-one docker image that runs the latest ComfyUI with SageAttention.