image-server-wheels
Prebuilt Python 3.11 wheels.
Contents
| File | OS | CUDA | Torch | Source | Notes |
|---|---|---|---|---|---|
ace_step-1.6.0-py3-none-any.whl |
any | โ | โ | built by us | Pure-Python, cross-platform |
block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl |
Windows x64 | 12.8 | 2.8 | built by us | Used by video pipeline |
q8_kernels-0.0.5-cp311-cp311-win_amd64.whl |
Windows x64 | 12.8 | 2.8 | built by us | Used by LTX video |
flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl |
Windows x64 | 12.8 | 2.8 | mjun0812/flash-attention-prebuild-wheels | Mirror of upstream release |
flash_attn-2.8.3+cu130torch2.10-cp311-cp311-win_amd64.whl |
Windows x64 | 13.0 | 2.10 | mjun0812/flash-attention-prebuild-wheels | Mirror of upstream release |
flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl |
Linux x86_64 | 12.8 | 2.8 | mjun0812/flash-attention-prebuild-wheels | Mirror of upstream release |
Direct install
BASE=https://huggingface.co/deAPI-ai/image-server-wheels/resolve/main
# Windows
pip install $BASE/q8_kernels-0.0.5-cp311-cp311-win_amd64.whl
pip install $BASE/block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl
pip install $BASE/flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl
pip install --no-deps $BASE/ace_step-1.6.0-py3-none-any.whl
# Linux
pip install --no-deps $BASE/flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl
Credits
flash_attn wheels are mirrored from mjun0812/flash-attention-prebuild-wheels โ all credit for those builds goes to the upstream author. We mirror them here so the install scripts have a single source of truth and do not break if upstream release URLs change.
The remaining wheels (ace_step, block_sparse_attn, q8_kernels) were built in-house.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support