image-server-wheels

Prebuilt Python 3.11 wheels.

Contents

File OS CUDA Torch Source Notes
ace_step-1.6.0-py3-none-any.whl any โ€” โ€” built by us Pure-Python, cross-platform
block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 built by us Used by video pipeline
q8_kernels-0.0.5-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 built by us Used by LTX video
flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release
flash_attn-2.8.3+cu130torch2.10-cp311-cp311-win_amd64.whl Windows x64 13.0 2.10 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release
flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl Linux x86_64 12.8 2.8 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release

Direct install

BASE=https://huggingface.co/deAPI-ai/image-server-wheels/resolve/main

# Windows
pip install $BASE/q8_kernels-0.0.5-cp311-cp311-win_amd64.whl
pip install $BASE/block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl
pip install $BASE/flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl
pip install --no-deps $BASE/ace_step-1.6.0-py3-none-any.whl

# Linux
pip install --no-deps $BASE/flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl

Credits

flash_attn wheels are mirrored from mjun0812/flash-attention-prebuild-wheels โ€” all credit for those builds goes to the upstream author. We mirror them here so the install scripts have a single source of truth and do not break if upstream release URLs change.

The remaining wheels (ace_step, block_sparse_attn, q8_kernels) were built in-house.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support