image-server-wheels / README.md
dawidope's picture
Trim README
c7d2034 verified
metadata
license: other
tags:
  - wheels
  - cuda
  - pytorch
  - windows
  - linux

image-server-wheels

Prebuilt Python 3.11 wheels.

Contents

File OS CUDA Torch Source Notes
ace_step-1.6.0-py3-none-any.whl any built by us Pure-Python, cross-platform
block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 built by us Used by video pipeline
q8_kernels-0.0.5-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 built by us Used by LTX video
flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl Windows x64 12.8 2.8 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release
flash_attn-2.8.3+cu130torch2.10-cp311-cp311-win_amd64.whl Windows x64 13.0 2.10 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release
flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl Linux x86_64 12.8 2.8 mjun0812/flash-attention-prebuild-wheels Mirror of upstream release

Direct install

BASE=https://huggingface.co/deAPI-ai/image-server-wheels/resolve/main

# Windows
pip install $BASE/q8_kernels-0.0.5-cp311-cp311-win_amd64.whl
pip install $BASE/block_sparse_attn-0.0.2-cp311-cp311-win_amd64.whl
pip install $BASE/flash_attn-2.8.2+cu128torch2.8-cp311-cp311-win_amd64.whl
pip install --no-deps $BASE/ace_step-1.6.0-py3-none-any.whl

# Linux
pip install --no-deps $BASE/flash_attn-2.8.3+cu128torch2.8-cp311-cp311-linux_x86_64.whl

Credits

flash_attn wheels are mirrored from mjun0812/flash-attention-prebuild-wheels — all credit for those builds goes to the upstream author. We mirror them here so the install scripts have a single source of truth and do not break if upstream release URLs change.

The remaining wheels (ace_step, block_sparse_attn, q8_kernels) were built in-house.