Skip to content
Merged
Changes from 1 commit
Commits
Show all changes
34 commits
Select commit Hold shift + click to select a range
b57f399
feat: Add backend gallery
mudler Jun 9, 2025
5226227
Add backends docs
mudler Jun 11, 2025
3099afb
wip: Backend Dockerfile for python backends
mudler Jun 11, 2025
2ba3899
feat: drop extras images, build python backends separately
mudler Jun 12, 2025
eb5707c
fixup on all backends
mudler Jun 12, 2025
9bd1478
test CI
mudler Jun 12, 2025
389d4e2
Tweaks
mudler Jun 12, 2025
ee692be
Drop old backends leftovers
mudler Jun 12, 2025
c07c577
Fixup CI
mudler Jun 12, 2025
6d96f00
Move dockerfile upper
mudler Jun 12, 2025
6b90ef5
Fix proto
mudler Jun 12, 2025
97d3176
Feature dropped for consistency - we prefer model galleries
mudler Jun 13, 2025
06234d6
Merge branch 'master' into feat/backend_gallery
mudler Jun 13, 2025
4493bc4
Add missing packages in the build image
mudler Jun 13, 2025
98573b4
exllama is ponly available on cublas
mudler Jun 13, 2025
06628f8
pin torch on chatterbox
mudler Jun 13, 2025
427bbf4
Fixups to index
mudler Jun 13, 2025
c36deef
CI
mudler Jun 13, 2025
1a924b2
Debug CI
mudler Jun 13, 2025
5751d2e
Install accellerators deps
mudler Jun 13, 2025
7402a4e
Add target arch
mudler Jun 13, 2025
28f90e5
Add cuda minor version
mudler Jun 13, 2025
5e83023
Use self-hosted runners
mudler Jun 13, 2025
ed39960
ci: use quay for test images
mudler Jun 14, 2025
1545d31
fixups for vllm and chatterbox
mudler Jun 14, 2025
a427c7a
Small fixups on CI
mudler Jun 14, 2025
7da886d
chatterbox is only available for nvidia
mudler Jun 14, 2025
1699c96
Simplify CI builds
mudler Jun 14, 2025
3b63a71
Adapt test, use qwen3
mudler Jun 14, 2025
5d1241d
Merge branch 'master' into feat/backend_gallery
mudler Jun 14, 2025
ff820f8
chore(model gallery): add jina-reranker-v1-tiny-en-gguf
mudler Jun 14, 2025
211891c
fix(gguf-parser): recover from potential panics that can happen while…
mudler Jun 14, 2025
ad05b49
Use reranker from llama.cpp in AIO images
mudler Jun 14, 2025
f12e009
Limit concurrent jobs
mudler Jun 15, 2025
File filter

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Install accellerators deps
Signed-off-by: Ettore Di Giacinto <[email protected]>
  • Loading branch information
mudler committed Jun 13, 2025
commit 5751d2e793524b500e809ff817f192185acffb1a
63 changes: 63 additions & 0 deletions backend/Dockerfile.python
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,7 @@ ARG BACKEND=rerankers
ARG BUILD_TYPE
ENV BUILD_TYPE=${BUILD_TYPE}
ARG CUDA_MAJOR_VERSION
ARG SKIP_DRIVERS=false
ENV CUDA_MAJOR_VERSION=${CUDA_MAJOR_VERSION}
ENV DEBIAN_FRONTEND=noninteractive

Expand All @@ -28,7 +29,69 @@ RUN apt-get update && \
rm -rf /var/lib/apt/lists/* && \
pip install --upgrade pip

# Vulkan requirements
RUN <<EOT bash
if [ "${BUILD_TYPE}" = "vulkan" ] && [ "${SKIP_DRIVERS}" = "false" ]; then
apt-get update && \
apt-get install -y --no-install-recommends \
software-properties-common pciutils wget gpg-agent && \
wget -qO - https://packages.lunarg.com/lunarg-signing-key-pub.asc | apt-key add - && \
wget -qO /etc/apt/sources.list.d/lunarg-vulkan-jammy.list https://packages.lunarg.com/vulkan/lunarg-vulkan-jammy.list && \
apt-get update && \
apt-get install -y \
vulkan-sdk && \
apt-get clean && \
rm -rf /var/lib/apt/lists/*
fi
EOT

# CuBLAS requirements
RUN <<EOT bash
if [ "${BUILD_TYPE}" = "cublas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then
apt-get update && \
apt-get install -y --no-install-recommends \
software-properties-common pciutils
if [ "amd64" = "$TARGETARCH" ]; then
curl -O https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/cuda-keyring_1.1-1_all.deb
fi
if [ "arm64" = "$TARGETARCH" ]; then
curl -O https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/arm64/cuda-keyring_1.1-1_all.deb
fi
dpkg -i cuda-keyring_1.1-1_all.deb && \
rm -f cuda-keyring_1.1-1_all.deb && \
apt-get update && \
apt-get install -y --no-install-recommends \
cuda-nvcc-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} \
libcufft-dev-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} \
libcurand-dev-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} \
libcublas-dev-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} \
libcusparse-dev-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} \
libcusolver-dev-${CUDA_MAJOR_VERSION}-${CUDA_MINOR_VERSION} && \
apt-get clean && \
rm -rf /var/lib/apt/lists/*
fi
EOT

# If we are building with clblas support, we need the libraries for the builds
RUN if [ "${BUILD_TYPE}" = "clblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \
apt-get update && \
apt-get install -y --no-install-recommends \
libclblast-dev && \
apt-get clean && \
rm -rf /var/lib/apt/lists/* \
; fi

RUN if [ "${BUILD_TYPE}" = "hipblas" ] && [ "${SKIP_DRIVERS}" = "false" ]; then \
apt-get update && \
apt-get install -y --no-install-recommends \
hipblas-dev \
rocblas-dev && \
apt-get clean && \
rm -rf /var/lib/apt/lists/* && \
# I have no idea why, but the ROCM lib packages don't trigger ldconfig after they install, which results in local-ai and others not being able
# to locate the libraries. We run ldconfig ourselves to work around this packaging deficiency
ldconfig \
; fi
# Install uv as a system package
RUN curl -LsSf https://astral.sh/uv/install.sh | UV_INSTALL_DIR=/usr/bin sh
ENV PATH="/root/.cargo/bin:${PATH}"
Expand Down
Loading