From 8dd5715e10d5a7f22f437a794eddd844e6b8d686 Mon Sep 17 00:00:00 2001 From: Chengjie Li <109656400+ChengjieLi28@users.noreply.github.com> Date: Fri, 29 Nov 2024 18:20:18 +0800 Subject: [PATCH] BUG: llama-cpp-python 0.3.2 build issue in cpu Docker (#2613) --- xinference/deploy/docker/cpu.Dockerfile | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/xinference/deploy/docker/cpu.Dockerfile b/xinference/deploy/docker/cpu.Dockerfile index 21518a5272..716e0a9f9e 100644 --- a/xinference/deploy/docker/cpu.Dockerfile +++ b/xinference/deploy/docker/cpu.Dockerfile @@ -21,7 +21,7 @@ ARG PIP_INDEX=https://pypi.org/simple RUN python -m pip install --upgrade -i "$PIP_INDEX" pip && \ pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu && \ pip install -i "$PIP_INDEX" --upgrade-strategy only-if-needed -r /opt/inference/xinference/deploy/docker/requirements_cpu.txt && \ - CMAKE_ARGS="-DGGML_BLAS=ON -DGGML_BLAS_VENDOR=OpenBLAS" pip install llama-cpp-python && \ + pip install llama-cpp-python && \ cd /opt/inference && \ python setup.py build_web && \ git restore . && \