From df723e646b9eb73180f7632e464d7dbfd3b21a03 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Adrien=20Gallou=C3=ABt?= Date: Thu, 6 Feb 2025 13:24:36 +0000 Subject: [PATCH] Bump llama.cpp & cuda MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: Adrien Gallouët --- Dockerfile_llamacpp | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/Dockerfile_llamacpp b/Dockerfile_llamacpp index 67fb82b5269..6fba85e9d12 100644 --- a/Dockerfile_llamacpp +++ b/Dockerfile_llamacpp @@ -1,6 +1,6 @@ -FROM nvidia/cuda:12.6.3-cudnn-devel-ubuntu24.04 AS deps +FROM nvidia/cuda:12.8.0-cudnn-devel-ubuntu24.04 AS deps -ARG llamacpp_version=b4628 +ARG llamacpp_version=b4651 ARG llamacpp_cuda=OFF ARG cuda_arch=75-real;80-real;86-real;89-real;90-real ENV TGI_LLAMA_PKG_CUDA=cuda-${CUDA_VERSION%.*} @@ -56,7 +56,7 @@ RUN cargo build \ --profile release-opt \ --package text-generation-router-llamacpp --frozen -FROM nvidia/cuda:12.6.3-cudnn-runtime-ubuntu24.04 +FROM nvidia/cuda:12.8.0-cudnn-runtime-ubuntu24.04 RUN apt update && apt install -y \ python3-venv \