Replies: 1 comment
-
I meet this problem and find a solution.
FROM ghcr.io/ggml-org/llama.cpp:server-vulkan
RUN apt-get update && apt-get install -y --no-install-recommends libglvnd0 libgl1 libglx0 libegl1 libgles2 && rm -rf /var/lib/apt/lists/* (saved as
outputs
Story:When |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
My starting assumption is that I missed something although I tried following the instructions for compiling Vulkan support for a docker image at https://github.com/ggml-org/llama.cpp/blob/master/docs/build.md#for-docker-users.
Image compiles and the container runs using:
docker run -d --name llama-cpp-tmp -p 8010:8080 -v "/mnt/ai-models-llama:/models" --rm -ti --device /dev/dri/renderD128:/dev/dri/renderD128 --device /dev/dri/renderD129:/dev/dri/renderD129 llama-cpp-vulkan:b6503 ...
... but it loads everything onto the CPU...
running the command inside the docker container llama-server --list-devices returns
rather odd considering that on the host vulkaninfo reports multiple gpus and ls -la /dev/dri:
crw-rw---- 1 nobody nogroup 226, 128 Sep 20 00:37 renderD128
crw-rw---- 1 nobody nogroup 226, 129 Sep 20 00:37 renderD129
crw-rw---- 1 nobody nogroup 226, 130 Sep 20 00:37 renderD130
crw-rw---- 1 nobody nogroup 226, 131 Sep 20 00:37 renderD131
crw-rw---- 1 nobody nogroup 226, 132 Sep 20 00:37 renderD132
crw-rw---- 1 nobody nogroup 226, 133 Sep 20 00:37 renderD133
ls -la /dev/dri inside the docker container reports:
which matches the two devices the container was started with...
The host has libvulkan1, mesa-vulkan-drivers, vulkan-tools, mesa-utils installed, as well as cuda-toolkit for docker... this host works fine with CUDA based llama.cpp container... ?!?
the host's uname -a :
Linux gpu-lxc-vllm-1 6.8.12-15-pve #1 SMP PREEMPT_DYNAMIC PMX 6.8.12-15 (2025-09-12T11:02Z) x86_64 GNU/Linux
Any suggestions are welcome.
Beta Was this translation helpful? Give feedback.
All reactions