From 1e935fdec5fab31617f1ca881dae102b8e187a27 Mon Sep 17 00:00:00 2001 From: Andy Bunce Date: Thu, 23 Apr 2026 22:58:27 +0100 Subject: [PATCH] [fix] --- llama.cpp/compose.yml | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/llama.cpp/compose.yml b/llama.cpp/compose.yml index 1cf3342..0549b82 100644 --- a/llama.cpp/compose.yml +++ b/llama.cpp/compose.yml @@ -5,7 +5,7 @@ services: ports: - 8878:8080 volumes: - - /home/andy/.lmstudio/models/lmstudio-community:/root/.cache/llama.cpp/ + - /home/andy/.lmstudio/models/lmstudio-community:/models # Optionally mount a local .gguf file directly: # - /absolute/path/to/your-model.gguf:/model.gguf:ro devices: @@ -33,6 +33,7 @@ services: restart: unless-stopped command: > + --models-dir /models --n-gpu-layers all --parallel 2 --flash-attn on