Files
local-llm/docker-compose.yml
2026-03-10 20:10:53 +01:00

27 lines
475 B
YAML

services:
llama-cpp:
image: ghcr.io/ggml-org/llama.cpp:server
container_name: llama-cpp-qwen
restart: unless-stopped
ports:
- "8000:8000"
volumes:
- ./models:/models
command:
- -m
- /models/qwen3.5-9B-Q5.gguf
- --host
- 0.0.0.0
- --port
- "8000"
- --ctx-size
- "16384"
- --temp
- "0.7"
- --top-p
- "0.8"
- --top-k
- "20"
- --min-p
- "0.00"