From 16c5cdc9a66945a76a536207a3f4778e5b821ee4 Mon Sep 17 00:00:00 2001 From: Andy Bunce Date: Tue, 17 Mar 2026 22:46:51 +0000 Subject: [PATCH] add: model caps --- opencode.cfg/opencode.json | 38 +++++++++++++++++++++++++++++--------- 1 file changed, 29 insertions(+), 9 deletions(-) diff --git a/opencode.cfg/opencode.json b/opencode.cfg/opencode.json index e5ccfec..a095e7e 100644 --- a/opencode.cfg/opencode.json +++ b/opencode.cfg/opencode.json @@ -1,5 +1,5 @@ { - "plugin": [], + "plugin": [], "theme": "system", "model": "halo/qwen3-coder-next:latest", "small_model": "halo/gemma3:4b", @@ -7,19 +7,39 @@ "halo": { "npm": "@ai-sdk/openai-compatible", "name": "(local)", - "options": { "baseURL": "http://host.docker.internal:11434/v1" }, - "models": { "hf.co/unsloth/GLM-4.7-Flash-GGUF:Q6_K": { "name": "glm4.7" }, - "hf.co/LiquidAI/LFM2-24B-A2B-GGUF:Q8_0": { "name": "lfm2" }, - "qwen3-coder-next:latest": { "name": "qwen3" }, - "gemma3:4b": {"name": "Gemma"} - } + "options": { + "baseURL": "http://host.docker.internal:11434/v1" + }, + "models": { + "hf.co/unsloth/GLM-4.7-Flash-GGUF:Q6_K": { + "name": "glm4.7" + }, + "hf.co/LiquidAI/LFM2-24B-A2B-GGUF:Q8_0": { + "name": "lfm2" + }, + "qwen3-coder-next:latest": { + "name": "qwen3" + }, + "gemma3:4b": { + "name": "Gemma", + "capabilities": { + "modalities": { + "input": [ + "text", + "image" + ] + } + } + } } - + } }, "server": { "port": 4096, "hostname": "0.0.0.0", "mdns": true, - "cors": ["https://example.com"] + "cors": [ + "https://example.com" + ] } } \ No newline at end of file