mod: config for gemma

This commit is contained in:
Andy Bunce 2026-03-19 11:08:38 +00:00
parent 7fa236923d
commit a342645c4b

View file

@ -1,8 +1,8 @@
{ {
"$schema": "https://opencode.ai/config.json",
"plugin": [], "plugin": [],
"theme": "system",
"model": "halo/qwen3-coder-next:latest", "model": "halo/qwen3-coder-next:latest",
"small_model": "halo/gemma3:4b", "small_model": "gemma3:4b",
"provider": { "provider": {
"halo": { "halo": {
"npm": "@ai-sdk/openai-compatible", "npm": "@ai-sdk/openai-compatible",
@ -22,43 +22,29 @@
}, },
"gemma3:4b": { "gemma3:4b": {
"name": "Gemma *", "name": "Gemma *",
"capabilities": {
"temperature": false, "temperature": false,
"reasoning": false, "reasoning": false,
"attachment": false, "attachment": true,
"toolcall": true, "tool_call": false,
"input": { "modalities": {
"text": true, "input": [
"audio": false, "text",
"image": true, "image"
"video": false, ], // Explicitly declare image support
"pdf": false "output": [
"text"
]
}, },
"output": {
"text": true,
"audio": false,
"image": false,
"video": false,
"pdf": false
},
"interleaved": false
},
"cost": {
"input": 0,
"output": 0,
"cache": {
"read": 0,
"write": 0
}
},
"options": {},
"limit": { "limit": {
"context": 0, "context": 100000, // Maximum context window (input + output)
"output": 0 "output": 8192 // Maximum output tokens
},
"options": {
"num_ctx": 16000 // Context window size for Ollama
}, },
"headers": {}, "headers": {},
"family": "", "family": "gemma",
"release_date": "", "release_date": "13 March 2025",
"variants": {} "variants": {}
} }
} }
@ -67,7 +53,7 @@
"server": { "server": {
"port": 4096, "port": 4096,
"hostname": "0.0.0.0", "hostname": "0.0.0.0",
"mdns": true, "mdns": false,
"cors": [ "cors": [
"https://example.com" "https://example.com"
] ]