diff --git a/opencode.cfg/opencode.json b/opencode.cfg/opencode.json index 231919c..41ba81a 100644 --- a/opencode.cfg/opencode.json +++ b/opencode.cfg/opencode.json @@ -36,68 +36,69 @@ "output": [ "text" ] - }, - "qwen3-coder-next:latest": { - "name": "qwen3", - "family": "qwen3next", - "reasoning": false, - "tool_call": true, - "limit": { - "context": 128000, // Context window size - "output": 8192 // Output token limit - }, - "options": { - "num_ctx": 16000, // Context window size for Ollama - "temperature": 1.0, - "top_p": 0.95, - "top_k": 40, - "max_tokens": 8192 // Max response length - }, - "modalities": { - "input": [ - "text" - ], - "output": [ - "text" - ] - } - }, - "gemma3:4b": { - "name": "Gemma 3:4b", - "family": "gemma", - "temperature": false, - "reasoning": false, - "attachment": true, - "tool_call": false, - "modalities": { - "input": [ - "text", - "image" - ], // Explicitly declare image support - "output": [ - "text" - ] - }, - "limit": { - "context": 100000, // Maximum context window (input + output) - "output": 8192 // Maximum output tokens - }, - "options": { - "num_ctx": 16000 // Context window size for Ollama - }, - "headers": {}, - "release_date": "13 March 2025", - "variants": {} } + }, + "qwen3-coder-next:latest": { + "name": "qwen3", + "family": "qwen3next", + "reasoning": false, + "tool_call": true, + "limit": { + "context": 128000, // Context window size + "output": 8192 // Output token limit + }, + "options": { + "num_ctx": 16000, // Context window size for Ollama + "temperature": 1.0, + "top_p": 0.95, + "top_k": 40, + "max_tokens": 8192 // Max response length + }, + "modalities": { + "input": [ + "text" + ], + "output": [ + "text" + ] + } + }, + "gemma3:4b": { + "name": "Gemma 3:4b", + "family": "gemma", + "temperature": false, + "reasoning": false, + "attachment": true, + "tool_call": false, + "modalities": { + "input": [ + "text", + "image" + ], // Explicitly declare image support + "output": [ + "text" + ] + }, + "limit": { + "context": 100000, // Maximum context window (input + output) + "output": 8192 // Maximum output tokens + }, + "options": { + "num_ctx": 16000 // Context window size for Ollama + }, + "headers": {}, + "release_date": "13 March 2025", + "variants": {} } } - }, - "server": { - "port": 4096, - "hostname": "0.0.0.0", - "mdns": false, - "cors": [ - "https://example.com" - ] } - } \ No newline at end of file + }, + "server": { + "port": 4096, + "hostname": "0.0.0.0", + "mdns": false, + "cors": [ + "https://example.com" + ] + } +} \ No newline at end of file