fix: use updated ollama ai package
This commit is contained in:
@@ -133,7 +133,7 @@ Environment variable overrides:
|
||||
"openai": { "apiKey": "..." },
|
||||
"google": { "apiKey": "..." },
|
||||
"openrouter": { "apiKey": "..." },
|
||||
"ollama": { "apiBase": "http://localhost:11434/api" }
|
||||
"ollama": { "apiBase": "http://localhost:11434" }
|
||||
},
|
||||
"channels": {
|
||||
"sendProgress": true,
|
||||
@@ -192,7 +192,7 @@ The `agent.model` field is also **required** and should be the model ID without
|
||||
| `openrouter` | `anthropic/claude-sonnet-4-5` (OpenRouter uses its own model IDs) |
|
||||
| `ollama` | `llama3.2`, `qwen2.5` |
|
||||
|
||||
For Ollama, set `providers.ollama.apiBase` (default: `http://localhost:11434/api`).
|
||||
For Ollama, set `providers.ollama.apiBase` (default: `http://localhost:11434`).
|
||||
|
||||
### Mattermost setup
|
||||
|
||||
|
||||
Reference in New Issue
Block a user