Compare commits
1 Commits
main
...
feat/skill
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
66fb080297 |
4
.gitignore
vendored
4
.gitignore
vendored
@@ -1,10 +1,6 @@
|
||||
# dependencies (bun install)
|
||||
node_modules
|
||||
|
||||
# editors
|
||||
.vscode
|
||||
.openvscode-server
|
||||
|
||||
# output
|
||||
out
|
||||
dist
|
||||
|
||||
@@ -5,8 +5,7 @@
|
||||
"correctness": "warn"
|
||||
},
|
||||
"rules": {
|
||||
"eslint/no-unused-vars": "error",
|
||||
"unicorn/no-nested-ternary": "error"
|
||||
"eslint/no-unused-vars": "error"
|
||||
},
|
||||
"options": {
|
||||
"typeAware": true,
|
||||
|
||||
92
README.md
92
README.md
@@ -18,17 +18,13 @@ bun install # or use `mise install`
|
||||
|
||||
## Quick start
|
||||
|
||||
**1. Initialize workspace**
|
||||
**1. Create a config file**
|
||||
|
||||
```bash
|
||||
bun run nanobot onboard
|
||||
mkdir -p ~/.nanobot
|
||||
```
|
||||
|
||||
This creates `~/.config/nanobot/` with a config file and templates.
|
||||
|
||||
**2. Edit config**
|
||||
|
||||
Add your API key and set provider/model:
|
||||
`~/.nanobot/config.json`:
|
||||
|
||||
```json
|
||||
{
|
||||
@@ -38,16 +34,15 @@ Add your API key and set provider/model:
|
||||
}
|
||||
},
|
||||
"agent": {
|
||||
"provider": "openrouter",
|
||||
"model": "anthropic/claude-sonnet-4-5"
|
||||
"model": "openrouter/anthropic/claude-sonnet-4-5"
|
||||
}
|
||||
}
|
||||
```
|
||||
|
||||
**3. Chat**
|
||||
**2. Chat**
|
||||
|
||||
```bash
|
||||
bun run nanobot agent
|
||||
bun run start agent
|
||||
```
|
||||
|
||||
That's it.
|
||||
@@ -59,21 +54,22 @@ That's it.
|
||||
Chat with the agent from your terminal. Does not require a running gateway.
|
||||
|
||||
```
|
||||
bun run nanobot agent [options]
|
||||
bun run start agent [options]
|
||||
```
|
||||
|
||||
| Option | Description |
|
||||
|--------|-------------|
|
||||
| `-c, --config <path>` | Path to `config.json` (default: `~/.config/nanobot/config.json`) |
|
||||
| `-c, --config <path>` | Path to `config.json` (default: `~/.nanobot/config.json`) |
|
||||
| `-m, --message <text>` | Send a single message and exit (non-interactive) |
|
||||
| `-w, --workspace <path>` | Override the workspace directory |
|
||||
| `-M, --model <model>` | Override the model for this session |
|
||||
|
||||
**Interactive mode** (default when no `-m` is given):
|
||||
|
||||
```bash
|
||||
bun run nanobot agent
|
||||
bun run nanobot agent -c ~/.config/nanobot-work/config.json
|
||||
bun run nanobot agent -w /tmp/scratch
|
||||
bun run start agent
|
||||
bun run start agent -c ~/.nanobot-work/config.json
|
||||
bun run start agent -w /tmp/scratch
|
||||
```
|
||||
|
||||
Press `Ctrl+C` to exit.
|
||||
@@ -81,8 +77,8 @@ Press `Ctrl+C` to exit.
|
||||
**Single-shot mode:**
|
||||
|
||||
```bash
|
||||
bun run nanobot agent -m "What time is it in Tokyo?"
|
||||
bun run nanobot agent -m "Summarize the file ./notes.md"
|
||||
bun run start agent -m "What time is it in Tokyo?"
|
||||
bun run start agent -m "Summarize the file ./notes.md"
|
||||
```
|
||||
|
||||
### `gateway` — Mattermost bot
|
||||
@@ -90,23 +86,23 @@ bun run nanobot agent -m "Summarize the file ./notes.md"
|
||||
Runs the full stack: Mattermost WebSocket channel, agent loop, cron scheduler, and heartbeat.
|
||||
|
||||
```
|
||||
bun run nanobot gateway [options]
|
||||
bun run start gateway [options]
|
||||
```
|
||||
|
||||
| Option | Description |
|
||||
|--------|-------------|
|
||||
| `-c, --config <path>` | Path to `config.json` (default: `~/.config/nanobot/config.json`) |
|
||||
| `-c, --config <path>` | Path to `config.json` (default: `~/.nanobot/config.json`) |
|
||||
|
||||
```bash
|
||||
bun run nanobot gateway
|
||||
bun run nanobot gateway -c ~/.config/nanobot-work/config.json
|
||||
bun run start gateway
|
||||
bun run start gateway -c ~/.nanobot-work/config.json
|
||||
```
|
||||
|
||||
Handles `SIGINT` / `SIGTERM` for graceful shutdown.
|
||||
|
||||
## Configuration
|
||||
|
||||
Config file: `~/.config/nanobot/config.json` (or pass `-c <path>` to any command).
|
||||
Config file: `~/.nanobot/config.json` (or pass `-c <path>` to any command).
|
||||
|
||||
Environment variable overrides:
|
||||
|
||||
@@ -114,15 +110,15 @@ Environment variable overrides:
|
||||
|----------|-------------------|
|
||||
| `NANOBOT_CONFIG` | path to config file |
|
||||
| `NANOBOT_MODEL` | `agent.model` |
|
||||
| `NANOBOT_WORKSPACE` | `agent.workspacePath` |
|
||||
|
||||
### Full config reference
|
||||
|
||||
```json
|
||||
{
|
||||
"agent": {
|
||||
"provider": "openrouter",
|
||||
"model": "anthropic/claude-sonnet-4-5",
|
||||
"workspacePath": "~/.config/nanobot",
|
||||
"model": "openrouter/anthropic/claude-sonnet-4-5",
|
||||
"workspacePath": "~/.nanobot",
|
||||
"maxTokens": 4096,
|
||||
"contextWindowTokens": 65536,
|
||||
"temperature": 0.7,
|
||||
@@ -133,7 +129,7 @@ Environment variable overrides:
|
||||
"openai": { "apiKey": "..." },
|
||||
"google": { "apiKey": "..." },
|
||||
"openrouter": { "apiKey": "..." },
|
||||
"ollama": { "apiBase": "http://localhost:11434" }
|
||||
"ollama": { "apiBase": "http://localhost:11434/api" }
|
||||
},
|
||||
"channels": {
|
||||
"sendProgress": true,
|
||||
@@ -170,29 +166,19 @@ Environment variable overrides:
|
||||
}
|
||||
```
|
||||
|
||||
### Provider
|
||||
### Providers
|
||||
|
||||
The `agent.provider` field is **required** and must be one of:
|
||||
Model names use a `provider/model` prefix scheme:
|
||||
|
||||
| Provider | Description |
|
||||
|----------|-------------|
|
||||
| `anthropic` | Anthropic direct (Claude models) |
|
||||
| `openai` | OpenAI direct (GPT models) |
|
||||
| `google` | Google direct (Gemini models) |
|
||||
| `openrouter` | OpenRouter (access to many models) |
|
||||
| `ollama` | Local Ollama instance |
|
||||
| Prefix | Provider | Example |
|
||||
|--------|----------|---------|
|
||||
| `anthropic/` | Anthropic direct | `anthropic/claude-opus-4-5` |
|
||||
| `openai/` | OpenAI direct | `openai/gpt-4o` |
|
||||
| `google/` | Google direct | `google/gemini-2.5-pro` |
|
||||
| `openrouter/` | OpenRouter (any model) | `openrouter/anthropic/claude-sonnet-4-5` |
|
||||
| `ollama/` | Local Ollama | `ollama/llama3.2` |
|
||||
|
||||
The `agent.model` field is also **required** and should be the model ID without any provider prefix:
|
||||
|
||||
| Provider | Example Model |
|
||||
|----------|---------------|
|
||||
| `anthropic` | `claude-sonnet-4-5`, `claude-opus-4-5` |
|
||||
| `openai` | `gpt-4o`, `gpt-4o-mini` |
|
||||
| `google` | `gemini-2.5-pro`, `gemini-2.0-flash` |
|
||||
| `openrouter` | `anthropic/claude-sonnet-4-5` (OpenRouter uses its own model IDs) |
|
||||
| `ollama` | `llama3.2`, `qwen2.5` |
|
||||
|
||||
For Ollama, set `providers.ollama.apiBase` (default: `http://localhost:11434`).
|
||||
For Ollama, set `providers.ollama.apiBase` (default: `http://localhost:11434/api`).
|
||||
|
||||
### Mattermost setup
|
||||
|
||||
@@ -212,7 +198,7 @@ For Ollama, set `providers.ollama.apiBase` (default: `http://localhost:11434`).
|
||||
}
|
||||
```
|
||||
|
||||
4. Run `bun run nanobot gateway`
|
||||
4. Run `bun run start gateway`
|
||||
|
||||
`allowFrom` controls which users the bot responds to. Use `["*"]` to allow all users.
|
||||
|
||||
@@ -247,10 +233,10 @@ Run separate instances with different configs — useful for isolated workspaces
|
||||
|
||||
```bash
|
||||
# Instance A
|
||||
bun run nanobot gateway -c ~/.config/nanobot-a/config.json
|
||||
bun run start gateway -c ~/.nanobot-a/config.json
|
||||
|
||||
# Instance B
|
||||
bun run nanobot gateway -c ~/.config/nanobot-b/config.json
|
||||
bun run start gateway -c ~/.nanobot-b/config.json
|
||||
```
|
||||
|
||||
Each instance needs its own config file. Set a different `agent.workspacePath` per instance to keep memory, sessions, and cron jobs isolated:
|
||||
@@ -258,7 +244,7 @@ Each instance needs its own config file. Set a different `agent.workspacePath` p
|
||||
```json
|
||||
{
|
||||
"agent": {
|
||||
"workspacePath": "~/.config/nanobot-a"
|
||||
"workspacePath": "~/.nanobot-a"
|
||||
}
|
||||
}
|
||||
```
|
||||
@@ -266,10 +252,10 @@ Each instance needs its own config file. Set a different `agent.workspacePath` p
|
||||
To run a local CLI session against a specific instance:
|
||||
|
||||
```bash
|
||||
bun run nanobot agent -c ~/.config/nanobot-a/config.json -m "Hello"
|
||||
bun run start agent -c ~/.nanobot-a/config.json -m "Hello"
|
||||
|
||||
# Temporarily override the workspace for a one-off run
|
||||
bun run nanobot agent -c ~/.config/nanobot-a/config.json -w /tmp/scratch
|
||||
bun run start agent -c ~/.nanobot-a/config.json -w /tmp/scratch
|
||||
```
|
||||
|
||||
## Linux service (systemd)
|
||||
|
||||
18
bun.lock
18
bun.lock
@@ -11,12 +11,12 @@
|
||||
"@mozilla/readability": "^0.6.0",
|
||||
"@openrouter/ai-sdk-provider": "^2.3.0",
|
||||
"ai": "^6.0.116",
|
||||
"ai-sdk-ollama": "^3.8.0",
|
||||
"commander": "^14.0.3",
|
||||
"cron-parser": "^5.5.0",
|
||||
"js-tiktoken": "^1.0.21",
|
||||
"jsonrepair": "^3.13.3",
|
||||
"node-html-parser": "^7.1.0",
|
||||
"ollama-ai-provider": "^1.2.0",
|
||||
"picocolors": "^1.1.1",
|
||||
"zod": "^4.3.6",
|
||||
},
|
||||
@@ -151,8 +151,6 @@
|
||||
|
||||
"ai": ["ai@6.0.116", "", { "dependencies": { "@ai-sdk/gateway": "3.0.66", "@ai-sdk/provider": "3.0.8", "@ai-sdk/provider-utils": "4.0.19", "@opentelemetry/api": "1.9.0" }, "peerDependencies": { "zod": "^3.25.76 || ^4.1.8" } }, "sha512-7yM+cTmyRLeNIXwt4Vj+mrrJgVQ9RMIW5WO0ydoLoYkewIvsMcvUmqS4j2RJTUXaF1HphwmSKUMQ/HypNRGOmA=="],
|
||||
|
||||
"ai-sdk-ollama": ["ai-sdk-ollama@3.8.0", "", { "dependencies": { "@ai-sdk/provider": "^3.0.8", "@ai-sdk/provider-utils": "^4.0.15", "jsonrepair": "^3.13.2", "ollama": "^0.6.3" }, "peerDependencies": { "ai": "^6.0.89" } }, "sha512-Nlla8FpK8QFMNh9m8sPCZoNqnr+n+Ud0QTqpXNds4j/b/lbVZGaji13ZcRuuFvBwPwd4xnFkNrijJzi70Ih1Tg=="],
|
||||
|
||||
"base64-js": ["base64-js@1.5.1", "", {}, "sha512-AKpaYlHn8t4SVbOHCy+b5+KKgvR4vrsD8vbvrbiQJps7fKDTkjkDry6ji0rUJjC0kzbNePLwzxq8iypo41qeWA=="],
|
||||
|
||||
"boolbase": ["boolbase@1.0.0", "", {}, "sha512-JZOSA7Mo9sNGB8+UjSgzdLtokWAky1zbztM3WRLCbZ70/3cTANmQmOdR7y2g+J0e2WXywy1yS468tY+IruqEww=="],
|
||||
@@ -189,11 +187,13 @@
|
||||
|
||||
"luxon": ["luxon@3.7.2", "", {}, "sha512-vtEhXh/gNjI9Yg1u4jX/0YVPMvxzHuGgCm6tC5kZyb08yjGWGnqAjGJvcXbqQR2P3MyMEFnRbpcdFS6PBcLqew=="],
|
||||
|
||||
"nanoid": ["nanoid@3.3.11", "", { "bin": { "nanoid": "bin/nanoid.cjs" } }, "sha512-N8SpfPUnUp1bK+PMYW8qSWdl9U+wwNWI4QKxOYDy9JAro3WMX7p2OeVRF9v+347pnakNevPmiHhNmZ2HbFA76w=="],
|
||||
|
||||
"node-html-parser": ["node-html-parser@7.1.0", "", { "dependencies": { "css-select": "^5.1.0", "he": "1.2.0" } }, "sha512-iJo8b2uYGT40Y8BTyy5ufL6IVbN8rbm/1QK2xffXU/1a/v3AAa0d1YAoqBNYqaS4R/HajkWIpIfdE6KcyFh1AQ=="],
|
||||
|
||||
"nth-check": ["nth-check@2.1.1", "", { "dependencies": { "boolbase": "^1.0.0" } }, "sha512-lqjrjmaOoAnWfMmBPL+XNnynZh2+swxiX3WUE0s4yEHI6m+AwrK2UZOimIRl3X/4QctVqS8AiZjFqyOGrMXb/w=="],
|
||||
|
||||
"ollama": ["ollama@0.6.3", "", { "dependencies": { "whatwg-fetch": "^3.6.20" } }, "sha512-KEWEhIqE5wtfzEIZbDCLH51VFZ6Z3ZSa6sIOg/E/tBV8S51flyqBOXi+bRxlOYKDf8i327zG9eSTb8IJxvm3Zg=="],
|
||||
"ollama-ai-provider": ["ollama-ai-provider@1.2.0", "", { "dependencies": { "@ai-sdk/provider": "^1.0.0", "@ai-sdk/provider-utils": "^2.0.0", "partial-json": "0.1.7" }, "peerDependencies": { "zod": "^3.0.0" }, "optionalPeers": ["zod"] }, "sha512-jTNFruwe3O/ruJeppI/quoOUxG7NA6blG3ZyQj3lei4+NnJo7bi3eIRWqlVpRlu/mbzbFXeJSBuYQWF6pzGKww=="],
|
||||
|
||||
"oxfmt": ["oxfmt@0.40.0", "", { "dependencies": { "tinypool": "2.1.0" }, "optionalDependencies": { "@oxfmt/binding-android-arm-eabi": "0.40.0", "@oxfmt/binding-android-arm64": "0.40.0", "@oxfmt/binding-darwin-arm64": "0.40.0", "@oxfmt/binding-darwin-x64": "0.40.0", "@oxfmt/binding-freebsd-x64": "0.40.0", "@oxfmt/binding-linux-arm-gnueabihf": "0.40.0", "@oxfmt/binding-linux-arm-musleabihf": "0.40.0", "@oxfmt/binding-linux-arm64-gnu": "0.40.0", "@oxfmt/binding-linux-arm64-musl": "0.40.0", "@oxfmt/binding-linux-ppc64-gnu": "0.40.0", "@oxfmt/binding-linux-riscv64-gnu": "0.40.0", "@oxfmt/binding-linux-riscv64-musl": "0.40.0", "@oxfmt/binding-linux-s390x-gnu": "0.40.0", "@oxfmt/binding-linux-x64-gnu": "0.40.0", "@oxfmt/binding-linux-x64-musl": "0.40.0", "@oxfmt/binding-openharmony-arm64": "0.40.0", "@oxfmt/binding-win32-arm64-msvc": "0.40.0", "@oxfmt/binding-win32-ia32-msvc": "0.40.0", "@oxfmt/binding-win32-x64-msvc": "0.40.0" }, "bin": { "oxfmt": "bin/oxfmt" } }, "sha512-g0C3I7xUj4b4DcagevM9kgH6+pUHytikxUcn3/VUkvzTNaaXBeyZqb7IBsHwojeXm4mTBEC/aBjBTMVUkZwWUQ=="],
|
||||
|
||||
@@ -201,16 +201,22 @@
|
||||
|
||||
"oxlint-tsgolint": ["oxlint-tsgolint@0.16.0", "", { "optionalDependencies": { "@oxlint-tsgolint/darwin-arm64": "0.16.0", "@oxlint-tsgolint/darwin-x64": "0.16.0", "@oxlint-tsgolint/linux-arm64": "0.16.0", "@oxlint-tsgolint/linux-x64": "0.16.0", "@oxlint-tsgolint/win32-arm64": "0.16.0", "@oxlint-tsgolint/win32-x64": "0.16.0" }, "bin": { "tsgolint": "bin/tsgolint.js" } }, "sha512-4RuJK2jP08XwqtUu+5yhCbxEauCm6tv2MFHKEMsjbosK2+vy5us82oI3VLuHwbNyZG7ekZA26U2LLHnGR4frIA=="],
|
||||
|
||||
"partial-json": ["partial-json@0.1.7", "", {}, "sha512-Njv/59hHaokb/hRUjce3Hdv12wd60MtM9Z5Olmn+nehe0QDAsRtRbJPvJ0Z91TusF0SuZRIvnM+S4l6EIP8leA=="],
|
||||
|
||||
"picocolors": ["picocolors@1.1.1", "", {}, "sha512-xceH2snhtb5M9liqDsmEw56le376mTZkEX/jEb/RxNFyegNul7eNslCXP9FDj/Lcu0X8KEyMceP2ntpaHrDEVA=="],
|
||||
|
||||
"secure-json-parse": ["secure-json-parse@2.7.0", "", {}, "sha512-6aU+Rwsezw7VR8/nyvKTx8QpWH9FrcYiXXlqC4z5d5XQBDRqtbfsRjnwGyqbi3gddNtWHuEk9OANUotL26qKUw=="],
|
||||
|
||||
"tinypool": ["tinypool@2.1.0", "", {}, "sha512-Pugqs6M0m7Lv1I7FtxN4aoyToKg1C4tu+/381vH35y8oENM/Ai7f7C4StcoK4/+BSw9ebcS8jRiVrORFKCALLw=="],
|
||||
|
||||
"typescript": ["typescript@5.9.3", "", { "bin": { "tsc": "bin/tsc", "tsserver": "bin/tsserver" } }, "sha512-jl1vZzPDinLr9eUt3J/t7V6FgNEw9QjvBPdysz9KfQDD41fQrC2Y4vKQdiaUpFT4bXlb1RHhLpp8wtm6M5TgSw=="],
|
||||
|
||||
"undici-types": ["undici-types@7.18.2", "", {}, "sha512-AsuCzffGHJybSaRrmr5eHr81mwJU3kjw6M+uprWvCXiNeN9SOGwQ3Jn8jb8m3Z6izVgknn1R0FTCEAP2QrLY/w=="],
|
||||
|
||||
"whatwg-fetch": ["whatwg-fetch@3.6.20", "", {}, "sha512-EqhiFU6daOA8kpjOWTL0olhVOF3i7OrFzSYiGsEMB8GcXS+RrzauAERX65xMeNWVqxA6HXH2m69Z9LaKKdisfg=="],
|
||||
|
||||
"zod": ["zod@4.3.6", "", {}, "sha512-rftlrkhHZOcjDwkGlnUtZZkvaPHCsDATp4pGpuOOMDaTdDDXF91wuVDJoWoPsKX/3YPQ5fHuF3STjcYyKr+Qhg=="],
|
||||
|
||||
"ollama-ai-provider/@ai-sdk/provider": ["@ai-sdk/provider@1.1.3", "", { "dependencies": { "json-schema": "^0.4.0" } }, "sha512-qZMxYJ0qqX/RfnuIaab+zp8UAeJn/ygXXAffR5I4N0n1IrvA6qBsjc8hXLmBiMV2zoXlifkacF7sEFnYnjBcqg=="],
|
||||
|
||||
"ollama-ai-provider/@ai-sdk/provider-utils": ["@ai-sdk/provider-utils@2.2.8", "", { "dependencies": { "@ai-sdk/provider": "1.1.3", "nanoid": "^3.3.8", "secure-json-parse": "^2.7.0" }, "peerDependencies": { "zod": "^3.23.8" } }, "sha512-fqhG+4sCVv8x7nFzYnFo19ryhAa3w096Kmc3hWxMQfW/TubPOmt3A6tYZhl4mUfQWWQMsuSkLrtjlWuXBVSGQA=="],
|
||||
}
|
||||
}
|
||||
|
||||
@@ -123,7 +123,7 @@ Wraps Vercel AI SDK `generateText()` with:
|
||||
- Normalized `LLMResponse` type
|
||||
|
||||
### SessionManager
|
||||
Persists conversation history to JSONL files in `~/.config/nanobot/sessions/`.
|
||||
Persists conversation history to JSONL files in `~/.nanobot/sessions/`.
|
||||
|
||||
- Key format: `{channel}:{chatId}` (e.g., `mattermost:abc123`)
|
||||
- Supports history truncation for context window limits
|
||||
@@ -136,7 +136,7 @@ When session history exceeds token limits, summarizes old messages and archives
|
||||
|
||||
## Configuration
|
||||
|
||||
- File: `~/.config/nanobot/config.json`
|
||||
- File: `~/.nanobot/config.json`
|
||||
- Validation: Zod schemas in `src/config/types.ts`
|
||||
- Env overrides: `NANOBOT_MODEL`, `NANOBOT_WORKSPACE`, `NANOBOT_CONFIG`
|
||||
|
||||
|
||||
@@ -129,7 +129,7 @@ const timeout = parseInt(strArg(args, 'timeout', '30'), 10);
|
||||
## Session Persistence
|
||||
|
||||
- Format: JSONL (one JSON object per line)
|
||||
- Location: `~/.config/nanobot/sessions/{sessionKey}.jsonl`
|
||||
- Location: `~/.nanobot/sessions/{sessionKey}.jsonl`
|
||||
- Tool results truncated at 16,000 characters
|
||||
- Memory consolidation triggered when approaching context window limit
|
||||
|
||||
@@ -147,5 +147,5 @@ Max 3 attempts with exponential backoff.
|
||||
|
||||
1. CLI flags (`-c`, `-m`, `-w`, `-M`)
|
||||
2. Environment variables (`NANOBOT_CONFIG`, `NANOBOT_MODEL`, `NANOBOT_WORKSPACE`)
|
||||
3. Config file (`~/.config/nanobot/config.json`)
|
||||
3. Config file (`~/.nanobot/config.json`)
|
||||
4. Zod schema defaults
|
||||
|
||||
@@ -19,22 +19,14 @@ Docs directory created with 4 files (PRD.md, Architecture.md, API.md, Discoverie
|
||||
- **lint**: all `${err}` in template literals → `${String(err)}`; `String(args['key'] ?? '')` → `strArg(args, 'key')` helper; unused `onProgress` param → `_onProgress`; WebSocket `onerror` `err` type is `Event` → use `err.type`
|
||||
|
||||
## Work Queue (next steps)
|
||||
1. [x] Create workspace helper module (src/cli/utils.ts) with ensureWorkspace() and syncTemplates()
|
||||
2. [x] Create onboard command (src/cli/onboard.ts) with path argument and directory-not-empty guard
|
||||
3. [x] Agent/gateway commands check workspace exists (throw if not found)
|
||||
4. [x] Added required `provider` field to agent config (values: anthropic, openai, google, openrouter, ollama)
|
||||
5. [x] Provider resolution uses explicit provider from config (no model prefix parsing)
|
||||
6. [x] Typecheck and lint pass (0 errors)
|
||||
7. [x] Test onboard and agent commands work correctly
|
||||
8. [x] Updated Ollama provider from `ollama-ai-provider` to `ai-sdk-ollama`
|
||||
9. [ ] Test with a real Mattermost config (optional — user can do this)
|
||||
1. [ ] Runtime smoke test: `bun run start --help`
|
||||
2. [ ] Test with a real Mattermost config (optional — user can do this)
|
||||
3. [ ] Write sample `~/.nanobot/config.json` in README or docs
|
||||
|
||||
## Key Decisions Made
|
||||
- Mattermost channel uses raw WebSocket + fetch (no mattermostdriver, no SSL hack)
|
||||
- No MCP support (use shell tools / CLI instead)
|
||||
- No reasoning/thinking token handling (can add later)
|
||||
- Config is fresh Zod schema (no migration from Python config needed)
|
||||
- `ai-sdk-ollama` package for Ollama provider (replaced old `ollama-ai-provider`)
|
||||
- `ollama-ai-provider` package (not `@ai-sdk/ollama` which 404s on npm)
|
||||
- `strArg(args, key, fallback?)` helper exported from `agent/tools/base.ts` for safe unknown→string extraction
|
||||
- Agent config requires explicit `provider` field (no more model prefix like "anthropic/claude-...")
|
||||
- Model names are now just the raw model ID (e.g., "claude-sonnet-4-5" not "anthropic/claude-sonnet-4-5")
|
||||
|
||||
@@ -12,7 +12,7 @@ A personal AI assistant that connects to Mattermost (via WebSocket) and runs an
|
||||
## Key design principles (from Python codebase)
|
||||
- Ultra-lightweight: minimal dependencies, small codebase
|
||||
- Provider-agnostic: works with Anthropic, OpenAI, Google, Ollama, OpenRouter
|
||||
- Workspace-centric: everything lives in a configurable workspace directory (`~/.config/nanobot/`)
|
||||
- Workspace-centric: everything lives in a configurable workspace directory (`~/.nanobot/`)
|
||||
- SOUL/AGENTS/USER/TOOLS.md: workspace markdown files that define the bot's personality and rules
|
||||
- Memory is just markdown files (`MEMORY.md`, `HISTORY.md`) — no database
|
||||
|
||||
|
||||
@@ -8,21 +8,41 @@
|
||||
- All dependencies installed
|
||||
- `src/` directory structure scaffolded
|
||||
- Memory bank initialized
|
||||
- All source files written (first pass)
|
||||
- All source files written (first pass):
|
||||
- `src/config/types.ts` + `src/config/loader.ts`
|
||||
- `src/bus/types.ts` + `src/bus/queue.ts`
|
||||
- `src/provider/types.ts` + `src/provider/index.ts`
|
||||
- `src/session/types.ts` + `src/session/manager.ts`
|
||||
- `src/agent/tools/base.ts` (+ `strArg` helper)
|
||||
- `src/agent/tools/filesystem.ts`
|
||||
- `src/agent/tools/shell.ts`
|
||||
- `src/agent/tools/web.ts`
|
||||
- `src/agent/tools/message.ts`
|
||||
- `src/agent/tools/spawn.ts` + `src/agent/subagent.ts`
|
||||
- `src/agent/tools/cron.ts`
|
||||
- `src/cron/types.ts` + `src/cron/service.ts`
|
||||
- `src/heartbeat/service.ts`
|
||||
- `src/agent/memory.ts`
|
||||
- `src/agent/skills.ts`
|
||||
- `src/agent/context.ts`
|
||||
- `src/agent/loop.ts`
|
||||
- `src/channels/base.ts` + `src/channels/mattermost.ts`
|
||||
- `src/channels/manager.ts`
|
||||
- `src/cli/commands.ts`
|
||||
- `index.ts`
|
||||
- Templates and skills copied from Python repo
|
||||
- **Full typecheck pass**: `tsc --noEmit` → 0 errors
|
||||
- **Full lint pass**: `oxlint` → 0 errors, 0 warnings
|
||||
- `package.json` scripts added: `start`, `dev`, `typecheck`
|
||||
- **Docs created**: `/docs/PRD.md`, `Architecture.md`, `API.md`, `Discoveries.md`
|
||||
- **Onboard command**: Created `src/cli/onboard.ts` with workspace initialization
|
||||
- **Provider config**: Added required `provider` field to agent config
|
||||
- **Workspace validation**: Agent/gateway commands throw if workspace doesn't exist
|
||||
|
||||
### 🔄 In Progress
|
||||
- Nothing
|
||||
|
||||
### ⏳ Pending
|
||||
- Runtime smoke test: `bun run start --help`
|
||||
- Integration test with a real Mattermost server
|
||||
- Sample `~/.nanobot/config.json` documentation
|
||||
|
||||
## Known Issues / Risks
|
||||
- `ollama-ai-provider` v1.2.0 returns `LanguageModelV1` (not V2/V3 as expected by AI SDK v6) — cast used at call site. Works at runtime.
|
||||
|
||||
@@ -34,7 +34,7 @@ Inbound and outbound messages are passed through a typed `AsyncQueue<T>`. The qu
|
||||
- Returns a normalized `LLMResponse` type
|
||||
|
||||
## Config Pattern
|
||||
- Config file: `~/.config/nanobot/config.json` (camelCase JSON)
|
||||
- Config file: `~/.nanobot/config.json` (camelCase JSON)
|
||||
- Loaded with `loadConfig()`, validated by Zod, returns inferred `Config` type
|
||||
- `NANOBOT_` env vars can override fields (e.g. `NANOBOT_MODEL`)
|
||||
|
||||
@@ -49,28 +49,6 @@ Inbound and outbound messages are passed through a typed `AsyncQueue<T>`. The qu
|
||||
## Logging Pattern
|
||||
Use `console.error` / `console.warn` / `console.info` / `console.debug` — no external logger. Color via `picocolors` in CLI output only.
|
||||
|
||||
## CLI Command Pattern
|
||||
Each command is in its own file with a registration function:
|
||||
```ts
|
||||
// src/cli/agent.ts
|
||||
export function agentCommand(program: Command, config: Config, workspace: string): void {
|
||||
program.command('agent')
|
||||
.description('...')
|
||||
.option('-m, --message <text>', 'Single message to process')
|
||||
.action(async (opts) => { /* ... */ })
|
||||
}
|
||||
|
||||
// src/cli/commands.ts (bootstrap)
|
||||
export function createCli(): Command {
|
||||
const program = new Command('nanobot')...
|
||||
const config = loadConfig(opts.config);
|
||||
const workspace = resolveWorkspacePath(config.agent.workspacePath);
|
||||
gatewayCommand(program, config, workspace);
|
||||
agentCommand(program, config, workspace);
|
||||
return program;
|
||||
}
|
||||
```
|
||||
|
||||
## File Layout
|
||||
```
|
||||
src/
|
||||
@@ -89,12 +67,8 @@ src/
|
||||
tools/base.ts + filesystem.ts + shell.ts + web.ts + message.ts + spawn.ts + cron.ts
|
||||
channels/
|
||||
base.ts + mattermost.ts + manager.ts
|
||||
cli/
|
||||
types.ts # CommandHandler type
|
||||
commands.ts # Bootstrap - loads config, registers commands
|
||||
agent.ts # agentCommand() - interactive/single-shot mode
|
||||
gateway.ts # gatewayCommand() - full runtime with Mattermost
|
||||
index.ts
|
||||
cli/commands.ts
|
||||
index.ts
|
||||
templates/ (SOUL.md, AGENTS.md, USER.md, TOOLS.md, HEARTBEAT.md, memory/MEMORY.md)
|
||||
skills/ (copied from Python repo)
|
||||
```
|
||||
|
||||
@@ -4,14 +4,13 @@
|
||||
"type": "module",
|
||||
"module": "index.ts",
|
||||
"scripts": {
|
||||
"nanobot": "bun run index.ts",
|
||||
"start": "bun run index.ts",
|
||||
"dev": "bun --watch run index.ts",
|
||||
"typecheck": "tsc --noEmit",
|
||||
"fmt": "oxfmt --check",
|
||||
"fmt:fix": "oxfmt",
|
||||
"lint": "oxlint",
|
||||
"lint:fix": "oxlint --fix",
|
||||
"checks": "bun run lint && bun run fmt"
|
||||
"lint:fix": "oxlint --fix"
|
||||
},
|
||||
"dependencies": {
|
||||
"@ai-sdk/anthropic": "^3.0.58",
|
||||
@@ -20,12 +19,12 @@
|
||||
"@mozilla/readability": "^0.6.0",
|
||||
"@openrouter/ai-sdk-provider": "^2.3.0",
|
||||
"ai": "^6.0.116",
|
||||
"ai-sdk-ollama": "^3.8.0",
|
||||
"commander": "^14.0.3",
|
||||
"cron-parser": "^5.5.0",
|
||||
"js-tiktoken": "^1.0.21",
|
||||
"jsonrepair": "^3.13.3",
|
||||
"node-html-parser": "^7.1.0",
|
||||
"ollama-ai-provider": "^1.2.0",
|
||||
"picocolors": "^1.1.1",
|
||||
"zod": "^4.3.6"
|
||||
},
|
||||
|
||||
@@ -27,21 +27,21 @@ npx --yes clawhub@latest search "web scraping" --limit 5
|
||||
## Install
|
||||
|
||||
```bash
|
||||
npx --yes clawhub@latest install <slug> --workdir ~/.config/nanobot/workspace
|
||||
npx --yes clawhub@latest install <slug> --workdir ~/.nanobot/workspace
|
||||
```
|
||||
|
||||
Replace `<slug>` with the skill name from search results. This places the skill into `~/.config/nanobot/workspace/skills/`, where nanobot loads workspace skills from. Always include `--workdir`.
|
||||
Replace `<slug>` with the skill name from search results. This places the skill into `~/.nanobot/workspace/skills/`, where nanobot loads workspace skills from. Always include `--workdir`.
|
||||
|
||||
## Update
|
||||
|
||||
```bash
|
||||
npx --yes clawhub@latest update --all --workdir ~/.config/nanobot/workspace
|
||||
npx --yes clawhub@latest update --all --workdir ~/.nanobot/workspace
|
||||
```
|
||||
|
||||
## List installed
|
||||
|
||||
```bash
|
||||
npx --yes clawhub@latest list --workdir ~/.config/nanobot/workspace
|
||||
npx --yes clawhub@latest list --workdir ~/.nanobot/workspace
|
||||
```
|
||||
|
||||
## Notes
|
||||
@@ -49,5 +49,5 @@ npx --yes clawhub@latest list --workdir ~/.config/nanobot/workspace
|
||||
- Requires Node.js (`npx` comes with it).
|
||||
- No API key needed for search and install.
|
||||
- Login (`npx --yes clawhub@latest login`) is only required for publishing.
|
||||
- `--workdir ~/.config/nanobot/workspace` is critical — without it, skills install to the current directory instead of the nanobot workspace.
|
||||
- `--workdir ~/.nanobot/workspace` is critical — without it, skills install to the current directory instead of the nanobot workspace.
|
||||
- After install, remind the user to start a new session to load the skill.
|
||||
|
||||
@@ -364,13 +364,13 @@ export class AgentLoop {
|
||||
if (response.content) await onProgress(response.content);
|
||||
const hint = response.toolCalls
|
||||
.map((tc) => {
|
||||
let display = '';
|
||||
|
||||
const firstVal = Object.values(tc.arguments)[0];
|
||||
if (typeof firstVal === 'string') {
|
||||
display = `"${firstVal.slice(0, 40) + (firstVal.length > 40 ? '…' : '')}"`;
|
||||
}
|
||||
|
||||
const display =
|
||||
typeof firstVal === 'string'
|
||||
? firstVal.length > 40
|
||||
? `"${firstVal.slice(0, 40)}…"`
|
||||
: `"${firstVal}"`
|
||||
: '';
|
||||
return `${tc.name}(${display})`;
|
||||
})
|
||||
.join(', ');
|
||||
|
||||
150
src/cli/agent.ts
150
src/cli/agent.ts
@@ -1,94 +1,92 @@
|
||||
import { mkdirSync } from 'node:fs';
|
||||
import { createInterface } from 'node:readline';
|
||||
import { Command } from 'commander';
|
||||
import pc from 'picocolors';
|
||||
import { AgentLoop } from '../agent/loop.ts';
|
||||
import { MessageBus } from '../bus/queue.ts';
|
||||
import type { Config } from '../config/types.ts';
|
||||
import { makeProvider } from '../provider/index.ts';
|
||||
import { loadConfig } from '../config/loader.ts';
|
||||
import { ensureWorkspace } from './utils.ts';
|
||||
|
||||
export function agentCommand(program: Command): void {
|
||||
export function agentCommand(program: Command, config: Config, workspace: string): void {
|
||||
mkdirSync(workspace, { recursive: true });
|
||||
|
||||
program
|
||||
.command('agent')
|
||||
.description('Run the agent interactively or send a single message.')
|
||||
.option('-c, --config <path>', 'Path to config.json')
|
||||
.option('-m, --message <text>', 'Single message to process (non-interactive)')
|
||||
.option('-w, --workspace <path>', 'Workspace path override')
|
||||
.option('-M, --model <model>', 'Model override')
|
||||
.action(async (opts: { config?: string; message?: string; model?: string }) => {
|
||||
const config = loadConfig(opts.config);
|
||||
const workspace = config.agent.workspacePath;
|
||||
ensureWorkspace(workspace);
|
||||
.action(
|
||||
async (opts: { config?: string; message?: string; workspace?: string; model?: string }) => {
|
||||
const model = opts.model ?? config.agent.model;
|
||||
const provider = makeProvider(
|
||||
config.providers,
|
||||
model,
|
||||
config.agent.maxTokens,
|
||||
config.agent.temperature,
|
||||
);
|
||||
const bus = new MessageBus();
|
||||
|
||||
console.info(pc.magenta(`workspace path: ${workspace}`));
|
||||
|
||||
const model = opts.model ?? config.agent.model;
|
||||
const provider = makeProvider(
|
||||
config.providers,
|
||||
config.agent.provider,
|
||||
model,
|
||||
config.agent.maxTokens,
|
||||
config.agent.temperature,
|
||||
);
|
||||
const bus = new MessageBus();
|
||||
|
||||
const agentLoop = new AgentLoop({
|
||||
bus,
|
||||
provider,
|
||||
workspace,
|
||||
model,
|
||||
maxIterations: config.agent.maxToolIterations,
|
||||
contextWindowTokens: config.agent.contextWindowTokens,
|
||||
braveApiKey: config.tools.web.braveApiKey,
|
||||
webProxy: config.tools.web.proxy,
|
||||
execConfig: config.tools.exec,
|
||||
restrictToWorkspace: config.tools.restrictToWorkspace,
|
||||
});
|
||||
|
||||
// Single-shot mode
|
||||
if (opts.message) {
|
||||
const result = await agentLoop.processDirect(opts.message);
|
||||
console.log(result);
|
||||
return;
|
||||
}
|
||||
|
||||
// Interactive mode
|
||||
console.info(pc.green('nanobot interactive mode. Type your message, Ctrl+C to exit.'));
|
||||
|
||||
const rl = createInterface({ input: process.stdin, output: process.stdout });
|
||||
|
||||
const promptUser = () => {
|
||||
rl.question(pc.cyan('You: '), async (input) => {
|
||||
const text = input.trim();
|
||||
if (!text) {
|
||||
promptUser();
|
||||
return;
|
||||
}
|
||||
|
||||
const onProgress = async (content: string, opts?: { toolHint?: boolean }) => {
|
||||
if (opts?.toolHint) {
|
||||
process.stdout.write(pc.dim(` [${content}]\n`));
|
||||
} else {
|
||||
process.stdout.write(pc.dim(` ${content}\n`));
|
||||
}
|
||||
};
|
||||
|
||||
const result = await agentLoop.processDirect(
|
||||
text,
|
||||
'cli:interactive',
|
||||
'cli',
|
||||
'interactive',
|
||||
onProgress,
|
||||
);
|
||||
console.log(pc.bold('Bot:'), result);
|
||||
promptUser();
|
||||
const agentLoop = new AgentLoop({
|
||||
bus,
|
||||
provider,
|
||||
workspace,
|
||||
model,
|
||||
maxIterations: config.agent.maxToolIterations,
|
||||
contextWindowTokens: config.agent.contextWindowTokens,
|
||||
braveApiKey: config.tools.web.braveApiKey,
|
||||
webProxy: config.tools.web.proxy,
|
||||
execConfig: config.tools.exec,
|
||||
restrictToWorkspace: config.tools.restrictToWorkspace,
|
||||
});
|
||||
};
|
||||
|
||||
rl.on('close', () => {
|
||||
agentLoop.stop();
|
||||
process.exit(0);
|
||||
});
|
||||
// Single-shot mode
|
||||
if (opts.message) {
|
||||
const result = await agentLoop.processDirect(opts.message);
|
||||
console.log(result);
|
||||
return;
|
||||
}
|
||||
|
||||
promptUser();
|
||||
});
|
||||
// Interactive mode
|
||||
console.info(pc.green('nanobot interactive mode. Type your message, Ctrl+C to exit.'));
|
||||
|
||||
const rl = createInterface({ input: process.stdin, output: process.stdout });
|
||||
|
||||
const promptUser = () => {
|
||||
rl.question(pc.cyan('You: '), async (input) => {
|
||||
const text = input.trim();
|
||||
if (!text) {
|
||||
promptUser();
|
||||
return;
|
||||
}
|
||||
|
||||
const onProgress = async (content: string, opts?: { toolHint?: boolean }) => {
|
||||
if (opts?.toolHint) {
|
||||
process.stdout.write(pc.dim(` [${content}]\n`));
|
||||
} else {
|
||||
process.stdout.write(pc.dim(` ${content}\n`));
|
||||
}
|
||||
};
|
||||
|
||||
const result = await agentLoop.processDirect(
|
||||
text,
|
||||
'cli:interactive',
|
||||
'cli',
|
||||
'interactive',
|
||||
onProgress,
|
||||
);
|
||||
console.log(pc.bold('Bot:'), result);
|
||||
promptUser();
|
||||
});
|
||||
};
|
||||
|
||||
rl.on('close', () => {
|
||||
agentLoop.stop();
|
||||
process.exit(0);
|
||||
});
|
||||
|
||||
promptUser();
|
||||
},
|
||||
);
|
||||
}
|
||||
|
||||
@@ -1,16 +1,21 @@
|
||||
import { mkdirSync } from 'node:fs';
|
||||
import { Command } from 'commander';
|
||||
import { loadConfig, resolveWorkspacePath } from '../config/loader.ts';
|
||||
import { agentCommand } from './agent.ts';
|
||||
import { gatewayCommand } from './gateway.ts';
|
||||
import { onboardCommand } from './onboard.ts';
|
||||
|
||||
export function createCli(): Command {
|
||||
const program = new Command('nanobot')
|
||||
.description('nanobot — personal AI assistant')
|
||||
.version('1.0.0');
|
||||
|
||||
onboardCommand(program);
|
||||
gatewayCommand(program);
|
||||
agentCommand(program);
|
||||
const globalOpts = program.opts();
|
||||
const config = loadConfig(globalOpts.config);
|
||||
const workspace = resolveWorkspacePath(config.agent.workspacePath);
|
||||
mkdirSync(workspace, { recursive: true });
|
||||
|
||||
gatewayCommand(program, config, workspace);
|
||||
agentCommand(program, config, workspace);
|
||||
|
||||
return program;
|
||||
}
|
||||
|
||||
@@ -1,30 +1,25 @@
|
||||
import { ChannelManager } from '../channels/manager.ts';
|
||||
import { mkdirSync } from 'node:fs';
|
||||
import { Command } from 'commander';
|
||||
import pc from 'picocolors';
|
||||
import { AgentLoop } from '../agent/loop.ts';
|
||||
import { MessageBus } from '../bus/queue.ts';
|
||||
import { MattermostChannel } from '../channels/mattermost.ts';
|
||||
import { ChannelManager } from '../channels/manager.ts';
|
||||
import type { Config } from '../config/types.ts';
|
||||
import { CronService } from '../cron/service.ts';
|
||||
import { HeartbeatService } from '../heartbeat/service.ts';
|
||||
import { makeProvider } from '../provider/index.ts';
|
||||
import { loadConfig } from '../config/loader.ts';
|
||||
import { ensureWorkspace } from './utils.ts';
|
||||
|
||||
export function gatewayCommand(program: Command): void {
|
||||
export function gatewayCommand(program: Command, config: Config, workspace: string): void {
|
||||
mkdirSync(workspace, { recursive: true });
|
||||
|
||||
program
|
||||
.command('gateway')
|
||||
.option('-c, --config <path>', 'Path to config.json')
|
||||
.description('Start the full gateway: Mattermost channel, agent loop, cron, and heartbeat.')
|
||||
.action(async (opts: { config?: string }) => {
|
||||
const config = loadConfig(opts.config);
|
||||
const workspace = config.agent.workspacePath;
|
||||
ensureWorkspace(workspace);
|
||||
|
||||
console.info(pc.magenta(`workspace path: ${workspace}`));
|
||||
|
||||
.option('-c, --config <path>', 'Path to config.json')
|
||||
.action(async (_opts: { config?: string }) => {
|
||||
const provider = makeProvider(
|
||||
config.providers,
|
||||
config.agent.provider,
|
||||
config.agent.model,
|
||||
config.agent.maxTokens,
|
||||
config.agent.temperature,
|
||||
|
||||
@@ -1,69 +0,0 @@
|
||||
import { writeFileSync } from 'node:fs';
|
||||
import { join } from 'node:path';
|
||||
import { Command } from 'commander';
|
||||
import pc from 'picocolors';
|
||||
import { WORKSPACE_PATH } from '../config/constants.ts';
|
||||
import { ensureWorkspace, resolvePath, checkWorkspaceEmpty, syncTemplates } from './utils.ts';
|
||||
|
||||
function logCreated(item: string) {
|
||||
console.info(pc.green(` ✓ Created ${item}`));
|
||||
}
|
||||
|
||||
export function onboardCommand(program: Command): void {
|
||||
program
|
||||
.command('onboard [path]')
|
||||
.description('Initialize a new nanobot workspace with config and templates')
|
||||
.action(async (rawPath?: string) => {
|
||||
try {
|
||||
// Create a minimal config template - users must fill in provider and model
|
||||
const defaultConfig = {
|
||||
providers: {},
|
||||
agent: {
|
||||
provider: '',
|
||||
model: '',
|
||||
},
|
||||
};
|
||||
|
||||
const targetPath = resolvePath(rawPath ?? WORKSPACE_PATH);
|
||||
const configPath = join(targetPath, 'config.json');
|
||||
|
||||
console.info(pc.blue('Initializing nanobot workspace...'));
|
||||
console.info(pc.dim(`Target path: ${targetPath}`));
|
||||
|
||||
// Check if directory exists and is not empty
|
||||
checkWorkspaceEmpty(targetPath);
|
||||
|
||||
// Create workspace directory
|
||||
ensureWorkspace(targetPath, true);
|
||||
logCreated('workspace directory');
|
||||
|
||||
// Write default config
|
||||
writeFileSync(configPath, JSON.stringify(defaultConfig, null, 2), 'utf8');
|
||||
logCreated('config.json');
|
||||
|
||||
// Sync templates
|
||||
const createdFiles = syncTemplates(targetPath);
|
||||
for (const file of createdFiles) {
|
||||
logCreated(file);
|
||||
}
|
||||
|
||||
console.info();
|
||||
console.info(pc.green('nanobot workspace initialized successfully!'));
|
||||
console.info();
|
||||
console.info(pc.bold('Next steps:'));
|
||||
console.info(` 1. Edit ${pc.cyan(configPath)} to add your API keys`);
|
||||
console.info(
|
||||
` 2. Customize ${pc.cyan(join(targetPath, 'USER.md'))} and ${pc.cyan(join(targetPath, 'SOUL.md'))} with your preferences`,
|
||||
);
|
||||
console.info(` 3. Start chatting: ${pc.cyan('bun run nanobot agent')}`);
|
||||
console.info();
|
||||
console.info(` -- For gateway mode:`);
|
||||
console.info(` 1. Edit ${pc.cyan(configPath)} to add your channel config (Mattermost)`);
|
||||
console.info(` 2. Connect your agent: ${pc.cyan('bun run nanobot gateway')}`);
|
||||
console.info();
|
||||
} catch (err) {
|
||||
console.error(pc.red(String(err)));
|
||||
process.exit(1);
|
||||
}
|
||||
});
|
||||
}
|
||||
@@ -1,89 +0,0 @@
|
||||
import { existsSync, mkdirSync, readdirSync, readFileSync, writeFileSync } from 'node:fs';
|
||||
import { dirname, join, resolve } from 'node:path';
|
||||
import { fileURLToPath } from 'node:url';
|
||||
import { homedir } from 'node:os';
|
||||
import pc from 'picocolors';
|
||||
|
||||
export function resolvePath(raw: string): string {
|
||||
if (raw.startsWith('~/') || raw === '~') {
|
||||
return resolve(homedir(), raw.slice(2));
|
||||
}
|
||||
return resolve(raw);
|
||||
}
|
||||
|
||||
export function ensureWorkspace(rawPath: string, createIfMissing = false): string {
|
||||
const path = resolvePath(rawPath);
|
||||
if (!existsSync(path)) {
|
||||
if (createIfMissing) {
|
||||
mkdirSync(path, { recursive: true });
|
||||
} else {
|
||||
console.error(
|
||||
pc.red(`Workspace does not exist: ${path}\nRun 'nanobot onboard' to initialize.`),
|
||||
);
|
||||
process.exit(1);
|
||||
}
|
||||
}
|
||||
return path;
|
||||
}
|
||||
|
||||
export function syncTemplates(workspacePath: string): string[] {
|
||||
// Get project root relative to this file (src/cli/utils.ts)
|
||||
const currentFile = fileURLToPath(import.meta.url);
|
||||
const srcDir = dirname(currentFile);
|
||||
const projectRoot = resolve(srcDir, '..', '..');
|
||||
const templatesDir = resolve(projectRoot, 'templates');
|
||||
|
||||
if (!existsSync(templatesDir)) {
|
||||
throw new Error(`Templates directory not found at ${templatesDir}`);
|
||||
}
|
||||
|
||||
const created: string[] = [];
|
||||
|
||||
function copyTemplate(src: string, dest: string) {
|
||||
if (existsSync(dest)) return;
|
||||
mkdirSync(dirname(dest), { recursive: true });
|
||||
const content = readFileSync(src, 'utf8');
|
||||
writeFileSync(dest, content, 'utf8');
|
||||
created.push(dest.slice(workspacePath.length + 1));
|
||||
}
|
||||
|
||||
function copyDir(srcDir: string, destDir: string) {
|
||||
if (!existsSync(srcDir)) return;
|
||||
const entries = readdirSync(srcDir, { withFileTypes: true });
|
||||
for (const entry of entries) {
|
||||
const srcPath = join(srcDir, entry.name);
|
||||
const destPath = join(destDir, entry.name);
|
||||
if (entry.isDirectory()) {
|
||||
copyDir(srcPath, destPath);
|
||||
} else if (entry.name.endsWith('.md')) {
|
||||
copyTemplate(srcPath, destPath);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
copyDir(templatesDir, workspacePath);
|
||||
|
||||
// Create empty HISTORY.md
|
||||
const historyPath = join(workspacePath, 'memory', 'HISTORY.md');
|
||||
if (!existsSync(historyPath)) {
|
||||
mkdirSync(dirname(historyPath), { recursive: true });
|
||||
writeFileSync(historyPath, '# Conversation History\n\n', 'utf8');
|
||||
created.push('memory/HISTORY.md');
|
||||
}
|
||||
|
||||
// Create skills directory
|
||||
const skillsPath = join(workspacePath, 'skills');
|
||||
if (!existsSync(skillsPath)) {
|
||||
mkdirSync(skillsPath, { recursive: true });
|
||||
}
|
||||
|
||||
return created;
|
||||
}
|
||||
|
||||
export function checkWorkspaceEmpty(path: string): void {
|
||||
if (!existsSync(path)) return;
|
||||
const entries = readdirSync(path);
|
||||
if (entries.length > 0) {
|
||||
throw new Error(pc.red(`Directory not empty: ${path}`));
|
||||
}
|
||||
}
|
||||
@@ -1 +0,0 @@
|
||||
export const WORKSPACE_PATH = '~/.config/nanobot';
|
||||
@@ -1,10 +1,9 @@
|
||||
import { existsSync, mkdirSync, readFileSync, writeFileSync } from 'node:fs';
|
||||
import { homedir } from 'node:os';
|
||||
import { dirname, resolve } from 'node:path';
|
||||
import pc from 'picocolors';
|
||||
import { type Config, ConfigSchema } from './types.ts';
|
||||
|
||||
const DEFAULT_CONFIG_PATH = resolve(homedir(), '.config', 'nanobot', 'config.json');
|
||||
const DEFAULT_CONFIG_PATH = resolve(homedir(), '.nanobot', 'config.json');
|
||||
|
||||
export function getConfigPath(override?: string): string {
|
||||
return override ?? process.env['NANOBOT_CONFIG'] ?? DEFAULT_CONFIG_PATH;
|
||||
@@ -14,17 +13,16 @@ export function loadConfig(configPath?: string): Config {
|
||||
const path = getConfigPath(configPath);
|
||||
|
||||
if (!existsSync(path)) {
|
||||
console.error(pc.red(`Failed to load config from ${configPath}`));
|
||||
process.exit(1);
|
||||
return ConfigSchema.parse({});
|
||||
}
|
||||
|
||||
const raw = readFileSync(path, 'utf8');
|
||||
let json: unknown;
|
||||
try {
|
||||
json = JSON.parse(raw);
|
||||
} catch (error) {
|
||||
} catch {
|
||||
console.error(`Failed to parse config at ${path}`);
|
||||
throw error;
|
||||
return ConfigSchema.parse({});
|
||||
}
|
||||
|
||||
// Apply NANOBOT_ env var overrides before validation
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import { z } from 'zod';
|
||||
import { WORKSPACE_PATH } from './constants.ts';
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Mattermost
|
||||
@@ -40,19 +39,9 @@ export type ChannelsConfig = z.infer<typeof ChannelsConfigSchema>;
|
||||
// Agent
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
export const AgentProviderSchema = z.enum([
|
||||
'anthropic',
|
||||
'openai',
|
||||
'google',
|
||||
'openrouter',
|
||||
'ollama',
|
||||
]);
|
||||
export type AgentProvider = z.infer<typeof AgentProviderSchema>;
|
||||
|
||||
export const AgentConfigSchema = z.object({
|
||||
provider: AgentProviderSchema,
|
||||
model: z.string(),
|
||||
workspacePath: z.string().default(WORKSPACE_PATH),
|
||||
model: z.string().default('anthropic/claude-sonnet-4-5'),
|
||||
workspacePath: z.string().default('~/.nanobot'),
|
||||
maxTokens: z.number().int().default(4096),
|
||||
contextWindowTokens: z.number().int().default(65536),
|
||||
temperature: z.number().default(0.7),
|
||||
@@ -123,14 +112,21 @@ export type HeartbeatConfig = z.infer<typeof HeartbeatConfigSchema>;
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
export const ConfigSchema = z.object({
|
||||
agent: AgentConfigSchema.default(() => ({
|
||||
model: 'anthropic/claude-sonnet-4-5',
|
||||
workspacePath: '~/.nanobot',
|
||||
maxTokens: 4096,
|
||||
contextWindowTokens: 65536,
|
||||
temperature: 0.7,
|
||||
maxToolIterations: 40,
|
||||
})),
|
||||
providers: ProvidersConfigSchema.default(() => ({})),
|
||||
agent: AgentConfigSchema,
|
||||
heartbeat: HeartbeatConfigSchema.default(() => ({ enabled: false, intervalMinutes: 30 })),
|
||||
channels: ChannelsConfigSchema.default(() => ({ sendProgress: true, sendToolHints: true })),
|
||||
tools: ToolsConfigSchema.default(() => ({
|
||||
exec: { timeout: 120, denyPatterns: [], restrictToWorkspace: false },
|
||||
web: {},
|
||||
restrictToWorkspace: false,
|
||||
})),
|
||||
heartbeat: HeartbeatConfigSchema.default(() => ({ enabled: false, intervalMinutes: 30 })),
|
||||
});
|
||||
export type Config = z.infer<typeof ConfigSchema>;
|
||||
|
||||
@@ -4,8 +4,8 @@ import { createOpenAI } from '@ai-sdk/openai';
|
||||
import { createOpenRouter } from '@openrouter/ai-sdk-provider';
|
||||
import { type ModelMessage, generateText, stepCountIs } from 'ai';
|
||||
import { jsonrepair } from 'jsonrepair';
|
||||
import { createOllama } from 'ai-sdk-ollama';
|
||||
import type { AgentProvider, ProvidersConfig } from '../config/types.ts';
|
||||
import { createOllama } from 'ollama-ai-provider';
|
||||
import type { ProvidersConfig } from '../config/types.ts';
|
||||
import type { ChatOptions, LLMResponse, ToolDefinition } from './types.ts';
|
||||
|
||||
export type { ToolDefinition };
|
||||
@@ -66,20 +66,17 @@ import type { LanguageModel } from 'ai';
|
||||
|
||||
export class LLMProvider {
|
||||
private _providers: ProvidersConfig;
|
||||
private _provider: AgentProvider;
|
||||
private _defaultModel: string;
|
||||
private _maxTokens: number;
|
||||
private _temperature: number;
|
||||
|
||||
constructor(
|
||||
providers: ProvidersConfig,
|
||||
provider: AgentProvider,
|
||||
defaultModel: string,
|
||||
maxTokens = 4096,
|
||||
temperature = 0.7,
|
||||
) {
|
||||
this._providers = providers;
|
||||
this._provider = provider;
|
||||
this._defaultModel = defaultModel;
|
||||
this._maxTokens = maxTokens;
|
||||
this._temperature = temperature;
|
||||
@@ -90,26 +87,38 @@ export class LLMProvider {
|
||||
}
|
||||
|
||||
private _resolveModel(model: string): LanguageModel {
|
||||
switch (this._provider) {
|
||||
const slashIdx = model.indexOf('/');
|
||||
const prefix = slashIdx >= 0 ? model.slice(0, slashIdx) : model;
|
||||
const remainder = slashIdx >= 0 ? model.slice(slashIdx + 1) : model;
|
||||
|
||||
switch (prefix) {
|
||||
case 'anthropic': {
|
||||
const cfg = this._providers.anthropic;
|
||||
return createAnthropic({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
return createAnthropic({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(remainder);
|
||||
}
|
||||
case 'openai': {
|
||||
const cfg = this._providers.openai;
|
||||
return createOpenAI({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
return createOpenAI({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(remainder);
|
||||
}
|
||||
case 'google': {
|
||||
const cfg = this._providers.google;
|
||||
return createGoogleGenerativeAI({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
return createGoogleGenerativeAI({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(remainder);
|
||||
}
|
||||
case 'openrouter': {
|
||||
const cfg = this._providers.openrouter;
|
||||
return createOpenRouter({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
return createOpenRouter({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(remainder);
|
||||
}
|
||||
case 'ollama': {
|
||||
const cfg = this._providers.ollama;
|
||||
return createOllama({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
// ollama-ai-provider returns LanguageModelV1; cast to LanguageModel (compatible at runtime)
|
||||
return createOllama({ baseURL: cfg?.apiBase ?? 'http://localhost:11434/api' })(
|
||||
remainder,
|
||||
) as unknown as LanguageModel;
|
||||
}
|
||||
default: {
|
||||
// No recognized prefix — fall through to openai-compatible
|
||||
const cfg = this._providers.openai;
|
||||
return createOpenAI({ apiKey: cfg?.apiKey, baseURL: cfg?.apiBase })(model);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -137,15 +146,17 @@ export class LLMProvider {
|
||||
: undefined;
|
||||
|
||||
try {
|
||||
let toolChoice: 'required' | 'none' | 'auto' = 'auto';
|
||||
if (opts.toolChoice === 'required' || opts.toolChoice === 'none')
|
||||
toolChoice = opts.toolChoice;
|
||||
const result = await generateText({
|
||||
model,
|
||||
messages: opts.messages as ModelMessage[],
|
||||
// biome-ignore lint/suspicious/noExplicitAny: AI SDK tools type is complex
|
||||
tools: aiTools as any,
|
||||
toolChoice,
|
||||
toolChoice:
|
||||
opts.toolChoice === 'required'
|
||||
? 'required'
|
||||
: opts.toolChoice === 'none'
|
||||
? 'none'
|
||||
: 'auto',
|
||||
maxOutputTokens: maxTokens,
|
||||
temperature,
|
||||
stopWhen: stepCountIs(1),
|
||||
@@ -204,12 +215,11 @@ export class LLMProvider {
|
||||
|
||||
export function makeProvider(
|
||||
providers: ProvidersConfig,
|
||||
provider: AgentProvider,
|
||||
model: string,
|
||||
maxTokens: number,
|
||||
temperature: number,
|
||||
): LLMProvider {
|
||||
return new LLMProvider(providers, provider, model, maxTokens, temperature);
|
||||
return new LLMProvider(providers, model, maxTokens, temperature);
|
||||
}
|
||||
|
||||
/** Build a tool-result message to append after executing a tool call. */
|
||||
|
||||
Reference in New Issue
Block a user