Nicolas Cantu cb87e283a1 Add AnythingLLM workspaces VS Code extension scaffold
**Motivations:**
- Expose AnythingLLM API workspaces from the editor against ia.enso public URL

**Root causes:**
- N/A (new capability)

**Correctifs:**
- N/A

**Evolutions:**
- Extension folder with list/open UI commands and API client
- Docs index and feature note

**Pages affectées:**
- extensions/anythingllm-workspaces/*
- docs/README.md
- docs/features/anythingllm-vscode-extension.md
2026-03-23 11:10:15 +01:00

39 lines
1.4 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

# AnythingLLM Workspaces (VS Code / Cursor extension)
Minimal extension to call the **AnythingLLM developer API** and open a workspace in the browser.
## Prerequisites
- AnythingLLM reachable at your public base URL (e.g. `https://ia.enso.4nkweb.com/anythingllm`).
- An **API key** created in AnythingLLM: **Settings → API Keys**.
## Configuration
| Setting | Description |
|--------|-------------|
| `anythingllm.baseUrl` | Base URL without trailing slash (default matches `deploy/nginx/README-ia-enso.md`). |
| `anythingllm.apiKey` | Bearer token for `GET /api/v1/workspaces`. Use **User** settings to avoid committing secrets. |
## Commands
- **AnythingLLM: List workspaces** — Fetches workspaces, then opens the selected one in the default browser (`/workspace/<slug>` under your base URL).
- **AnythingLLM: Open web UI** — Opens the AnythingLLM base URL.
## Ollama
This extension targets **AnythingLLM** only. For OpenAI-compatible Ollama behind the same proxy, use Cursors model settings with `https://ia.enso.4nkweb.com/ollama/v1` and the nginx Bearer (see `deploy/nginx/README-ia-enso.md`).
## Build
```bash
cd extensions/anythingllm-workspaces
npm install
npm run compile
```
Load the folder in VS Code / Cursor with **Run Extension** or install the packaged `.vsix` after `vsce package`.
## API reference
Upstream routes (mounted under `/api`): `GET /v1/workspaces` — see Mintplex-Labs anything-llm `server/endpoints/api/workspace/index.js`.