- Implemented Upstash Vector as a cloud-based storage solution for document chunks, replacing the local LanceDB option. - Added auto-detection of storage mode based on environment variables for seamless integration. - Updated the chat API to utilize the new retrieval mechanism, enhancing response accuracy and performance. - Enhanced README with setup instructions for Upstash and updated environment variable requirements. - Introduced new scripts and configurations for managing the vector index and API interactions.
Docs Chat
Docs chatbot that uses RAG (Retrieval-Augmented Generation) to answer questions from the OpenClaw documentation via semantic search.
Architecture
docs/**/*.md
│
▼
┌─────────────────┐
│ build-vector- │ Chunking + OpenAI Embeddings
│ index.ts │
└────────┬────────┘
│
▼
┌─────────────────┐
│ Vector Store │ Upstash (cloud) or LanceDB (local)
└────────┬────────┘
│
▼
┌─────────────────┐
│ API Server │ Hybrid Retrieval (Vector + Keyword Boost)
│ serve.ts │ → GPT-4o-mini Streaming Response
└─────────────────┘
Storage Backends
The pipeline supports two vector storage backends, auto-detected based on environment variables:
| Backend | When Used | Best For |
|---|---|---|
| LanceDB | Default (no Upstash credentials) | Local dev, POC, testing |
| Upstash | When UPSTASH_VECTOR_REST_* env vars are set |
Production, Vercel deploy |
Recommendation: For production deployments, use Upstash Vector for its serverless scalability and Vercel compatibility. LanceDB is great for local development and proof-of-concept work without external dependencies.
Quick Start (Local with LanceDB)
For local development without external services:
# Only OPENAI_API_KEY is required - uses LanceDB automatically
OPENAI_API_KEY=sk-... pnpm docs:chat:index:vector
OPENAI_API_KEY=sk-... pnpm docs:chat:serve:vector
The index is stored locally in scripts/docs-chat/.lancedb/.
Production Setup (Upstash Vector)
1. Create Upstash Vector Index
- Go to Upstash Console
- Create a new Vector index with:
- Dimensions: 3072 (for
text-embedding-3-large) - Distance Metric: Cosine
- Dimensions: 3072 (for
- Copy the REST URL and token
2. Environment Variables
| Variable | Required | Description |
|---|---|---|
OPENAI_API_KEY |
Yes | OpenAI API key for embeddings + chat |
UPSTASH_VECTOR_REST_URL |
No* | Upstash Vector REST endpoint |
UPSTASH_VECTOR_REST_TOKEN |
No* | Upstash Vector REST token |
* Required for Upstash mode; omit both for LanceDB mode.
3. Build the Vector Index
OPENAI_API_KEY=sk-... \
UPSTASH_VECTOR_REST_URL=https://... \
UPSTASH_VECTOR_REST_TOKEN=... \
pnpm docs:chat:index:vector
This generates embeddings for all doc chunks and upserts them to Upstash Vector.
4. Deploy to Vercel
cd scripts/docs-chat
npm install
vercel
Set the environment variables in the Vercel dashboard.
Local Development
Run the API locally
# With Upstash (cloud):
OPENAI_API_KEY=sk-... \
UPSTASH_VECTOR_REST_URL=https://... \
UPSTASH_VECTOR_REST_TOKEN=... \
pnpm docs:chat:serve:vector
# With LanceDB (local):
OPENAI_API_KEY=sk-... pnpm docs:chat:serve:vector
Defaults to http://localhost:3001. Optional environment variables:
| Variable | Default | Description |
|---|---|---|
PORT |
3001 |
Server port |
RATE_LIMIT |
20 |
Max requests per window per IP (Upstash only) |
RATE_WINDOW_MS |
60000 |
Rate limit window in milliseconds (Upstash only) |
Note: Rate limiting is only enforced in Upstash (production) mode. Local development with LanceDB has no rate limits.
Health check
curl http://localhost:3001/health
# Returns: {"ok":true,"chunks":N,"mode":"upstash"} # or "lancedb"
Mintlify Widget
Mintlify loads .js files from the docs content directory on every page.
docs/assets/docs-chat-config.js- Sets the API URLdocs/assets/docs-chat-widget.js- The chat widget
To configure the production API URL, edit docs/assets/docs-chat-config.js:
window.DOCS_CHAT_API_URL = "https://your-project.vercel.app";
API Endpoints
POST /chat
Send a message and receive a streaming response.
Request:
{ "message": "How do I configure the gateway?" }
Response: Streaming text/plain with the AI response.
GET /health
Check API health and vector count.
Response:
{ "ok": true, "chunks": 847, "mode": "upstash-vector" }
Legacy Pipelines
Keyword-based search
The keyword-based implementation is still available for backward compatibility:
pnpm docs:chat:index # Build keyword index
pnpm docs:chat:serve # Run keyword API
File Structure
scripts/docs-chat/
├── api/
│ ├── chat.ts # Vercel serverless function for chat
│ └── health.ts # Vercel serverless function for health check
├── rag/
│ ├── embeddings.ts # OpenAI embeddings wrapper
│ ├── retriever-factory.ts # Unified retriever (works with any store)
│ ├── retriever-upstash.ts # Legacy Upstash-specific retriever
│ ├── retriever.ts # Legacy LanceDB retriever
│ ├── store-factory.ts # Auto-selects Upstash or LanceDB
│ ├── store-upstash.ts # Upstash Vector store
│ └── store.ts # LanceDB store (local)
├── build-vector-index.ts # Index builder script
├── serve.ts # Local dev server
├── package.json # Standalone package for Vercel
├── tsconfig.json # TypeScript config
├── vercel.json # Vercel deployment config
└── README.md