Add initial project documentation
This commit is contained in:
35
docs/services/inference-service.md
Normal file
35
docs/services/inference-service.md
Normal file
@@ -0,0 +1,35 @@
|
||||
# Inference Service
|
||||
|
||||
**Package:** `@nexusai/inference-service`
|
||||
**Location:** `packages/inference-service`
|
||||
**Deployed on:** Main PC
|
||||
**Port:** 3001
|
||||
|
||||
## Purpose
|
||||
|
||||
Thin adapter layer around the local LLM runtime (Ollama). Receives
|
||||
assembled context packages from the orchestration service and returns
|
||||
model responses.
|
||||
|
||||
## Dependencies
|
||||
|
||||
- `express` — HTTP API
|
||||
- `ollama` — Ollama client
|
||||
- `dotenv` — environment variable loading
|
||||
- `@nexusai/shared` — shared utilities
|
||||
|
||||
## Environment Variables
|
||||
|
||||
| Variable | Required | Default | Description |
|
||||
|---|---|---|---|
|
||||
| PORT | No | 3001 | Port to listen on |
|
||||
| OLLAMA_URL | No | http://localhost:11434 | Ollama instance URL |
|
||||
| DEFAULT_MODEL | No | llama3 | Default model to use for inference |
|
||||
|
||||
## Endpoints
|
||||
|
||||
| Method | Path | Description |
|
||||
|---|---|---|
|
||||
| GET | /health | Service health check |
|
||||
|
||||
> Further endpoints will be documented as the service is built out.
|
||||
Reference in New Issue
Block a user