916 B
916 B
Inference Service
Package: @nexusai/inference-service
Location: packages/inference-service
Deployed on: Main PC
Port: 3001
Purpose
Thin adapter layer around the local LLM runtime (Ollama). Receives assembled context packages from the orchestration service and returns model responses.
Dependencies
express— HTTP APIollama— Ollama clientdotenv— environment variable loading@nexusai/shared— shared utilities
Environment Variables
| Variable | Required | Default | Description |
|---|---|---|---|
| PORT | No | 3001 | Port to listen on |
| OLLAMA_URL | No | http://localhost:11434 | Ollama instance URL |
| DEFAULT_MODEL | No | llama3 | Default model to use for inference |
Endpoints
| Method | Path | Description |
|---|---|---|
| GET | /health | Service health check |
Further endpoints will be documented as the service is built out.