Inference Options#
NemoClaw supports multiple inference providers.
During onboarding, the nemoclaw onboard wizard presents a numbered list of providers to choose from.
Your selection determines where the agent’s inference traffic is routed.
How Inference Routing Works#
The agent inside the sandbox talks to inference.local.
It never connects to a provider directly.
OpenShell intercepts inference traffic on the host and forwards it to the provider you selected.
Provider credentials stay on the host. The sandbox does not receive your API key.
Provider Options#
The onboard wizard presents the following provider options by default. The first six are always available. Ollama appears when it is installed or running on the host.
Option |
Description |
Curated models |
|---|---|---|
NVIDIA Endpoints |
Routes to models hosted on build.nvidia.com. You can also enter any model ID from the catalog. Set |
Nemotron 3 Super 120B, Kimi K2.5, GLM-5, MiniMax M2.5, GPT-OSS 120B |
OpenAI |
Routes to the OpenAI API. Set |
|
Other OpenAI-compatible endpoint |
Routes to any server that implements |
You provide the model name. |
Anthropic |
Routes to the Anthropic Messages API. Set |
|
Other Anthropic-compatible endpoint |
Routes to any server that implements the Anthropic Messages API ( |
You provide the model name. |
Google Gemini |
Routes to Google’s OpenAI-compatible endpoint. Set |
|
Local Ollama |
Routes to a local Ollama instance on |
Selected during onboarding. For more information, refer to Use a Local Inference Server. |
Experimental Options#
The following local inference options require NEMOCLAW_EXPERIMENTAL=1 and, when prerequisites are met, appear in the onboarding selection list.
Option |
Condition |
Notes |
|---|---|---|
Local NVIDIA NIM |
NIM-capable GPU detected |
Pulls and manages a NIM container. |
Local vLLM |
vLLM running on |
Auto-detects the loaded model. |
For setup instructions, refer to Use a Local Inference Server.
Validation#
NemoClaw validates the selected provider and model before creating the sandbox. If validation fails, the wizard returns to provider selection.
Provider type |
Validation method |
|---|---|
OpenAI-compatible |
Tries |
Anthropic-compatible |
Tries |
NVIDIA Endpoints (manual model entry) |
Validates the model name against the catalog API. |
Compatible endpoints |
Sends a real inference request because many proxies do not expose a |
Next Steps#
Use a Local Inference Server for Ollama, vLLM, NIM, and compatible-endpoint setup details.
Switch Inference Models for changing the model at runtime without re-onboarding.