Switch Inference Models at Runtime#
Change the active inference model while the sandbox is running. No restart is required.
Prerequisites#
A running NemoClaw sandbox.
The OpenShell CLI on your
PATH.
Switch to a Different Model#
Set the provider to nvidia-nim and specify a model from build.nvidia.com:
$ openshell inference set --provider nvidia-nim --model nvidia/nemotron-3-super-120b-a12b
This requires the NVIDIA_API_KEY environment variable.
The nemoclaw onboard command stores this key in ~/.nemoclaw/credentials.json on first run.
Verify the Active Model#
Run the status command to confirm the change:
$ nemoclaw <name> status
Add the --json flag for machine-readable output:
$ nemoclaw <name> status --json
The output includes the active provider, model, and endpoint.
Available Models#
The following table lists the models registered with the nvidia-nim provider.
You can switch to any of these models at runtime.
Model ID |
Label |
Context Window |
Max Output |
|---|---|---|---|
|
Nemotron 3 Super 120B |
131,072 |
8,192 |
|
Nemotron Ultra 253B |
131,072 |
4,096 |
|
Nemotron Super 49B v1.5 |
131,072 |
4,096 |
|
Nemotron 3 Nano 30B |
131,072 |
4,096 |