You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
NemoClaw installs the NVIDIA OpenShell runtime and Nemotron models, then uses a versioned blueprint to create a sandboxed environment where every network request, file access, and inference call is governed by declarative policy. The `nemoclaw` CLI orchestrates the full stack: OpenShell gateway, sandbox, inference provider, and network policy.
165
+
NemoClaw installs the NVIDIA OpenShell runtime, then creates a sandboxed OpenClaw environment where every network request, file access, and inference call is governed by declarative policy. The `nemoclaw` CLI orchestrates the full stack: OpenShell gateway, sandbox, inference provider, and network policy.
|**Plugin**| TypeScript CLI commands for launch, connect, status, and logs. |
170
170
|**Blueprint**| Versioned Python artifact that orchestrates sandbox creation, policy, and inference setup. |
171
171
|**Sandbox**| Isolated OpenShell container running OpenClaw with policy-enforced egress and filesystem. |
172
-
|**Inference**|NVIDIA Endpoint model calls, routed through the OpenShell gateway, transparent to the agent.|
172
+
|**Inference**|Provider-routed model calls, routed through the OpenShell gateway, transparent to the agent. |
173
173
174
174
The blueprint lifecycle follows four stages: resolve the artifact, verify its digest, plan the resources, and apply through the OpenShell CLI.
175
175
@@ -179,15 +179,28 @@ When something goes wrong, errors may originate from either NemoClaw or the Open
179
179
180
180
## Inference
181
181
182
-
Inference requests from the agent never leave the sandbox directly. OpenShell intercepts every call and routes it to the NVIDIA Endpoint provider.
182
+
Inference requests from the agent never leave the sandbox directly. OpenShell intercepts every call and routes it to the provider you selected during onboarding.
| NVIDIA Endpoint |`nvidia/nemotron-3-super-120b-a12b`| Production. Requires an NVIDIA API key. |
184
+
Supported non-experimental onboarding paths:
187
185
188
-
Get an API key from [build.nvidia.com](https://build.nvidia.com). The `nemoclaw onboard` command prompts for this key during setup.
186
+
| Provider | Notes |
187
+
|---|---|
188
+
| NVIDIA Endpoints | Curated hosted models on `integrate.api.nvidia.com`. |
189
+
| OpenAI | Curated GPT models plus `Other...` for manual model entry. |
190
+
| Other OpenAI-compatible endpoint | For proxies and compatible gateways. |
191
+
| Anthropic | Curated Claude models plus `Other...` for manual model entry. |
192
+
| Other Anthropic-compatible endpoint | For Claude proxies and compatible gateways. |
193
+
| Google Gemini | Google's OpenAI-compatible endpoint. |
189
194
190
-
Local inference options such as Ollama and vLLM are still experimental. On macOS, they also depend on OpenShell host-routing support in addition to the local service itself being reachable on the host.
195
+
During onboarding, NemoClaw validates the selected provider and model before it creates the sandbox:
196
+
197
+
- OpenAI-compatible providers: tries `/responses` first, then `/chat/completions`
- If validation fails, the wizard prompts you to fix the selection before continuing
200
+
201
+
Credentials stay on the host in `~/.nemoclaw/credentials.json`. The sandbox only sees the routed `inference.local` endpoint, not your raw provider key.
202
+
203
+
Local Ollama is supported in the standard onboarding flow. Local vLLM remains experimental, and local host-routed inference on macOS still depends on OpenShell host-routing support in addition to the local service itself being reachable on the host.
191
204
192
205
---
193
206
@@ -252,7 +265,7 @@ Refer to the documentation for more information on NemoClaw.
252
265
-[Overview](https://docs.nvidia.com/nemoclaw/latest/about/overview.html): Learn what NemoClaw does and how it fits together.
253
266
-[How It Works](https://docs.nvidia.com/nemoclaw/latest/about/how-it-works.html): Learn about the plugin, blueprint, and sandbox lifecycle.
254
267
-[Architecture](https://docs.nvidia.com/nemoclaw/latest/reference/architecture.html): Learn about the plugin structure, blueprint lifecycle, and sandbox environment.
255
-
-[Inference Profiles](https://docs.nvidia.com/nemoclaw/latest/reference/inference-profiles.html): Learn about the NVIDIA Endpoint inference configuration.
268
+
-[Inference Profiles](https://docs.nvidia.com/nemoclaw/latest/reference/inference-profiles.html): Learn how NemoClaw configures routed inference providers.
256
269
-[Network Policies](https://docs.nvidia.com/nemoclaw/latest/reference/network-policies.html): Learn about egress control and policy customization.
257
270
-[CLI Commands](https://docs.nvidia.com/nemoclaw/latest/reference/commands.html): Learn about the full command reference.
258
271
-[Troubleshooting](https://docs.nvidia.com/nemoclaw/latest/reference/troubleshooting.html): Troubleshoot common issues and resolution steps.
0 commit comments