diff --git a/docs/features/bring-your-own-llm.mdx b/docs/features/bring-your-own-llm.mdx index 8b15cd527..c3b4cf979 100644 --- a/docs/features/bring-your-own-llm.mdx +++ b/docs/features/bring-your-own-llm.mdx @@ -131,6 +131,29 @@ Connect to powerful AI models using your API keys. Your keys stay on your machin ![Gemini config](/images/byollm--gemini-provider-config.png) +
+ + NVIDIA's [build.nvidia.com](https://build.nvidia.com/models) hosts 80+ models — including GLM 5.1, MiniMax M2.7, GPT-OSS-120B, Qwen 3.5, Mistral, and Nemotron — behind a **free OpenAI-compatible API endpoint**. Great for chatting, prototyping, and personal projects. + + **Get your API key:** + 1. Go to [build.nvidia.com/models](https://build.nvidia.com/models) and sign in with a free NVIDIA developer account + 2. Pick any model tagged **Free Endpoint** (e.g. [`minimaxai/minimax-m2.7`](https://build.nvidia.com/minimaxai/minimax-m2.7), [`z-ai/glm-5.1`](https://build.nvidia.com/z-ai/glm-5.1), [`qwen/qwen3.5-122b-a10b`](https://build.nvidia.com/qwen/qwen3.5-122b-a10b)) + 3. Click **Get API Key** on the model page and copy the `nvapi-...` key + + **Add to BrowserOS:** + 1. Go to `chrome://browseros/settings` + 2. Click **USE** on the **OpenAI Compatible** card + 3. Set **Base URL** to `https://integrate.api.nvidia.com/v1` + 4. Set **Model ID** to a model from the catalog (e.g. `minimaxai/minimax-m2.7`, `z-ai/glm-5.1`, `qwen/qwen3.5-122b-a10b`) + 5. Paste your NVIDIA API key + 6. Set **Context Window** based on the model (most are `128000` or higher) + 7. Click **Save** + + + NVIDIA's free endpoints share GPU capacity across all developers, so throughput is slower than a paid API. They're best for Chat Mode, exploring new open-source models, and personal projects. For production agent workloads, use a paid provider like Claude or Kimi. + + +
Claude Opus 4.5 gives the best results for Agent Mode.