From 392cd589325c4d7a6e10e71693b97621b46ceb85 Mon Sep 17 00:00:00 2001 From: Felarof Date: Wed, 22 Apr 2026 12:45:37 -0700 Subject: [PATCH] docs(byollm): add NVIDIA free endpoint provider (#784) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Document NVIDIA's free OpenAI-compatible API at build.nvidia.com — 80+ free models including GLM 5.1, MiniMax M2.7, Qwen 3.5, Mistral, and Nemotron — wired through BrowserOS's OpenAI Compatible provider template. Co-authored-by: Claude Opus 4.7 (1M context) --- docs/features/bring-your-own-llm.mdx | 23 +++++++++++++++++++++++ 1 file changed, 23 insertions(+) diff --git a/docs/features/bring-your-own-llm.mdx b/docs/features/bring-your-own-llm.mdx index 8b15cd527..c3b4cf979 100644 --- a/docs/features/bring-your-own-llm.mdx +++ b/docs/features/bring-your-own-llm.mdx @@ -131,6 +131,29 @@ Connect to powerful AI models using your API keys. Your keys stay on your machin ![Gemini config](/images/byollm--gemini-provider-config.png) +
+ + NVIDIA's [build.nvidia.com](https://build.nvidia.com/models) hosts 80+ models — including GLM 5.1, MiniMax M2.7, GPT-OSS-120B, Qwen 3.5, Mistral, and Nemotron — behind a **free OpenAI-compatible API endpoint**. Great for chatting, prototyping, and personal projects. + + **Get your API key:** + 1. Go to [build.nvidia.com/models](https://build.nvidia.com/models) and sign in with a free NVIDIA developer account + 2. Pick any model tagged **Free Endpoint** (e.g. [`minimaxai/minimax-m2.7`](https://build.nvidia.com/minimaxai/minimax-m2.7), [`z-ai/glm-5.1`](https://build.nvidia.com/z-ai/glm-5.1), [`qwen/qwen3.5-122b-a10b`](https://build.nvidia.com/qwen/qwen3.5-122b-a10b)) + 3. Click **Get API Key** on the model page and copy the `nvapi-...` key + + **Add to BrowserOS:** + 1. Go to `chrome://browseros/settings` + 2. Click **USE** on the **OpenAI Compatible** card + 3. Set **Base URL** to `https://integrate.api.nvidia.com/v1` + 4. Set **Model ID** to a model from the catalog (e.g. `minimaxai/minimax-m2.7`, `z-ai/glm-5.1`, `qwen/qwen3.5-122b-a10b`) + 5. Paste your NVIDIA API key + 6. Set **Context Window** based on the model (most are `128000` or higher) + 7. Click **Save** + + + NVIDIA's free endpoints share GPU capacity across all developers, so throughput is slower than a paid API. They're best for Chat Mode, exploring new open-source models, and personal projects. For production agent workloads, use a paid provider like Claude or Kimi. + + +
Claude Opus 4.5 gives the best results for Agent Mode.