From b9dc117309fb9f3ad4d0ac1e00ede0154a24b5c6 Mon Sep 17 00:00:00 2001 From: Peter Steinberger Date: Sun, 25 Jan 2026 01:49:31 +0000 Subject: [PATCH] docs: refine venice highlight --- CHANGELOG.md | 1 + docs/providers/index.md | 11 ++++++- docs/providers/models.md | 10 +++++++ docs/providers/venice.md | 64 ++++++++++++++++++++++++++++++++++++---- 4 files changed, 80 insertions(+), 6 deletions(-) diff --git a/CHANGELOG.md b/CHANGELOG.md index 8a8737dd8..84bd00bd4 100644 --- a/CHANGELOG.md +++ b/CHANGELOG.md @@ -6,6 +6,7 @@ Docs: https://docs.clawd.bot ### Highlights - Ollama: provider discovery + docs. (#1606) Thanks @abhaymundhara. https://docs.clawd.bot/providers/ollama +- Venius (Venice AI): highlight provider guide + cross-links + expanded guidance. https://docs.clawd.bot/providers/venice ### Changes - TTS: add Edge TTS provider fallback, defaulting to keyless Edge with MP3 retry on format failures. (#1668) Thanks @steipete. https://docs.clawd.bot/tts diff --git a/docs/providers/index.md b/docs/providers/index.md index c6a1f1b5c..c4f020192 100644 --- a/docs/providers/index.md +++ b/docs/providers/index.md @@ -11,6 +11,15 @@ default model as `provider/model`. Looking for chat channel docs (WhatsApp/Telegram/Discord/Slack/Mattermost (plugin)/etc.)? See [Channels](/channels). +## Highlight: Venius (Venice AI) + +Venius is our recommended Venice AI setup for privacy-first inference with an option to use Opus for hard tasks. + +- Default: `venice/llama-3.3-70b` +- Best overall: `venice/claude-opus-45` (Opus remains the strongest) + +See [Venice AI](/providers/venice). + ## Quick start 1) Authenticate with the provider (usually via `clawdbot onboard`). @@ -35,7 +44,7 @@ Looking for chat channel docs (WhatsApp/Telegram/Discord/Slack/Mattermost (plugi - [Z.AI](/providers/zai) - [GLM models](/providers/glm) - [MiniMax](/providers/minimax) -- [Venice AI (privacy-focused)](/providers/venice) +- [Venius (Venice AI, privacy-focused)](/providers/venice) - [Ollama (local models)](/providers/ollama) ## Transcription providers diff --git a/docs/providers/models.md b/docs/providers/models.md index 0a50b0d5b..e581740a7 100644 --- a/docs/providers/models.md +++ b/docs/providers/models.md @@ -9,6 +9,15 @@ read_when: Clawdbot can use many LLM providers. Pick one, authenticate, then set the default model as `provider/model`. +## Highlight: Venius (Venice AI) + +Venius is our recommended Venice AI setup for privacy-first inference with an option to use Opus for the hardest tasks. + +- Default: `venice/llama-3.3-70b` +- Best overall: `venice/claude-opus-45` (Opus remains the strongest) + +See [Venice AI](/providers/venice). + ## Quick start (two steps) 1) Authenticate with the provider (usually via `clawdbot onboard`). @@ -32,6 +41,7 @@ model as `provider/model`. - [Z.AI](/providers/zai) - [GLM models](/providers/glm) - [MiniMax](/providers/minimax) +- [Venius (Venice AI)](/providers/venice) - [Amazon Bedrock](/bedrock) For the full provider catalog (xAI, Groq, Mistral, etc.) and advanced configuration, diff --git a/docs/providers/venice.md b/docs/providers/venice.md index d73eba621..bd91e6da6 100644 --- a/docs/providers/venice.md +++ b/docs/providers/venice.md @@ -1,7 +1,22 @@ -# Venice AI Provider +--- +summary: "Use Venice AI privacy-focused models in Clawdbot" +read_when: + - You want privacy-focused inference in Clawdbot + - You want Venice AI setup guidance +--- +# Venice AI (Venius highlight) + +**Venius** is our highlight Venice setup for privacy-first inference with optional anonymized access to proprietary models. Venice AI provides privacy-focused AI inference with support for uncensored models and access to major proprietary models through their anonymized proxy. All inference is private by default—no training on your data, no logging. +## Why Venice in Clawdbot + +- **Private inference** for open-source models (no logging). +- **Uncensored models** when you need them. +- **Anonymized access** to proprietary models (Opus/GPT/Gemini) when quality matters. +- OpenAI-compatible `/v1` endpoints. + ## Privacy Modes Venice offers two privacy levels — understanding this is key to choosing your model: @@ -20,6 +35,7 @@ Venice offers two privacy levels — understanding this is key to choosing your - **Streaming**: ✅ Supported on all models - **Function calling**: ✅ Supported on select models (check model capabilities) - **Vision**: ✅ Supported on models with vision capability +- **No hard rate limits**: Fair-use throttling may apply for extreme usage ## Setup @@ -54,8 +70,7 @@ This will: ```bash clawdbot onboard --non-interactive \ --auth-choice venice-api-key \ - --token "vapi_xxxxxxxxxxxx" \ - --token-provider venice + --venice-api-key "vapi_xxxxxxxxxxxx" ``` ### 3. Verify Setup @@ -68,8 +83,10 @@ clawdbot chat --model venice/llama-3.3-70b "Hello, are you working?" After setup, Clawdbot shows all available Venice models. Pick based on your needs: -- **Privacy**: Choose "private" models for fully private inference -- **Capability**: Choose "anonymized" models to access Claude, GPT, Gemini via Venice's proxy +- **Default (our pick)**: `venice/llama-3.3-70b` for private, balanced performance. +- **Best overall quality**: `venice/claude-opus-45` for hard jobs (Opus remains the strongest). +- **Privacy**: Choose "private" models for fully private inference. +- **Capability**: Choose "anonymized" models to access Claude, GPT, Gemini via Venice's proxy. Change your default model anytime: @@ -84,11 +101,18 @@ List all available models: clawdbot models list | grep venice ``` +## Configure via `clawdbot configure` + +1. Run `clawdbot configure` +2. Select **Model/auth** +3. Choose **Venice AI** + ## Which Model Should I Use? | Use Case | Recommended Model | Why | |----------|-------------------|-----| | **General chat** | `llama-3.3-70b` | Good all-around, fully private | +| **Best overall quality** | `claude-opus-45` | Opus remains the strongest for hard tasks | | **Privacy + Claude quality** | `claude-opus-45` | Best reasoning via anonymized proxy | | **Coding** | `qwen3-coder-480b-a35b-instruct` | Code-optimized, 262k context | | **Vision tasks** | `qwen3-vl-235b-a22b` | Best private vision model | @@ -202,6 +226,36 @@ The Venice model catalog updates dynamically. Run `clawdbot models list` to see Venice API is at `https://api.venice.ai/api/v1`. Ensure your network allows HTTPS connections. +## Config file example + +```json5 +{ + env: { VENICE_API_KEY: "vapi_..." }, + agents: { defaults: { model: { primary: "venice/llama-3.3-70b" } } }, + models: { + mode: "merge", + providers: { + venice: { + baseUrl: "https://api.venice.ai/api/v1", + apiKey: "${VENICE_API_KEY}", + api: "openai-completions", + models: [ + { + id: "llama-3.3-70b", + name: "Llama 3.3 70B", + reasoning: false, + input: ["text"], + cost: { input: 0, output: 0, cacheRead: 0, cacheWrite: 0 }, + contextWindow: 131072, + maxTokens: 8192 + } + ] + } + } + } +} +``` + ## Links - [Venice AI](https://venice.ai)