Files
openclaw/docs/zh-CN/providers/index.md
Seb Slight 929a3725d3 docs: canonicalize docs paths and align zh navigation (#11428)
* docs(navigation): canonicalize paths and align zh nav

* chore(docs): remove stray .DS_Store

* docs(scripts): add non-mint docs link audit

* docs(nav): fix zh source paths and preserve legacy redirects (#11428) (thanks @sebslight)

* chore(docs): satisfy lint for docs link audit script (#11428) (thanks @sebslight)
2026-02-07 15:40:35 -05:00

69 lines
2.1 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
read_when:
- 你想选择一个模型提供商
- 你需要快速了解支持的 LLM 后端
summary: OpenClaw 支持的模型提供商LLM
title: 模型提供商
x-i18n:
generated_at: "2026-02-03T07:53:32Z"
model: claude-opus-4-5
provider: pi
source_hash: eb4a97438adcf610499253afcf8b2af6624f4be098df389a6c3746f14c4a901b
source_path: providers/index.md
workflow: 15
---
# 模型提供商
OpenClaw 可以使用许多 LLM 提供商。选择一个提供商,进行认证,然后将默认模型设置为 `provider/model`
正在寻找聊天渠道文档WhatsApp/Telegram/Discord/Slack/Mattermost插件参见[渠道](/channels)。
## 亮点VeniceVenice AI
Venice 是我们推荐的 Venice AI 设置,用于隐私优先的推理,并可选择使用 Opus 处理困难任务。
- 默认:`venice/llama-3.3-70b`
- 最佳综合:`venice/claude-opus-45`Opus 仍然是最强的)
参见 [Venice AI](/providers/venice)。
## 快速开始
1. 与提供商进行认证(通常通过 `openclaw onboard`)。
2. 设置默认模型:
```json5
{
agents: { defaults: { model: { primary: "anthropic/claude-opus-4-5" } } },
}
```
## 提供商文档
- [OpenAIAPI + Codex](/providers/openai)
- [AnthropicAPI + Claude Code CLI](/providers/anthropic)
- [QwenOAuth](/providers/qwen)
- [OpenRouter](/providers/openrouter)
- [Vercel AI Gateway](/providers/vercel-ai-gateway)
- [Moonshot AIKimi + Kimi Coding](/providers/moonshot)
- [OpenCode Zen](/providers/opencode)
- [Amazon Bedrock](/providers/bedrock)
- [Z.AI](/providers/zai)
- [Xiaomi](/providers/xiaomi)
- [GLM 模型](/providers/glm)
- [MiniMax](/providers/minimax)
- [VeniceVenice AI注重隐私](/providers/venice)
- [Ollama本地模型](/providers/ollama)
## 转录提供商
- [Deepgram音频转录](/providers/deepgram)
## 社区工具
- [Claude Max API Proxy](/providers/claude-max-api-proxy) - 将 Claude Max/Pro 订阅作为 OpenAI 兼容的 API 端点使用
有关完整的提供商目录xAI、Groq、Mistral 等)和高级配置,
参见[模型提供商](/concepts/model-providers)。