Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs(karpor): add api proxy parameters in helm installation document #599

Merged
Merged
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
15 changes: 14 additions & 1 deletion docs/karpor/1-getting-started/2-installation.md
Original file line number Diff line number Diff line change
Expand Up @@ -118,6 +118,14 @@ helm install karpor-release kusionstack/karpor \
--set server.ai.model=deepseek-chat \
--set server.ai.topP=0.5 \
--set server.ai.temperature=0.2

# Example using AI Proxy
helm install karpor kusionstack/karpor \
--set server.ai.authToken={YOUR_AI_TOKEN} \
--set server.ai.proxy.enabled=true \
--set server.ai.proxy.httpProxy={YOUR_HTTP_PROXY} \
--set server.ai.proxy.httpsProxy={YOUR_HTTPS_PROXY} \
--set server.ai.proxy.noProxy={YOUR_NO_PROXY}
```

## Chart Parameters
Expand All @@ -144,11 +152,16 @@ The Karpor Server Component is main backend server. It itself is an `apiserver`,

| Key | Type | Default | Description |
|-----|------|---------|-------------|
| server.ai | object | `{"authToken":"","backend":"openai","baseUrl":"","model":"gpt-3.5-turbo","temperature":1,"topP":1}` | AI configuration section. The AI analysis feature requires that [authToken, baseUrl] be assigned values. |
| server.ai | object | `{"authToken":"","backend":"openai","baseUrl":"","model":"gpt-3.5-turbo","proxy":{"enabled":false,"httpProxy":"","httpsProxy":"","noProxy":""},"temperature":1,"topP":1}` | AI configuration section. The AI analysis feature requires that [authToken, baseUrl] be assigned values. |
| server.ai.authToken | string | `""` | Authentication token for accessing the AI service. |
| server.ai.backend | string | `"openai"` | Backend service or platform that the AI model is hosted on. Available options: <br/>- `"openai"`: OpenAI API (default)<br/>- `"azureopenai"`: Azure OpenAI Service<br/>- `"huggingface"`: Hugging Face API<br/> If the backend you are using is compatible with OpenAI, then there is no need to make any changes here. |
| server.ai.baseUrl | string | `""` | Base URL of the AI service. e.g., "https://api.openai.com/v1". |
| server.ai.model | string | `"gpt-3.5-turbo"` | Name or identifier of the AI model to be used. e.g., "gpt-3.5-turbo". |
| server.ai.proxy | object | `{"enabled":false,"httpProxy":"","httpsProxy":"","noProxy":""}` | Proxy configuration for AI service connections. |
| server.ai.proxy.enabled | bool | `false` | Enable proxy settings for AI service connections. When false, proxy settings will be ignored. |
| server.ai.proxy.httpProxy | string | `""` | HTTP proxy URL for AI service connections (e.g., "http://proxy.example.com:8080"). |
| server.ai.proxy.httpsProxy | string | `""` | HTTPS proxy URL for AI service connections (e.g., "https://proxy.example.com:8080"). |
| server.ai.proxy.noProxy | string | `""` | No proxy configuration for AI service connections (e.g., "localhost,127.0.0.1,example.com"). |
| server.ai.temperature | float | `1` | Temperature parameter for the AI model. This controls the randomness of the output, where a higher value (e.g., 1.0) makes the output more random, and a lower value (e.g., 0.0) makes it more deterministic. |
| server.ai.topP | float | `1` | Top-p (nucleus sampling) parameter for the AI model. This controls Controls the probability mass to consider for sampling, where a higher value leads to greater diversity in the generated content (typically ranging from 0 to 1) |
| server.enableRbac | bool | `false` | Enable RBAC authorization if set to true. |
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -118,6 +118,14 @@ helm install karpor-release kusionstack/karpor \
--set server.ai.model=deepseek-chat \
--set server.ai.topP=0.5 \
--set server.ai.temperature=0.2

# 使用 AI Proxy 的样例
helm install karpor kusionstack/karpor \
--set server.ai.authToken={YOUR_AI_TOKEN} \
--set server.ai.proxy.enabled=true \
--set server.ai.proxy.httpProxy={YOUR_HTTP_PROXY} \
--set server.ai.proxy.httpsProxy={YOUR_HTTPS_PROXY} \
--set server.ai.proxy.noProxy={YOUR_NO_PROXY}
```

## Chart 参数
Expand All @@ -144,11 +152,16 @@ Karpor 服务器组件是主要的后端服务器。它本身是一个 `apiserve

| 键 | 类型 | 默认值 | 描述 |
|-----|------|---------|-------------|
| server.ai | object | `{"authToken":"","backend":"openai","baseUrl":"","model":"gpt-3.5-turbo","temperature":1,"topP":1}` | AI 配置部分。AI 分析功能需要为 [authToken, baseUrl] 赋值。 |
| server.ai | object | `{"authToken":"","backend":"openai","baseUrl":"","model":"gpt-3.5-turbo","proxy":{"enabled":false,"httpProxy":"","httpsProxy":"","noProxy":""},"temperature":1,"topP":1}` | AI 配置部分。AI 分析功能需要为 [authToken, baseUrl] 赋值。 |
| server.ai.authToken | string | `""` | 访问 AI 服务的认证令牌。 |
| server.ai.backend | string | `"openai"` | 托管 AI 模型的后端服务或平台。可用选项:<br/>- `"openai"`: OpenAI API(默认)<br/>- `"azureopenai"`: Azure OpenAI 服务<br/>- `"huggingface"`: Hugging Face API<br/>如果您使用的后端与 OpenAI 兼容,则无需在此处进行任何更改。 |
| server.ai.baseUrl | string | `""` | AI 服务的基础 URL。例如:"https://api.openai.com/v1"。 |
| server.ai.model | string | `"gpt-3.5-turbo"` | 要使用的 AI 模型的名称或标识符。例如:"gpt-3.5-turbo"。 |
| server.ai.proxy | object | `{"enabled":false,"httpProxy":"","httpsProxy":"","noProxy":""}` | AI 服务连接的代理配置。 |
| server.ai.proxy.enabled | bool | `false` | 启用 AI 服务连接的代理设置。如果为 false,则将忽略代理设置。 |
| server.ai.proxy.httpProxy | string | `""` | AI 服务连接的 HTTP 代理 URL(例如“http://proxy.example.com:8080”)。 |
| server.ai.proxy.httpsProxy | string | `""` | AI 服务连接的 HTTPS 代理 URL(例如“https://proxy.example.com:8080”)。 |
| server.ai.proxy.noProxy | string | `""` | AI 服务连接无需代理配置(例如“localhost,127.0.0.1,example.com”)。 |
| server.ai.temperature | float | `1` | AI 模型的温度参数。控制输出的随机性,较高的值(例如 1.0)使输出更随机,较低的值(例如 0.0)使输出更确定性。 |
| server.ai.topP | float | `1` | AI 模型的 Top-p(核采样)参数。控制采样的概率质量,较高的值导致生成内容的多样性更大(通常范围为 0 到 1)。 |
| server.enableRbac | bool | `false` | 如果设置为 true,则启用 RBAC 授权。 |
Expand Down