Skip to content

Commit c72a610

Browse files
fix
Signed-off-by: Abhishek Choudhary <shreemaan.abhishek@gmail.com>
1 parent ea985bb commit c72a610

File tree

6 files changed

+26
-12
lines changed

6 files changed

+26
-12
lines changed

docs/en/latest/plugins/ai-proxy-multi.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -60,15 +60,15 @@ In addition, the Plugin also supports logging LLM request information in the acc
6060
| instances.name | string | True | | | Name of the LLM service instance. |
6161
| instances.provider | string | True | | [openai, deepseek, azure-openai, aimlapi, anthropic, openrouter, gemini, vertex-ai, openai-compatible] | LLM service provider. When set to `openai`, the Plugin will proxy the request to `api.openai.com`. When set to `deepseek`, the Plugin will proxy the request to `api.deepseek.com`. When set to `aimlapi`, the Plugin uses the OpenAI-compatible driver and proxies the request to `api.aimlapi.com` by default. When set to `anthropic`, the Plugin will proxy the request to `api.anthropic.com` by default. When set to `openrouter`, the Plugin uses the OpenAI-compatible driver and proxies the request to `openrouter.ai` by default. When set to `gemini`, the Plugin uses the OpenAI-compatible driver and proxies the request to `generativelanguage.googleapis.com` by default. When set to `vertex-ai`, the Plugin will proxy the request to `aiplatform.googleapis.com` by default and requires `provider_conf` or `override`. When set to `openai-compatible`, the Plugin will proxy the request to the custom endpoint configured in `override`. |
6262
| instances.provider_conf | object | False | | | Configuration for the specific provider. Required when `provider` is set to `vertex-ai` and `override` is not configured. |
63-
| instances.provider_conf.project_id | string | True | | | Google Cloud Project ID. Required when `provider` is set to `vertex-ai`. |
64-
| instances.provider_conf.region | string | True | | | Google Cloud Region. Required when `provider` is set to `vertex-ai`. |
63+
| instances.provider_conf.project_id | string | True | | | Google Cloud Project ID. |
64+
| instances.provider_conf.region | string | True | | | Google Cloud Region. |
6565
| instances.priority | integer | False | 0 | | Priority of the LLM instance in load balancing. `priority` takes precedence over `weight`. |
6666
| instances.weight | string | True | 0 | greater or equal to 0 | Weight of the LLM instance in load balancing. |
6767
| instances.auth | object | True | | | Authentication configurations. |
6868
| instances.auth.header | object | False | | | Authentication headers. At least one of the `header` and `query` should be configured. |
6969
| instances.auth.query | object | False | | | Authentication query parameters. At least one of the `header` and `query` should be configured. |
7070
| instances.auth.gcp | object | False | | | Configuration for Google Cloud Platform (GCP) authentication. |
71-
| instances.auth.gcp.service_account_json | string | False | | | Content of the GCP service account JSON file. |
71+
| instances.auth.gcp.service_account_json | string | False | | | Content of the GCP service account JSON file. This can also be configured by setting the `GCP_SERVICE_ACCOUNT` environment variable. |
7272
| instances.auth.gcp.max_ttl | integer | False | | minimum = 1 | Maximum TTL (in seconds) for caching the GCP access token. |
7373
| instances.auth.gcp.expire_early_secs| integer | False | 60 | minimum = 0 | Seconds to expire the access token before its actual expiration time to avoid edge cases. |
7474
| instances.options | object | False | | | Model configurations. In addition to `model`, you can configure additional parameters and they will be forwarded to the upstream LLM service in the request body. For instance, if you are working with OpenAI, DeepSeek, or AIMLAPI, you can configure additional parameters such as `max_tokens`, `temperature`, `top_p`, and `stream`. See your LLM provider's API documentation for more available options. |

docs/en/latest/plugins/ai-proxy.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -53,13 +53,13 @@ In addition, the Plugin also supports logging LLM request information in the acc
5353
|--------------------|--------|----------|---------|------------------------------------------|-------------|
5454
| provider | string | True | | [openai, deepseek, azure-openai, aimlapi, anthropic, openrouter, gemini, vertex-ai, openai-compatible] | LLM service provider. When set to `openai`, the Plugin will proxy the request to `https://api.openai.com/chat/completions`. When set to `deepseek`, the Plugin will proxy the request to `https://api.deepseek.com/chat/completions`. When set to `aimlapi`, the Plugin uses the OpenAI-compatible driver and proxies the request to `https://api.aimlapi.com/v1/chat/completions` by default. When set to `anthropic`, the Plugin will proxy the request to `https://api.anthropic.com/v1/chat/completions` by default. When set to `openrouter`, the Plugin uses the OpenAI-compatible driver and proxies the request to `https://openrouter.ai/api/v1/chat/completions` by default. When set to `gemini`, the Plugin uses the OpenAI-compatible driver and proxies the request to `https://generativelanguage.googleapis.com/v1beta/openai/chat/completions` by default. When set to `vertex-ai`, the Plugin will proxy the request to `https://aiplatform.googleapis.com` by default and requires `provider_conf` or `override`. When set to `openai-compatible`, the Plugin will proxy the request to the custom endpoint configured in `override`. |
5555
| provider_conf | object | False | | | Configuration for the specific provider. Required when `provider` is set to `vertex-ai` and `override` is not configured. |
56-
| provider_conf.project_id | string | True | | | Google Cloud Project ID. Required when `provider` is set to `vertex-ai`. |
57-
| provider_conf.region | string | True | | | Google Cloud Region. Required when `provider` is set to `vertex-ai`. |
56+
| provider_conf.project_id | string | True | | | Google Cloud Project ID. |
57+
| provider_conf.region | string | True | | | Google Cloud Region. |
5858
| auth | object | True | | | Authentication configurations. |
5959
| auth.header | object | False | | | Authentication headers. At least one of `header` or `query` must be configured. |
6060
| auth.query | object | False | | | Authentication query parameters. At least one of `header` or `query` must be configured. |
6161
| auth.gcp | object | False | | | Configuration for Google Cloud Platform (GCP) authentication. |
62-
| auth.gcp.service_account_json | string | False | | | Content of the GCP service account JSON file. |
62+
| auth.gcp.service_account_json | string | False | | | Content of the GCP service account JSON file. This can also be configured by setting the `GCP_SERVICE_ACCOUNT` environment variable. |
6363
| auth.gcp.max_ttl | integer | False | | minimum = 1 | Maximum TTL (in seconds) for caching the GCP access token. |
6464
| auth.gcp.expire_early_secs | integer | False | 60 | minimum = 0 | Seconds to expire the access token before its actual expiration time to avoid edge cases. |
6565
| options | object | False | | | Model configurations. In addition to `model`, you can configure additional parameters and they will be forwarded to the upstream LLM service in the request body. For instance, if you are working with OpenAI, you can configure additional parameters such as `temperature`, `top_p`, and `stream`. See your LLM provider's API documentation for more available options. |

docs/en/latest/plugins/ai-request-rewrite.md

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,9 +37,16 @@ The `ai-request-rewrite` plugin intercepts client requests before they are forwa
3737
| ------------------------- | ------------ | -------- | ------------------------------------------------------------------------------------ |
3838
| prompt | Yes | String | The prompt send to LLM service. |
3939
| provider | Yes | String | Name of the LLM service. Available options: openai, deekseek, azure-openai, aimlapi, anthropic, openrouter, gemini, vertex-ai, and openai-compatible. When `aimlapi` is selected, the plugin uses the OpenAI-compatible driver with a default endpoint of `https://api.aimlapi.com/v1/chat/completions`. |
40+
| provider_conf | No | Object | Configuration for the specific provider. Required when `provider` is set to `vertex-ai` and `override` is not configured. |
41+
| provider_conf.project_id | Yes | String | Google Cloud Project ID. |
42+
| provider_conf.region | Yes | String | Google Cloud Region. |
4043
| auth | Yes | Object | Authentication configuration |
4144
| auth.header | No | Object | Authentication headers. Key must match pattern `^[a-zA-Z0-9._-]+$`. |
4245
| auth.query | No | Object | Authentication query parameters. Key must match pattern `^[a-zA-Z0-9._-]+$`. |
46+
| auth.gcp | No | Object | Configuration for Google Cloud Platform (GCP) authentication. |
47+
| auth.gcp.service_account_json | No | String | Content of the GCP service account JSON file. This can also be configured by setting the `GCP_SERVICE_ACCOUNT` environment variable. |
48+
| auth.gcp.max_ttl | No | Integer | Maximum TTL (in seconds) for caching the GCP access token. Minimum: 1. |
49+
| auth.gcp.expire_early_secs| No | Integer | Seconds to expire the access token before its actual expiration time to avoid edge cases. Minimum: 0. Default: 60. |
4350
| options | No | Object | Key/value settings for the model |
4451
| options.model | No | String | Model to execute. Examples: "gpt-3.5-turbo" for openai, "deepseek-chat" for deekseek, or "qwen-turbo" for openai-compatible or aimlapi services |
4552
| override.endpoint | No | String | Override the default endpoint when using OpenAI-compatible services (e.g., self-hosted models or third-party LLM services). When the provider is 'openai-compatible', the endpoint field is required. |

docs/zh/latest/plugins/ai-proxy-multi.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -60,16 +60,16 @@ description: ai-proxy-multi 插件通过负载均衡、重试、故障转移和
6060
| instances.name | string || | | LLM 服务实例的名称。 |
6161
| instances.provider | string | 是 | | [openai, deepseek, azure-openai, aimlapi, anthropic, openrouter, gemini, vertex-ai, openai-compatible] | LLM 服务提供商。设置为 `openai` 时,插件将代理请求到 `api.openai.com`。设置为 `deepseek` 时,插件将代理请求到 `api.deepseek.com`。设置为 `aimlapi` 时,插件使用 OpenAI 兼容驱动程序,默认将请求代理到 `api.aimlapi.com`。设置为 `anthropic` 时,插件使用 OpenAI 兼容驱动程序,默认将请求代理到 `api.anthropic.com`。设置为 `openrouter` 时,插件使用 OpenAI 兼容驱动程序,默认将请求代理到 `openrouter.ai`。设置为 `gemini` 时,插件使用 OpenAI 兼容驱动程序,默认将请求代理到 `generativelanguage.googleapis.com`。设置为 `vertex-ai` 时,插件默认将请求代理到 `aiplatform.googleapis.com`,且需要配置 `provider_conf` 或 `override`。设置为 `openai-compatible` 时,插件将代理请求到在 `override` 中配置的自定义端点。 |
6262
| instances.provider_conf | object || | | 特定提供商的配置。当 `provider` 设置为 `vertex-ai` 且未配置 `override` 时必填。 |
63-
| instances.provider_conf.project_id | string || | | Google Cloud 项目 ID。`provider` 设置为 `vertex-ai` 时必填。 |
64-
| instances.provider_conf.region | string || | | Google Cloud 区域。`provider` 设置为 `vertex-ai` 时必填。 |
63+
| instances.provider_conf.project_id | string || | | Google Cloud 项目 ID。 |
64+
| instances.provider_conf.region | string || | | Google Cloud 区域。 |
6565
| instances.priority | integer || 0 | | LLM 实例在负载均衡中的优先级。`priority` 优先于 `weight`|
6666
| instances.weight | string || 0 | 大于或等于 0 | LLM 实例在负载均衡中的权重。 |
6767
| instances.auth | object || | | 身份验证配置。 |
6868
| instances.auth.header | object || | | 身份验证标头。应配置 `header``query` 中的至少一个。 |
6969
| instances.auth.query | object || | | 身份验证查询参数。应配置 `header``query` 中的至少一个。 |
7070
| instances.auth.gcp | object || | | Google Cloud Platform (GCP) 身份验证配置。 |
7171
| instances.auth.gcp.service_account_json | string || | | GCP 服务账号 JSON 文件的内容。 |
72-
| instances.auth.gcp.max_ttl | integer || | minimum = 1 | 缓存 GCP 访问令牌的最大 TTL(秒)。 |
72+
| instances.auth.gcp.max_ttl | integer || | minimum = 1 | GCP 服务帐户 JSON 文件的内容。也可以通过设置“GCP_SERVICE_ACCOUNT”环境变量来配置 |
7373
| instances.auth.gcp.expire_early_secs| integer || 60 | minimum = 0 | 在访问令牌实际过期时间之前使其过期的秒数,以避免边缘情况。 |
7474
| instances.options | object || | | 模型配置。除了 `model` 之外,您还可以配置其他参数,它们将在请求体中转发到上游 LLM 服务。例如,如果您使用 OpenAI、DeepSeek 或 AIMLAPI,可以配置其他参数,如 `max_tokens``temperature``top_p``stream`。有关更多可用选项,请参阅您的 LLM 提供商的 API 文档。 |
7575
| instances.options.model | string || | | LLM 模型的名称,如 `gpt-4``gpt-3.5`。有关更多可用模型,请参阅您的 LLM 提供商的 API 文档。 |

0 commit comments

Comments
 (0)