diff --git a/pages/blog/2023-11-30_litellm.mdx b/pages/blog/2023-11-30_litellm.mdx index 358f8aa0d..2180ddcac 100644 --- a/pages/blog/2023-11-30_litellm.mdx +++ b/pages/blog/2023-11-30_litellm.mdx @@ -268,7 +268,6 @@ custom: titleModel: "gpt-3.5-turbo" summarize: false summaryModel: "gpt-3.5-turbo" - forcePrompt: false modelDisplayLabel: "Lite LLM" ``` diff --git a/pages/docs/configuration/azure.mdx b/pages/docs/configuration/azure.mdx index 651e801b9..cc2e3f611 100644 --- a/pages/docs/configuration/azure.mdx +++ b/pages/docs/configuration/azure.mdx @@ -183,7 +183,6 @@ options={[ ['serverless', 'boolean', 'Specifies if the group is a serverless inference chat completions endpoint from Azure Model Catalog, for which only a model identifier, baseURL, and apiKey are needed. For more info, see serverless inference endpoints.', 'serverless: true'], ['addParams', 'object', 'Adds or overrides additional parameters for Azure OpenAI API requests. Useful for specifying API-specific options as key-value pairs.', 'addParams: {temperature: 0.7}'], ['dropParams', 'array', 'Allows for the exclusion of certain default parameters from Azure OpenAI API requests. Useful for APIs that do not accept or recognize specific parameters. This should be specified as a list of strings.', 'dropParams: [top_p, stop]'], -['forcePrompt', 'boolean', 'Dictates whether to send a prompt parameter instead of messages in the request body. This option is useful when needing to format the request in a manner consistent with OpenAI API expectations, particularly for scenarios preferring a single text payload.', 'forcePrompt: true'], ]} /> @@ -215,7 +214,6 @@ endpoints: dropParams: - "frequency_penalty" - "presence_penalty" - forcePrompt: false models: # ... (model-level configurations) ``` @@ -583,4 +581,3 @@ endpoints: - Compatibility with LibreChat relies on parity with OpenAI API specs, which at the time of writing, are typically **"Pay-as-you-go"** or "Models as a Service" (MaaS) deployments on Azure AI Studio, that are OpenAI-SDK-compatible with either `v1/completions` or `models/chat/completions` endpoint handling. - All models that offer serverless deployments ("Serverless APIs") are compatible from the Azure model catalog. You can filter by "Serverless API" under Deployment options and "Chat completion" under inference tasks to see the full list; however, real time endpoint models have not been tested. - These serverless inference endpoint/models may or may not support function calling according to OpenAI API specs, which enables their use with Agents. -- If using legacy "/v1/completions" (without "chat"), you need to set the `forcePrompt` field to `true` in your [group config.](#group-level-configuration) diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/anyscale.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/anyscale.mdx index 09061b2e1..c2505135a 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/anyscale.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/anyscale.mdx @@ -24,7 +24,6 @@ description: Example configuration for Anyscale titleModel: "meta-llama/Llama-2-7b-chat-hf" summarize: false summaryModel: "meta-llama/Llama-2-7b-chat-hf" - forcePrompt: false modelDisplayLabel: "Anyscale" ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/fireworks.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/fireworks.mdx index 5043f9c69..db055747b 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/fireworks.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/fireworks.mdx @@ -25,7 +25,6 @@ description: Example configuration for Fireworks titleModel: "accounts/fireworks/models/llama-v2-7b-chat" summarize: false summaryModel: "accounts/fireworks/models/llama-v2-7b-chat" - forcePrompt: false modelDisplayLabel: "Fireworks" dropParams: ["user"] ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/litellm.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/litellm.mdx index 9128dd3d0..c389b8828 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/litellm.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/litellm.mdx @@ -21,7 +21,6 @@ description: Example configuration for LiteLLM titleModel: "gpt-3.5-turbo" summarize: false summaryModel: "gpt-3.5-turbo" - forcePrompt: false modelDisplayLabel: "LiteLLM" ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/mlx.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/mlx.mdx index e4abc03b1..352ef3936 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/mlx.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/mlx.mdx @@ -27,7 +27,6 @@ description: Example configuration for Apple MLX titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "Apple MLX" addParams: max_tokens: 2000 diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/moonshot.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/moonshot.mdx new file mode 100644 index 000000000..ac6a7896c --- /dev/null +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/moonshot.mdx @@ -0,0 +1,25 @@ +--- +title: Moonshot +description: Example configuration for Moonshot AI (Kimi) +--- + +# [Moonshot](https://www.moonshot.ai/) + +> Moonshot API key: [platform.moonshot.cn](https://platform.moonshot.ai) + +**Notes:** + +- **Known:** icon provided. +- **Important:** For models with reasoning/thinking capabilities (e.g., `kimi-k2.5`, `kimi-k2-thinking`), the endpoint `name` **must** be set to `"Moonshot"` (case-insensitive) for interleaved reasoning to work correctly with tool calls. Using a different name will result in errors like `thinking is enabled but reasoning_content is missing in assistant tool call message`. See [Moonshot's documentation](https://platform.moonshot.ai/docs/guide/use-kimi-k2-thinking-model#frequently-asked-questions) for more details. + +```yaml filename="librechat.yaml" + - name: "Moonshot" + apiKey: "${MOONSHOT_API_KEY}" + baseURL: "https://api.moonshot.ai/v1" + models: + default: ["kimi-k2.5"] + fetch: true + titleConvo: true + titleModel: "current_model" + modelDisplayLabel: "Moonshot" +``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/neurochain.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/neurochain.mdx index e470c96a7..f0109b616 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/neurochain.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/neurochain.mdx @@ -26,7 +26,6 @@ description: Example configuration for NeurochainAI titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "NeurochainAI" iconURL: "https://raw.githubusercontent.com/LibreChat-AI/librechat-config-yaml/refs/heads/main/icons/NeurochainAI.png" ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/ollama.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/ollama.mdx index 0bca15267..1a0cc5e45 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/ollama.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/ollama.mdx @@ -36,7 +36,6 @@ description: Example configuration for Ollama titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "Ollama" ``` @@ -60,7 +59,6 @@ However, in case you experience the behavior where `llama3` does not stop genera titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "Ollama" addParams: "stop": [ diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/perplexity.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/perplexity.mdx index b7398ae9e..3073c2792 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/perplexity.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/perplexity.mdx @@ -32,7 +32,6 @@ description: Example configuration for Perplexity titleModel: "llama-3-sonar-small-32k-chat" summarize: false summaryModel: "llama-3-sonar-small-32k-chat" - forcePrompt: false dropParams: ["stop", "frequency_penalty"] modelDisplayLabel: "Perplexity" ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/portkey.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/portkey.mdx index cb8fc5922..286a0f990 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/portkey.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/portkey.mdx @@ -33,7 +33,6 @@ description: Example configuration for Portkey AI titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "Portkey:OpenAI" iconURL: https://images.crunchbase.com/image/upload/c_pad,f_auto,q_auto:eco,dpr_1/rjqy7ghvjoiu4cd1xjbf ``` @@ -55,7 +54,6 @@ description: Example configuration for Portkey AI titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "Portkey:Llama" iconURL: https://images.crunchbase.com/image/upload/c_pad,f_auto,q_auto:eco,dpr_1/rjqy7ghvjoiu4cd1xjbf ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/shuttleai.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/shuttleai.mdx index aee5a6057..e1919e35f 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/shuttleai.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/shuttleai.mdx @@ -24,7 +24,6 @@ description: Example configuration for ShuttleAI titleModel: "shuttle-2.5-mini" summarize: false summaryModel: "shuttle-2.5-mini" - forcePrompt: false modelDisplayLabel: "ShuttleAI" dropParams: ["user", "stop"] ``` diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/togetherai.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/togetherai.mdx index 0584cec94..b1c64d383 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/togetherai.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/togetherai.mdx @@ -168,6 +168,5 @@ with open("models_togetherai.json", "w") as file: titleModel: "togethercomputer/llama-2-7b-chat" summarize: false summaryModel: "togethercomputer/llama-2-7b-chat" - forcePrompt: false modelDisplayLabel: "together.ai" ``` \ No newline at end of file diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/truefoundry.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/truefoundry.mdx index 9f53a7e31..d14b74119 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/truefoundry.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/truefoundry.mdx @@ -25,7 +25,6 @@ To use [TrueFoundry's AI Gateway](https://www.truefoundry.com/ai-gateway) follow titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "TrueFoundry:OpenAI" ``` For more details you can check: [TrueFoundry Docs](https://docs.truefoundry.com/docs/introduction) \ No newline at end of file diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/vllm.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/vllm.mdx index 5489c0cba..b5c22eb5e 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/vllm.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/vllm.mdx @@ -27,7 +27,6 @@ description: Example configuration for vLLM titleMessageRole: "user" summarize: false summaryModel: "current_model" - forcePrompt: false ``` The configuration above connects LibreChat to a local vLLM server running on port 8023. It uses the Gemma 3 27B model as the default model, but will fetch all available models from your vLLM server. diff --git a/pages/docs/configuration/librechat_yaml/ai_endpoints/xai.mdx b/pages/docs/configuration/librechat_yaml/ai_endpoints/xai.mdx index 8299e62a9..623f9367f 100644 --- a/pages/docs/configuration/librechat_yaml/ai_endpoints/xai.mdx +++ b/pages/docs/configuration/librechat_yaml/ai_endpoints/xai.mdx @@ -23,7 +23,6 @@ description: Example configuration for xAI titleModel: "grok-beta" summarize: false summaryModel: "grok-beta" - forcePrompt: false modelDisplayLabel: "Grok" ``` diff --git a/pages/docs/configuration/librechat_yaml/example.mdx b/pages/docs/configuration/librechat_yaml/example.mdx index cc8bf7c85..27f457711 100644 --- a/pages/docs/configuration/librechat_yaml/example.mdx +++ b/pages/docs/configuration/librechat_yaml/example.mdx @@ -45,7 +45,6 @@ endpoints: titleModel: "meta-llama/Llama-2-7b-chat-hf" summarize: false summaryModel: "meta-llama/Llama-2-7b-chat-hf" - forcePrompt: false modelDisplayLabel: "Anyscale" # APIpie @@ -99,7 +98,6 @@ endpoints: titleModel: "accounts/fireworks/models/llama-v2-7b-chat" summarize: false summaryModel: "accounts/fireworks/models/llama-v2-7b-chat" - forcePrompt: false modelDisplayLabel: "Fireworks" dropParams: ["user"] @@ -148,7 +146,6 @@ endpoints: titleModel: "gpt-3.5-turbo" summarize: false summaryModel: "gpt-3.5-turbo" - forcePrompt: false modelDisplayLabel: "OpenRouter" # Perplexity @@ -168,7 +165,6 @@ endpoints: titleModel: "sonar-medium-chat" summarize: false summaryModel: "sonar-medium-chat" - forcePrompt: false dropParams: ["stop", "frequency_penalty"] modelDisplayLabel: "Perplexity" @@ -185,7 +181,6 @@ endpoints: titleModel: "gemini-pro" summarize: false summaryModel: "llama-summarize" - forcePrompt: false modelDisplayLabel: "ShuttleAI" dropParams: ["user"] @@ -245,7 +240,6 @@ endpoints: titleModel: "togethercomputer/llama-2-7b-chat" summarize: false summaryModel: "togethercomputer/llama-2-7b-chat" - forcePrompt: false modelDisplayLabel: "together.ai" ``` @@ -379,9 +373,6 @@ endpoints: # Summary Model: Specify the model to use if summarization is enabled. # summaryModel: "mistral-tiny" # Defaults to "gpt-3.5-turbo" if omitted. - # Force Prompt setting: If true, sends a `prompt` parameter instead of `messages`. - # forcePrompt: false - # The label displayed for the AI model in messages. modelDisplayLabel: 'Mistral' # Default is "AI" when not set. diff --git a/pages/docs/configuration/librechat_yaml/object_structure/azure_openai.mdx b/pages/docs/configuration/librechat_yaml/object_structure/azure_openai.mdx index 13d490d5c..7b1a1712a 100644 --- a/pages/docs/configuration/librechat_yaml/object_structure/azure_openai.mdx +++ b/pages/docs/configuration/librechat_yaml/object_structure/azure_openai.mdx @@ -228,22 +228,6 @@ addParams: dropParams: ["stop", "user", "frequency_penalty", "presence_penalty"] ``` -### forcePrompt - -**Key:** - - -**Default:** Not specified - -**Example:** -```yaml filename="endpoints / azureOpenAI / groups / {group_item} / forcePrompt" -forcePrompt: false -``` - ### models **Key:** diff --git a/pages/docs/configuration/librechat_yaml/object_structure/custom_endpoint.mdx b/pages/docs/configuration/librechat_yaml/object_structure/custom_endpoint.mdx index 0b220fa61..ae0baa46b 100644 --- a/pages/docs/configuration/librechat_yaml/object_structure/custom_endpoint.mdx +++ b/pages/docs/configuration/librechat_yaml/object_structure/custom_endpoint.mdx @@ -110,18 +110,25 @@ iconURL: https://github.com/danny-avila/LibreChat/raw/main/docs/assets/LibreChat * If you want to use existing project icons, define the endpoint `name` as one of the main endpoints (case-sensitive): - "openAI" | "azureOpenAI" | "google" | "anthropic" | "assistants" | "gptPlugins" * There are also "known endpoints" (case-insensitive), which have icons provided. If your endpoint `name` matches the following names, you should omit this field: - - "Mistral" - - "Deepseek" - - "OpenRouter" - - "groq" - - "APIpie" - "Anyscale" + - "APIpie" + - "Cohere" + - "Deepseek" - "Fireworks" + - "groq" + - "Helicone" + - "Huggingface" + - "Mistral" + - "MLX" + - "Moonshot" + - "ollama" + - "OpenRouter" - "Perplexity" + - "Qwen" + - "ShuttleAI" - "together.ai" - - "ollama" + - "Unify" - "xai" - - "MLX" ## models @@ -361,22 +368,6 @@ summarize: false summaryModel: "mistral-tiny" ``` -## forcePrompt - -**Key:** - - -**Default:** `false` - -**Example:** -```yaml filename="endpoints / custom / forcePrompt" -forcePrompt: false -``` - ## modelDisplayLabel **Key:** diff --git a/pages/docs/quick_start/custom_endpoints.mdx b/pages/docs/quick_start/custom_endpoints.mdx index 8dae05f65..d2f48a317 100644 --- a/pages/docs/quick_start/custom_endpoints.mdx +++ b/pages/docs/quick_start/custom_endpoints.mdx @@ -53,7 +53,6 @@ services: titleModel: "current_model" summarize: false summaryModel: "current_model" - forcePrompt: false modelDisplayLabel: "OpenRouter" - name: "Ollama" apiKey: "ollama"