A custom provider connects ACTIVATE AI Chat to any endpoint that implements the OpenAI chat completions API. This gives you flexibility to use self-hosted models, third-party inference services, or any other compatible backend.
Your endpoint must support the following:
/v1/chat/completions — The standard OpenAI chat completions endpoint. ACTIVATE sends all chat messages to this route./v1/models (optional) — If your endpoint exposes this route, ACTIVATE can list available models dynamically so users can switch models without reconfiguring the provider.The endpoint must accept an API key via the Authorization: Bearer <key> header.
Before creating a custom provider, ensure that:
https://my-server.example.com/v1/chat/completions).Custom providers are available immediately after creation — no provisioning step is required.
If the model list dropdown is populated, your endpoint supports the /v1/models route and ACTIVATE has successfully queried it.
You can update the endpoint, API key, and model for a custom provider at any time:
This is useful for rotating API keys or pointing the provider at a different model without creating a new provider.
vLLM — Start vLLM with the --api-key flag to enable API key authentication. The default port is 8000. Your endpoint URL will be http://<host>:8000/v1/chat/completions.
Ollama — Ollama exposes an OpenAI-compatible API on port 11434. Set the endpoint to http://<host>:11434/v1/chat/completions. The model name should match the Ollama model tag (for example, "llama3").
Text Generation Inference (TGI) — TGI provides an OpenAI-compatible Messages API. Set the endpoint to http://<host>:8080/v1/chat/completions.
Deleting a custom provider only removes it from ACTIVATE. It does not affect the external inference endpoint.