Skip to main content


OpenRouter provides a unified interface for LLM APIs.

The OpenRouter provider is compatible with all the options provided by the OpenAI provider.

Here's an example of how to configure the provider to use the mistralai/mistral-medium model from OpenRouter:

- id: openrouter:mistralai/mistral-medium # or any other chat model
temperature: 0.5
apiKey: insert_your_openrouter_api_key_here

If you prefer to use an environment variable, set OPENROUTER_API_KEY.

For more information on the available models and API usage, refer to the OpenRouter documentation for each specific model.