JFrog ML
The JFrog ML provider (formerly known as Qwak) allows you to interact with JFrog ML's LLM Model Library using the OpenAI protocol. It supports chat completion models hosted on JFrog ML's infrastructure.
Setup
To use the JFrog ML provider, you'll need:
- A JFrog ML account
- A JFrog ML token for authentication
- A deployed model from the JFrog ML Model Library
Set up your environment:
export QWAK_TOKEN="your-token-here"
Basic Usage
Here's a basic example of how to use the JFrog ML provider:
providers:
- id: jfrog:llama_3_8b_instruct
config:
temperature: 1.2
max_tokens: 500
You can also use the legacy qwak:
prefix:
providers:
- id: qwak:llama_3_8b_instruct
Configuration Options
The JFrog ML provider supports all the standard OpenAI configuration options plus these additional JFrog ML-specific options:
Parameter | Description |
---|---|
baseUrl | Optional. The full URL to your model endpoint. If not provided, it will be constructed using the model name: https://models.qwak-prod.qwak.ai/v1 |
Example with full configuration:
providers:
- id: jfrog:llama_3_8b_instruct
config:
# JFrog ML-specific options
baseUrl: https://models.qwak-prod.qwak.ai/v1
# Standard OpenAI options
temperature: 1.2
max_tokens: 500
top_p: 1
frequency_penalty: 0
presence_penalty: 0
Environment Variables
The following environment variables are supported:
Variable | Description |
---|---|
QWAK_TOKEN | The authentication token for JFrog ML API access |
API Compatibility
The JFrog ML provider is built on top of the OpenAI protocol, which means it supports the same message format and most of the same parameters as the OpenAI Chat API. This includes:
- Chat message formatting with roles (system, user, assistant)
- Temperature and other generation parameters
- Token limits and other constraints
Example chat conversation:
- role: system
content: 'You are a helpful assistant.'
- role: user
content: '{{user_input}}'
prompts:
- file://prompts.yaml
providers:
- id: jfrog:llama_3_8b_instruct
config:
temperature: 1.2
max_tokens: 500
tests:
- vars:
user_input: 'What should I do for a 4 day vacation in Spain?'