Skip to main content

JFrog ML

The JFrog ML provider (formerly known as Qwak) allows you to interact with JFrog ML's LLM Model Library using the OpenAI protocol. It supports chat completion models hosted on JFrog ML's infrastructure.

Setup

To use the JFrog ML provider, you'll need:

  1. A JFrog ML account
  2. A JFrog ML token for authentication
  3. A deployed model from the JFrog ML Model Library

Set up your environment:

export QWAK_TOKEN="your-token-here"

Basic Usage

Here's a basic example of how to use the JFrog ML provider:

promptfooconfig.yaml
providers:
- id: jfrog:llama_3_8b_instruct
config:
temperature: 1.2
max_tokens: 500

You can also use the legacy qwak: prefix:

promptfooconfig.yaml
providers:
- id: qwak:llama_3_8b_instruct

Configuration Options

The JFrog ML provider supports all the standard OpenAI configuration options plus these additional JFrog ML-specific options:

ParameterDescription
baseUrlOptional. The full URL to your model endpoint. If not provided, it will be constructed using the model name: https://models.qwak-prod.qwak.ai/v1

Example with full configuration:

promptfooconfig.yaml
providers:
- id: jfrog:llama_3_8b_instruct
config:
# JFrog ML-specific options
baseUrl: https://models.qwak-prod.qwak.ai/v1

# Standard OpenAI options
temperature: 1.2
max_tokens: 500
top_p: 1
frequency_penalty: 0
presence_penalty: 0

Environment Variables

The following environment variables are supported:

VariableDescription
QWAK_TOKENThe authentication token for JFrog ML API access

API Compatibility

The JFrog ML provider is built on top of the OpenAI protocol, which means it supports the same message format and most of the same parameters as the OpenAI Chat API. This includes:

  • Chat message formatting with roles (system, user, assistant)
  • Temperature and other generation parameters
  • Token limits and other constraints

Example chat conversation:

prompts.yaml
- role: system
content: 'You are a helpful assistant.'
- role: user
content: '{{user_input}}'
promptfooconfig.yaml
prompts:
- file://prompts.yaml

providers:
- id: jfrog:llama_3_8b_instruct
config:
temperature: 1.2
max_tokens: 500

tests:
- vars:
user_input: 'What should I do for a 4 day vacation in Spain?'