Skip to main content

Command line

The promptfoo command line utility supports the following subcommands:

  • init [directory] - Initialize a new project with dummy files.
  • eval - Evaluate prompts and models. This is the command you'll be using the most!
  • view - Start a browser UI for visualization of results.
  • share - Create a URL that can be shared online.
  • cache - Manage cache.
    • cache clear
  • list - List various resources like evaluations, prompts, and datasets.
    • list evals
    • list prompts
    • list datasets
  • show <id> - Show details of a specific resource (evaluation, prompt, dataset).
  • feedback <message> - Send feedback to the Promptfoo developers.

promptfoo eval

By default the eval command will read the promptfooconfig.yaml configuration file in your current directory. But, if you're looking to override certain parameters you can supply optional arguments:

OptionDescription
-p, --prompts <paths...>Paths to prompt files, directory, or glob
-r, --providers <name or path...>openai:chat, openai:completion, localai:chat:<model-name>, localai:completion:<model-name>, or one of the many other permutations per API providers
-o, --output <paths...>Path to output file (csv, json, yaml, html)
-t, --tests <path>Path to external test file
-c, --config <path>Path to one or more configuration files. promptfooconfig.js/json/yaml is automatically loaded if present. Wildcards and directories are supported.
--graderProvider that will conduct the evaluation, if you are using LLM to grade your output
--repeat <number>Number of times to repeat each test case. Disables cache if >1
--delay <number>Force the test runner to wait after each API call (milliseconds)
--no-cacheDisable cache
--no-tableDisable CLI table output
--no-progress-barDisable the progress bar
--no-writeDo not write the latest config to disk (used for web viewer and sharing)
--prompt-prefix <path>This prefix is prepended to every prompt
--prompt-suffix <path>This suffix is append to every prompt
--shareAutomatically create a share link
--table-cell-max-length <number>Truncate console table cells to this length
--verboseShow debug logs
--watchWatch the config and prompt files for changes
-j, --max-concurrency <number>Maximum number of concurrent API calls
--env-filePath to env file (defaults to .env)
--interactive-providersRun 1 provider at a time and prompt user to continue
--filter-failing <path>Run only failing tests from previous evaluation. Path to JSON output file from the previous evaluation.
-n, --filter-first-nRun the first N test cases
--filter-pattern <pattern>Run only test cases whose description matches the regex pattern

The eval command will return exit code 100 when there is at least 1 test case failure. It will return exit code 1 for any other error. The exit code for failed tests can be overridden with environment variable PROMPTFOO_FAILED_TEST_EXIT_CODE.

promptfoo init [directory]

Initialize a new project with dummy files.

OptionDescription
directoryDirectory to create files in

promptfoo view

Start a browser UI for visualization of results.

OptionDescription
-p, --port <number>Port number for the local server
-y, --yesSkip confirmation and auto-open the URL

If you've used PROMPTFOO_CONFIG_DIR to override the promptfoo output directory, run promptfoo view [directory].

promptfoo share

Create a URL that can be shared online.

OptionDescription
-y, --yesSkip confirmation before creating shareable URL

promptfoo cache

Manage cache.

OptionDescription
clearClear the cache

promptfoo feedback <message>

Send feedback to the promptfoo developers.

OptionDescription
messageFeedback message

promptfoo list

List various resources like evaluations, prompts, and datasets.

SubcommandDescription
evalsList evaluations
promptsList prompts
datasetsList datasets
OptionDescription
-nShow the first n records, sorted by descending date of creation

promptfoo show <id>

Show details of a specific resource.

OptionDescription
eval <id>Show details of a specific evaluation
prompt <id>Show details of a specific prompt
dataset <id>Show details of a specific dataset

promptfoo import <filepath>

Import an eval file from JSON format.

promptfoo export

Export an eval record to JSON format. Outputs to stdout by default.

OptionDescription
-o, --output <filepath>File to write

Environment variables

These general-purpose environment variables are supported:

NameDescriptionDefault
PROMPTFOO_CONFIG_DIRDirectory that stores eval history~/.promptfoo
PROMPTFOO_DISABLE_CONVERSATION_VARPrevents the _conversation variable from being set
PROMPTFOO_DISABLE_JSON_AUTOESCAPEIf set, disables smart variable substitution within JSON prompts
PROMPTFOO_DISABLE_REF_PARSERPrevents JSON schema dereferencing
PROMPTFOO_DISABLE_TEMPLATINGDisable Nunjucks rendering
PROMPTFOO_DISABLE_VAR_EXPANSIONPrevents Array-type vars from being expanded into multiple test cases
PROMPTFOO_ASSERTIONS_MAX_CONCURRENCYHow many assertions to run at a time3
tip

promptfoo will load environment variables from the .env in your current working directory.

promptfoo generate dataset

BETA: Generate synthetic test cases based on existing prompts and variables.

OptionDescriptionDefault
-c, --config <path>Path to the configuration filepromptfooconfig.yaml
-w, --writeWrite the generated test cases directly to the config filefalse
-i, --instructions <text>Custom instructions for test case generation
-o, --output <path>Path to write the generated test casesstdout
--numPersonas <number>Number of personas to generate5
--numTestCasesPerPersona <number>Number of test cases per persona3

For example, this command will modify your default config file (usually promptfooconfig.yaml) with new test cases:

promptfoo generate dataset -w

This command will generate test cases for a specific config and write them to a file, while following special instructions:

promptfoo generate dataset -c my_config.yaml -o new_tests.yaml -i 'All test cases for {{location}} must be European cities'
info

While in beta, this feature depends on OpenAI and requires the OPENAI_API_KEY environment variable.