- Click
Trigger experiment
in an open playground session. - Select the prompts from the session you want to benchmark.
- Optionally, select the evaluation metrics you want to use to score the outputs of the prompts. If the evaluation metrics rely on
log.inputs
the prompt templates’ variable names should match those expected by the evaluation metric. - Select the test case collection you want to use for the run.
- Select the test cases you want to use for the run.

You can only trigger an experiment if you have uploaded a dataset to the platform.