Define evaluation functions to test the outputs of your prompts, chains & agents via Python functions or webhooks. Upload a CSV file with your test cases as a test case collection to use them in the Lab and in the Prompt IDE as you iterate on your prompts, chains & agents. Finally, benchmark your prompts, chains & agents on a test case collection to find the best performing ones.