Eval Engine

Overview

The Eval Engine is used to evaluate the performance of an LLM's output against a given input. The engine returns a score between 0 and 1, along with a reason for the score.

Eval Engines require either criteria or evaluation_steps to be set. If criteria is set, Gen AI Builder will generate evaluation_steps for you. This is useful for getting started, but you may to explicitly set evaluation_steps for more complex evaluations. Either criteria or evaluation_steps must be set, but not both.


Could this page be better? Report a problem or suggest an addition!