Scorers
Built-in scorers for evaluating AI outputs
Scorers define how AI outputs are evaluated. Scout provides built-in scorers for both human review (collecting feedback) and automated evaluation (programmatic checks).
Human Review Scorers
Human review scorers are used with the Feedback API to collect ratings from users.
thumbs
The thumbs scorer collects simple thumbs up/down feedback.
Example:
Native Scorers
Native scorers are used for automated evaluation runs. They programmatically check workflow outputs against expected values, enabling you to test your AI workflows at scale.
exact_match
Checks if the output exactly matches the expected value.
Config Options:
Whether the comparison should be case-sensitive.
Whether to strip leading/trailing whitespace before comparing.
Example:
contains
Checks if the output contains the expected value as a substring.
Config Options:
Whether the search should be case-sensitive.
Example:
regex_match
Checks if the output matches a regular expression pattern.
Config Options:
The regular expression pattern to match against.
Regex flags (e.g., "i" for case-insensitive).
Example:
field_exists
Checks if a specific field exists in a JSON output.
Config Options:
Dot-notation path to the field (e.g., "response.data.id").
Example:
type_check
Validates that the output is of the expected type.
Config Options:
The expected type: "string", "number", "boolean", "array", "object", or "null".
Example: