LLM as judge
True/False
The True/False Evaluator is a specialized tool designed to assess responses based on strict pattern matching against provided descriptions. It uses an LLM (Large Language Model) as a judge to determine if a response matches the required patterns for being considered “true” while avoiding any patterns that would make it “false”.
Purpose
The True/False Evaluator is particularly useful when you need to:
- Verify strict compliance with specific content patterns
- Check for the presence or absence of particular phrases or content
- Evaluate responses against predefined criteria
- Ensure responses meet exact pattern requirements
- Handle cases where responses need to be evaluated based on literal matches
How It Works
The evaluator uses a binary scoring system:
- Score: 0 (True): Response must satisfy ALL elements of the True_description AND contain ZERO elements from the False_description
- Score: 1 (False): Response matches ANY aspect of the False_description OR fails to fully match the True_description
The evaluator performs literal pattern matching and does not consider real-world accuracy or ethical implications. It also automatically handles refusal patterns when the True_description requires harmful content.
Usage Example
The evaluator returns a tuple containing:
- A score (0 or 1) indicating whether the response is true or false
- A list of explanations for the given score, including specific text matches or mismatches
When to Use
Use the True/False Evaluator when you need to:
- Evaluate adversarial/malicious responses.
- Verify strict compliance with content requirements
- Check for specific pattern matches in responses
- Evaluate responses against predefined criteria
- Handle cases requiring literal pattern matching
- Assess responses that need to meet exact content requirements