Evaluates given input, output, or messages using Qualifire’s detectors. Supports checks for hallucinations, grounding, PII, prompt injections, content moderation, policy assertions, topic scoping, and tool use quality.
Evaluation request
At least one of input, output, or messages must be provided.
The input text (e.g., user prompt)
The output text (e.g., LLM response)
Conversation messages for multi-turn evaluation
List of tools available to the LLM. Required when tool_use_quality_check is enabled.
Check for factual inaccuracies or hallucinations
Quality/speed tradeoff for hallucination checks
speed, balanced, quality Check if the output is grounded in the provided input/context
Quality/speed tradeoff for grounding checks
speed, balanced, quality Use full conversation context for grounding checks
Check for prompt injection attempts
Check for personally identifiable information
Check for harmful content (dangerous content, harassment, hate speech, sexual content)
Check the quality of tool selection and usage. Requires messages and available_tools.
Quality/speed tradeoff for tool use quality checks
speed, balanced, quality Custom policy assertions to evaluate against (e.g., "don't give medical advice")
Quality/speed tradeoff for assertion checks
speed, balanced, quality Whether policy assertions apply to input, output, or both
input, output, both Use full conversation context for policy checks
Include tool definitions and tool calls in policy assertion context. When true, assertions can reference available_tools and tool_calls from messages.
List of allowed topics for topic scoping checks
Quality/speed tradeoff for topic scoping checks
speed, balanced, quality Whether topic scoping applies to input, output, or both
input, output, both Use full conversation context for topic scoping checks
Optional key-value pairs (string values only) to attach to the evaluation invocation