Open
Description
Description
When loading an eval dataset from a yaml file, it would be nice to also provide metadata to a specific eval if you are performing for example multiple LLMJudge evals.
The PDF data I'm trying to evaluate contains codes with texts which falls in different categories. The textsnippets are lengthy, so a Contains won't do it.
example:
101010:
- Header: textsnippet A
- Subtext: textsnippet B
cases:
- name: 101010
inputs:
prompt: "What is the data under code 101010
evaluators:-
LLMJudge:
rubric: "Answer should mention textsnippet A"
model: azure:gpt-4o
metadata:
type: header -
LLMJudge:
rubric: "Answer should mention tetsnippet B"
model: azure:gpt-4o
metadata:
type: subtext
-
References
No response