Skip to content

Commit 7325dac

Browse files
authored
Merge branch 'main' into pre-commit-ci-update-config
2 parents 18812f9 + fcf41a9 commit 7325dac

5 files changed

Lines changed: 515 additions & 227 deletions

File tree

implementations/report_generation/README.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -92,9 +92,9 @@ uv run --env-file .env python -m implementations.report_generation.evaluate --da
9292
```
9393

9494
This script will run the Report Generation Agent against each element of the dataset
95-
and then use an LLM-as-a-judge Evaluator Agent to evaluate each result. The evaluator
96-
will check the data generated to produce the report against the ground truth and
97-
produce a True/False score along with a reasoning.
95+
and then use LLM-as-a-judge Evaluator Agents to evaluate each result. The evaluator
96+
will check the data generated to produce the report and also the trajectory the
97+
agent used against the ground truth and produce True/False scores along with a reasoning.
9898

9999
At the end of the run, an evaluation report will be displayed along with a link
100100
to check details about the evaluation in Langfuse.

0 commit comments

Comments
 (0)