Description
To improve usability and reproducibility, this repository should include Jupyter notebooks that demonstrate how to run evaluations using the refactored YESciEval-based pipeline.
Tasks
Acceptance Criteria
- Two notebooks are added and committed
- Notebooks run successfully using the refactored evaluation pipeline
- Aggregated metrics are computed and visualized for report collections
- Notebooks serve as clear reference examples for users and evaluators
Rationale
These notebooks act as executable documentation and lower the barrier for:
- Reproducibility
- Qualitative analysis
- Method comparison
- Adoption of YESciEval-based evaluation workflows
Description
To improve usability and reproducibility, this repository should include Jupyter notebooks that demonstrate how to run evaluations using the refactored YESciEval-based pipeline.
Tasks
evaluation_single_report.ipynbevaluation_collection.ipynbAcceptance Criteria
Rationale
These notebooks act as executable documentation and lower the barrier for: