This repository contains the tools and results for evaluating AI generated summaries.
To use this repository, first you will need to set up a virtual environment.
In the root of the repository run:
python -m venv venv
On Windows run:
myenv\Scripts\activate
On mac/OS/Linux run:
source myenv/bin/activate
Within your virtual environment run:
pip install -r "requirements.txt"
This repository contains example notebooks in the example_notebooks
folder. These examples implement LLM-as-a-Judge and demonstrate the theory behind LLM-as-a-Judge.
The augmentation
folder allows you to apply basic augmentations to prompts such as typos, misspellings and poor punctuation.
This let's you test the quality of summaries using poor quality prompts.
- Will Poulett