You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Follow this guide to run the evaluation framework with all the German tasks provided by this repo.
Install
# create fresh conda environment
conda create -n lm-evaluation-harness python=3.8
conda activate lm-evaluation-harness
# clone repo
git clone https://github.com/OpenGPTX/lm-evaluation-harness.git
cd lm-evaluation-harness
# change to `german` branch
git checkout german
# install dependencies
pip install -r requirements.txt
# set environment variables (optional)
export DATASETS_DIR="./data" # TODO replace with your own path
export DATASETS_DIR="/data/datasets" # TODO replace with your own path
export CUDA_VISIBLE_DEVICES=1 # TODO change to your GPU id
export TRANSFORMERS_CACHE="${DATASETS_DIR}/transformers_cache"
export HF_DATASETS_CACHE="${DATASETS_DIR}/hf_datasets_cache"
export HF_DATASETS_OFFLINE=0
export TRANSFORMERS_OFFLINE=0
Follow this guide to run the evaluation framework with all the German tasks provided by this repo.
Install
Run evaluations
Evaluate all German tasks on
gpt2-xl-wechsel-german
(HF implementation).Arguments:
--no_tokenizer_check
is required if you want to run the evaluation with a custom / non-English tokenizer.--limit <int>
to run the script only on a subset of the test samples.--batch_size <int>
size of batch per GPUThe output should look as follows:
To run only a specific set of tasks, execute this command:
You can also evaluate model checkpoints from disk:
Run English GPT2:
The text was updated successfully, but these errors were encountered: