From 492109ee6ba61930578fc64e52f483bca0580744 Mon Sep 17 00:00:00 2001 From: Cemberk Date: Thu, 5 Sep 2024 09:19:29 -0500 Subject: [PATCH] Update fork-maintenance-docker.yml --- .github/workflows/fork-maintenance-docker.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/fork-maintenance-docker.yml b/.github/workflows/fork-maintenance-docker.yml index deeba59f63ad9b..45c71186383405 100644 --- a/.github/workflows/fork-maintenance-docker.yml +++ b/.github/workflows/fork-maintenance-docker.yml @@ -52,7 +52,7 @@ jobs: pip install huggingface_hub datasets ; pip install parameterized unit_test_command: > - cd tests; folders=($(python3 -c 'import os; tests = os.getcwd(); model_tests = os.listdir(os.path.join(tests, "models")); d1 = sorted(list(filter(os.path.isdir, os.listdir(tests)))); d2 = sorted(list(filter(os.path.isdir, [f"models/{x}" for x in model_tests]))); d1.remove("models"); d = d2 + d1; print("\n".join(d))')); cd ..; for folder in "${folders[@]}"; do pytest tests/${folder} -v --make-reports="huggingface_unit_tests_run_models_gpu_${folder}" -rfEs --continue-on-collection-errors -m "not not_device_test" -m "not multi_gpu" -p no:cacheprovider; done; allstats=($(find reports -name stats.txt)); for stat in "${allstats[@]}"; do echo "$stat"; cat $stat; done + cd tests; folders=($(python3 -c 'import os; tests = os.getcwd(); model_tests = os.listdir(os.path.join(tests, "models")); d1 = sorted(list(filter(os.path.isdir, os.listdir(tests)))); d2 = sorted(list(filter(os.path.isdir, [f"models/{x}" for x in model_tests]))); d1.remove("models"); d = d2 + d1; print("\n".join(d))')); cd ..; for folder in "${folders[@]}"; do pytest tests/${folder} -v --make-reports="huggingface_unit_tests_run_models_gpu_${folder}" --continue-on-collection-errors -m "not not_device_test" -p no:cacheprovider; done; allstats=($(find reports -name stats.txt)); for stat in "${allstats[@]}"; do echo "$stat"; cat $stat; done performance_test_command: > python examples/pytorch/language-modeling/run_mlm.py --model_name_or_path bert-base-uncased --dataset_name wikitext --dataset_config_name wikitext-2-raw-v1 --do_train --do_eval --output_dir /tmp/test-mlm --per_device_train_batch_size 8 --per_device_eval_batch_size 8 --max_steps 500