From a2978fb93b03154155431968e96ac1394fbeefb7 Mon Sep 17 00:00:00 2001 From: Howard Liberty Date: Fri, 19 Apr 2024 11:41:17 -0700 Subject: [PATCH] Style --- src/transformers/training_args.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/transformers/training_args.py b/src/transformers/training_args.py index 2095f2306a352b..5e81c22db93b50 100644 --- a/src/transformers/training_args.py +++ b/src/transformers/training_args.py @@ -1838,7 +1838,7 @@ def __post_init__(self): if str(sync_module_states).lower() == "false" and str(cpu_ram_efficient_loading).lower() == "true": # In this case, all the processes except the main process would have random weights leading # to unexpected behaviour during training, thus throwing error here to prevent it. - raise ValueError("`sync_module_states` must be `\"True\"` if `cpu_ram_efficient_loading` is `\"True\"`") + raise ValueError('`sync_module_states` must be `"True"` if `cpu_ram_efficient_loading` is `"True"`') os.environ[f"{prefix}SYNC_MODULE_STATES"] = sync_module_states os.environ[f"{prefix}CPU_RAM_EFFICIENT_LOADING"] = cpu_ram_efficient_loading