From 063fdd358adc8a71c2462041e91c495a613bfcd6 Mon Sep 17 00:00:00 2001 From: Zach Mueller Date: Thu, 25 Apr 2024 11:13:19 -0400 Subject: [PATCH] Comment --- src/transformers/training_args.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/transformers/training_args.py b/src/transformers/training_args.py index 7574c1c0536aa3..024e59dc870523 100644 --- a/src/transformers/training_args.py +++ b/src/transformers/training_args.py @@ -1619,7 +1619,7 @@ def __post_init__(self): if version.parse(version.parse(torch.__version__).base_version) == version.parse("2.0.0") and self.fp16: raise ValueError("--optim adamw_torch_fused with --fp16 requires PyTorch>2.0") - # We need to setup the accelerator config here + # We need to setup the accelerator config here *before* the first call to `self.device` if is_accelerate_available(): if not isinstance(self.accelerator_config, (AcceleratorConfig)): if self.accelerator_config is None: