From 11f273a16fbc0af313f13bf3da63014f2f58d0d0 Mon Sep 17 00:00:00 2001 From: Tom Aarsen <37621491+tomaarsen@users.noreply.github.com> Date: Mon, 15 Jan 2024 10:08:03 +0100 Subject: [PATCH] [`chore`] Update warning text, a word was missing (#28017) Update warning, a word was missing --- src/transformers/models/llama/modeling_llama.py | 4 ++-- src/transformers/models/mistral/modeling_mistral.py | 4 ++-- src/transformers/models/mixtral/modeling_mixtral.py | 4 ++-- src/transformers/models/persimmon/modeling_persimmon.py | 4 ++-- src/transformers/models/phi/modeling_phi.py | 4 ++-- 5 files changed, 10 insertions(+), 10 deletions(-) diff --git a/src/transformers/models/llama/modeling_llama.py b/src/transformers/models/llama/modeling_llama.py index b0ed793b7eeea5..bd2269ce1606b6 100644 --- a/src/transformers/models/llama/modeling_llama.py +++ b/src/transformers/models/llama/modeling_llama.py @@ -291,8 +291,8 @@ def __init__(self, config: LlamaConfig, layer_idx: Optional[int] = None): self.layer_idx = layer_idx if layer_idx is None: logger.warning_once( - f"Instantiating {self.__class__.__name__} without passing `layer_idx` is not recommended and will " - "to errors during the forward call, if caching is used. Please make sure to provide a `layer_idx` " + f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and will " + "lead to errors during the forward call if caching is used. Please make sure to provide a `layer_idx` " "when creating this class." ) diff --git a/src/transformers/models/mistral/modeling_mistral.py b/src/transformers/models/mistral/modeling_mistral.py index 5f1f85d74f1013..b0550943765034 100644 --- a/src/transformers/models/mistral/modeling_mistral.py +++ b/src/transformers/models/mistral/modeling_mistral.py @@ -202,8 +202,8 @@ def __init__(self, config: MistralConfig, layer_idx: Optional[int] = None): self.layer_idx = layer_idx if layer_idx is None: logger.warning_once( - f"Instantiating {self.__class__.__name__} without passing `layer_idx` is not recommended and will " - "to errors during the forward call, if caching is used. Please make sure to provide a `layer_idx` " + f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and will " + "lead to errors during the forward call if caching is used. Please make sure to provide a `layer_idx` " "when creating this class." ) diff --git a/src/transformers/models/mixtral/modeling_mixtral.py b/src/transformers/models/mixtral/modeling_mixtral.py index dc225a95360533..99bd2820869d93 100644 --- a/src/transformers/models/mixtral/modeling_mixtral.py +++ b/src/transformers/models/mixtral/modeling_mixtral.py @@ -246,8 +246,8 @@ def __init__(self, config: MixtralConfig, layer_idx: Optional[int] = None): self.layer_idx = layer_idx if layer_idx is None: logger.warning_once( - f"Instantiating {self.__class__.__name__} without passing `layer_idx` is not recommended and will " - "to errors during the forward call, if caching is used. Please make sure to provide a `layer_idx` " + f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and will " + "lead to errors during the forward call if caching is used. Please make sure to provide a `layer_idx` " "when creating this class." ) diff --git a/src/transformers/models/persimmon/modeling_persimmon.py b/src/transformers/models/persimmon/modeling_persimmon.py index 4bffbae38eb8d1..90d89a6d9e9f5a 100644 --- a/src/transformers/models/persimmon/modeling_persimmon.py +++ b/src/transformers/models/persimmon/modeling_persimmon.py @@ -185,8 +185,8 @@ def __init__(self, config: PersimmonConfig, layer_idx: Optional[int] = None): self.layer_idx = layer_idx if layer_idx is None: logger.warning_once( - f"Instantiating {self.__class__.__name__} without passing `layer_idx` is not recommended and will " - "to errors during the forward call, if caching is used. Please make sure to provide a `layer_idx` " + f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and will " + "lead to errors during the forward call if caching is used. Please make sure to provide a `layer_idx` " "when creating this class." ) diff --git a/src/transformers/models/phi/modeling_phi.py b/src/transformers/models/phi/modeling_phi.py index f4d227e67c86f1..d6ad4e46608eb9 100644 --- a/src/transformers/models/phi/modeling_phi.py +++ b/src/transformers/models/phi/modeling_phi.py @@ -237,8 +237,8 @@ def __init__(self, config: PhiConfig, layer_idx: Optional[int] = None): self.layer_idx = layer_idx if layer_idx is None: logger.warning_once( - f"Instantiating {self.__class__.__name__} without passing `layer_idx` is not recommended and will " - "to errors during the forward call, if caching is used. Please make sure to provide a `layer_idx` " + f"Instantiating {self.__class__.__name__} without passing a `layer_idx` is not recommended and will " + "lead to errors during the forward call if caching is used. Please make sure to provide a `layer_idx` " "when creating this class." )