From a694da55db259f3409706eaeaf18e44dfde703db Mon Sep 17 00:00:00 2001 From: Patrick von Platen Date: Mon, 2 Oct 2023 18:29:08 +0200 Subject: [PATCH] [Logging] Change warning to info --- src/transformers/modeling_utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/transformers/modeling_utils.py b/src/transformers/modeling_utils.py index e568f324f9a3..a548b019d83b 100644 --- a/src/transformers/modeling_utils.py +++ b/src/transformers/modeling_utils.py @@ -1614,7 +1614,7 @@ def _get_resized_embeddings( new_num_tokens = old_embeddings.weight.shape[0] new_num_tokens = ((new_num_tokens + pad_to_multiple_of - 1) // pad_to_multiple_of) * pad_to_multiple_of else: - logger.warning( + logger.info( "You are resizing the embedding layer without providing a `pad_to_multiple_of` parameter. This means that the new embedding" f" dimension will be {new_num_tokens}. This might induce some performance reduction as *Tensor Cores* will not be available." " For more details about this, or help on choosing the correct value for resizing, refer to this guide:"