You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
[rank3]: Traceback (most recent call last):
[rank3]: File "/mnt/nfs4/nlp/ft_llms/finetuning.py", line 299, in<module>
[rank3]: fire.Fire(main)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/fire/core.py", line 143, in Fire
[rank3]: component_trace = _Fire(component, args, parsed_flag_args, context, name)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/fire/core.py", line 477, in _Fire
[rank3]: component, remaining_args = _CallAndUpdateTrace(
[rank3]: File "/opt/conda/lib/python3.10/site-packages/fire/core.py", line 693, in _CallAndUpdateTrace
[rank3]: component = fn(*varargs, **kwargs)
[rank3]: File "/mnt/nfs4/nlp/ft_llms/finetuning.py", line 123, in main
[rank3]: model = LlamaForCausalLM(llama_config)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 1067, in __init__
[rank3]: self.model = LlamaModel(config)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 846, in __init__
[rank3]: [LlamaDecoderLayer(config, layer_idx) forlayer_idxin range(config.num_hidden_layers)]
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 846, in<listcomp>
[rank3]: [LlamaDecoderLayer(config, layer_idx) forlayer_idxin range(config.num_hidden_layers)]
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 633, in __init__
[rank3]: self.self_attn = LLAMA_ATTENTION_CLASSES[config._attn_implementation](config=config, layer_idx=layer_idx)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 306, in __init__
[rank3]: self.rotary_emb = LlamaRotaryEmbedding(config=self.config)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/models/llama/modeling_llama.py", line 119, in __init__
[rank3]: inv_freq, self.attention_scaling = self.rope_init_fn(self.config, device, **self.rope_kwargs)
[rank3]: File "/opt/conda/lib/python3.10/site-packages/transformers/modeling_rope_utils.py", line 330, in _compute_llama3_parameters
[rank3]: if wavelen < high_freq_wavelen:
[rank3]: File "/opt/conda/lib/python3.10/site-packages/torch/utils/_device.py", line 78, in __torch_function__
[rank3]: return func(*args, **kwargs)
[rank3]: NotImplementedError: aten::_local_scalar_dense: attempted to run this operator with Meta tensors, but there was no abstract impl or
Meta kernel registered. You may have run into this message while using an operator with PT2 compilation APIs (torch.compile/torch.export);in order to use this operator with those APIs you'll need to add an abstract impl.Please see the following doc for next steps: https://docs.google.com/document/d/1_W62p8WJOQQUzPsJYa7s701JXt0qf2OfLub2sbkHOaU/edit
Expected behavior
We expect that llama3 will be successfully generated on PyTorch's meta device type as well.
In the process of creating the rope embedding for llama3.1, the _compute_llama3_parameters function from modeling_repe_utils.py is called. However, this function does not work correctly on the meta device type.
During this process, a NotImplementedError occurs as follows.
NotImplementedError: aten::_local_scalar_dense: attempted to run this operator with Meta tensors, but there was no abstract impl or
Meta kernel registered. You may have run into this message while using an operator with PT2 compilation APIs (torch.compile/torch.export);in order to use this operator with those APIs you'll need to add an abstract impl.Please see the following doc for next steps: https://docs.google.com/document/d/1_W62p8WJOQQUzPsJYa7s701JXt0qf2OfLub2sbkHOaU/edit
The related commit hash was merged one day ago, and the commit hash is as follows. d5a99df
The text was updated successfully, but these errors were encountered:
System Info
transformers
version: 4.44.0.dev0Who can help?
@ArthurZucker
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
code
error messages
Expected behavior
We expect that llama3 will be successfully generated on PyTorch's meta device type as well.
In the process of creating the rope embedding for llama3.1, the
_compute_llama3_parameters
function from modeling_repe_utils.py is called. However, this function does not work correctly on the meta device type.During this process, a NotImplementedError occurs as follows.
The related commit hash was merged one day ago, and the commit hash is as follows.
d5a99df
The text was updated successfully, but these errors were encountered: