You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/_dynamo/variables/builder.py", line 2408, in handle_traced_output
wrap_fx_proxy_cls(
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/_dynamo/variables/builder.py", line 2230, in wrap_fx_proxy_cls
return handle_traced_output(
^^^^^^^^^^^^^^^^^^^^^
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/_dynamo/variables/builder.py", line 2517, in handle_traced_output
unimplemented(
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/_dynamo/exc.py", line 317, in unimplemented
raise Unsupported(msg, case_name=case_name)
torch._dynamo.exc.Unsupported: torch.* op returned non-Tensor device call_function <built-in function getitem>
from user code:
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/hooks.py", line 170, in new_forward
output = module._old_forward(*args, **kwargs)
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/transformers/models/llama/modeling_llama.py", line 834, in forward
outputs = self.model(
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/transformers/models/llama/modeling_llama.py", line 592, in forward
layer_outputs = decoder_layer(
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1750, in _call_impl
return forward_call(*args, **kwargs)
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/hooks.py", line 165, in new_forward
args, kwargs = module._hf_hook.pre_forward(module, *args, **kwargs)
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/hooks.py", line 364, in pre_forward
return send_to_device(args, self.execution_device), send_to_device(
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/utils/operations.py", line 183, in send_to_device
{
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/utils/operations.py", line 184, in <dictcomp>
k: t if k in skip_keys else send_to_device(t, device, non_blocking=non_blocking, skip_keys=skip_keys)
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/accelerate/utils/operations.py", line 155, in send_to_device
return tensor.to(device, non_blocking=non_blocking)
File "/home/vasiliy/.conda/envs/pytorch_2_6/lib/python3.11/site-packages/torch/nn/modules/module.py", line 1302, in to
device, dtype, non_blocking, convert_to_format = torch._C._nn._parse_to(
Set TORCH_LOGS="+dynamo" and TORCHDYNAMO_VERBOSE=1 for more information
You can suppress this exception and fall back to eager by setting:
import torch._dynamo
torch._dynamo.config.suppress_errors = True
The text was updated successfully, but these errors were encountered:
When I run the code snippet from https://huggingface.co/docs/transformers/main/en/quantization/torchao, I see a torch.compile error.
torch version: 2.6.0
torchao version: 0.8.0
transformers version: 4.48.3
Note that I see the same error when running this example even if I remove the quantiazation_config argument to
AutoModelForCausalLM.from_pretrained
.Repro script: https://gist.github.com/vkuzo/4dc367370246dc8e47b77f8f1bd92901
Error: https://gist.github.com/vkuzo/814a452a024c27a2f5b2e3fb7276dec3
The text was updated successfully, but these errors were encountered: