-
Notifications
You must be signed in to change notification settings - Fork 213
Open
Labels
bugSomething isn't workingSomething isn't workinginternal bug filedInternal nvbug has been filedInternal nvbug has been filedinvestigating
Description
Before submitting an issue, please make sure it hasn't been already addressed by searching through the existing and past issues.
Describe the bug
https://github.com/NVIDIA/TensorRT-Model-Optimizer/blob/main/modelopt/torch/opt/plugins/megatron.py#L80
When I try to use NeMo to load model for finetuning after PTQ, I got errors in this line.
I have no idea what this line, but I have tested in locally and it fails
import torch
import pickle
x = torch.randint(0, 10, (2,))
y = pickle.loads(x.numpy().tobytes())- ?
Steps/Code to reproduce bug
- ?
Expected behavior
Who can help?
- ?
System information
- Container used (if applicable): ?
- OS (e.g., Ubuntu 22.04, CentOS 7, Windows 10): ?
- CPU architecture (x86_64, aarch64): ?
- GPU name (e.g. H100, A100, L40S): ?
- GPU memory size: ?
- Number of GPUs: ?
- Library versions (if applicable):
- Python: ?
- ModelOpt version or commit hash: ?
- CUDA: ?
- PyTorch: ?
- Transformers: ?
- TensorRT-LLM: ?
- ONNXRuntime: ?
- TensorRT: ?
- Any other details that may help: ?
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't workinginternal bug filedInternal nvbug has been filedInternal nvbug has been filedinvestigating