-
Notifications
You must be signed in to change notification settings - Fork 372
Open
Labels
bugSomething isn't workingSomething isn't working
Description
Environment
Libtorch 2.5.0.dev (latest nightly) (built with CUDA 12.4)
CUDA 12.4
TensorRT 10.1.0.27
PyTorch 2.4.0+cu124
Torch-TensorRT 2.4.0
Python 3.12.8
Windows 10
Code
import torch
import torch_tensorrt
model = DeepLabv3+(num_classes=8, backbone='mobilenet', output_stride=16)
model.eval().cuda()
dynamic_input = torch_tensorrt.Input(
min_shape=[1, 3, 20, 20],
opt_shape=[1, 3, 512, 512],
max_shape=[1, 3, 2448, 2048],
dtype=torch.float32 )
static_input = [torch.randn((1, 3, 512, 512)).cuda()]
trt_gm = torch_tensorrt.compile(model, ir="dynamo", inputs = dynamic_input)
torch_tensorrt.save(trt_gm, "trt_dynamic.ts", output_format="torchscript",inputs=static_input)
ERROR
torch_tensorrt.compile succeeds, but torch_tensorrt.save fails.
torch_tensorrt.save(trt_gm, "trt_dynamic.ts", output_format="torchscript",inputs=static_input)
def __call__(self_, *args, **kwargs): # noqa: B902
# use `self_` to avoid naming collide with aten ops arguments that
# are named "self". This way, all the aten ops can be called by kwargs.
return self_._op(*args, **kwargs)

RuntimeError: aten::clamp() Expected a value of type 'Optional[number]' for argument 'max' but instead found type 'Tensor'.
Position: 2
Value: tensor([31], device='cuda:0')
Declaration: aten::clamp(Tensor self, Scalar? min=None, Scalar? max=None) -> Tensor
Cast error details: Cannot cast tensor([31], device='cuda:0') to number
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working