Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

TensorRT EP could not deserialize engine from binary data #22139

Open
adaber opened this issue Sep 18, 2024 · 0 comments
Open

TensorRT EP could not deserialize engine from binary data #22139

adaber opened this issue Sep 18, 2024 · 0 comments
Labels
api:CSharp issues related to the C# API ep:TensorRT issues related to TensorRT execution provider performance issues related to performance regressions

Comments

@adaber
Copy link

adaber commented Sep 18, 2024

Describe the issue

Hi,

I've wrapped a TensorRT engine in an _ctx.onnx file using an official python script (https://github.com/microsoft/onnxruntime/blob/main/onnxruntime/python/tools/tensorrt/gen_trt_engine_wrapper_onnx_model.py#L156-L187)

The problem is that I get the "TensorRT EP could not deserialize engine from binary data" error. The TensorRT model works well using the TensorRT API. I am kind of stuck since there is no other information to help me figure out why this happens.

I've tried using different ortTrtOptions but to no avail.

This error occurs when creating an inference session. I tried both the FP16 and INT8 version and I got the same error.

I've uploaded the FP16 version and it'd be great if you have time to look at it.

Thanks!

Edit:

Graphics Card: 3090

The trt engine was built using the following profile shapes:
min: 1x1024x128x3
opt: 1x4096x640x3
max: 1x8000x1400x3

To reproduce

EmbededTrtEngine_FP16_ctx.zip

Urgency

Both a workaround or a fix would help.

Platform

Windows

OS Version

10

ONNX Runtime Installation

Released Package

ONNX Runtime Version or Commit ID

1.19.0

ONNX Runtime API

C#

Architecture

X64

Execution Provider

TensorRT

Execution Provider Library Version

CUDA 11.8, CuDNN 8.9.7.29, TRT 10.4.0.26 and 10.1.0.27

Model File

EmbededTrtEngine_FP16_ctx.zip

Is this a quantized model?

No

@adaber adaber added the performance issues related to performance regressions label Sep 18, 2024
@github-actions github-actions bot added api:CSharp issues related to the C# API ep:TensorRT issues related to TensorRT execution provider labels Sep 18, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
api:CSharp issues related to the C# API ep:TensorRT issues related to TensorRT execution provider performance issues related to performance regressions
Projects
None yet
Development

No branches or pull requests

1 participant