Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Are some layers implemented in DFINE and DEIM which aren't supported? #4375

Open
horngjason opened this issue Mar 5, 2025 · 3 comments
Open
Assignees
Labels
Module:Engine Build Issues with engine build triaged Issue has been triaged by maintainers waiting for feedback Requires more information from user to make progress on the issue.

Comments

@horngjason
Copy link

horngjason commented Mar 5, 2025

Hi all,

While I was playing with DFINE and DEIM models, I couldn't convert the model from .pth -> ONNX -> TRT INT8. Has anyone met the same situation?

It's successful to convert the model to FP16 or FP8. Yet, once I set INT8 in the trtexec, what I got in the end would be a "FAILED TensorRT.....".

Image

Failed_profile_fp16int8.txt

Possible solutions I tried are:

  • Use simplified ONNX model
  • Set ONNX model in opset17 instead opset16
  • Set allocationStrategy=runtime in trtexec
    Neither of them solved my case.

My environment:

  • TensorRT: 10.5, 10.6.0.26 and 10.8.0.43
  • CUDA: 11.8
  • DFINE model type: dfine_hgnetv2_n_coco.yml

Many thanks!

@kevinch-nv
Copy link
Collaborator

Are you able to share your ONNX model? What GPU are you running on?

@kevinch-nv kevinch-nv added Module:Engine Build Issues with engine build triaged Issue has been triaged by maintainers waiting for feedback Requires more information from user to make progress on the issue. labels Mar 7, 2025
@kevinch-nv kevinch-nv self-assigned this Mar 7, 2025
@horngjason
Copy link
Author

horngjason commented Mar 11, 2025

Hi @kevinch-nv, thanks for the reply.

I'm testing on a RTX4060 Laptop GPU (on ASUS TUF GAMING A14).

Here are the official model that I used their official converter without simplifying: https://drive.google.com/drive/folders/1vMn1aqMOnS2w_W7FJKNaVbt72TukxfDR?usp=drive_link. They are the smallest two versions of DFINE and are in OPSET17.

I've tried to add arguments such as
● trtexec --onnx=./dfine/opset17/dfine_s_coco.onnx --saveEngine=./dfine/opset17/dfine_s_coco.engine --[Switch between fp16 or int8] --stronglyType --verbose
● trtexec --onnx=./dfine/opset17/dfine_s_coco.onnx --saveEngine=./dfine/opset17/dfine_s_coco_int8.trt --[Switch between fp16 or int8] --precisionConstraints=prefer --layerPrecisions=*:fp16 --verbose

@lix19937
Copy link

lix19937 commented Mar 21, 2025

Maybe a bug, would you like to try the latest version to check if the error still occurs?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Module:Engine Build Issues with engine build triaged Issue has been triaged by maintainers waiting for feedback Requires more information from user to make progress on the issue.
Projects
None yet
Development

No branches or pull requests

3 participants