-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unstable inference time #4382
Comments
Nobody here? |
Use the same inputs or different inputs ? Or you can use trtexec to profile it . |
The same inputs. |
Set warmup, and data copy not include the latency. Lock the gpu freq. |
The warm-up phase had already been integrated into the process. Moreover, the GPU frequency had been set to its maximum level. And data copy latency was not taken into consideration. Only the time of enqueueV3 was recorded, yet these times were unstable. |
Hi!
I run yolo11 model inference for 1000 times in Tesla T4, but I found the time cost was very unstable.
From the cached records, I found most of the time cost was even and looked normal, but the normal ones were always interleaved with a few abnormal ones. For example, of the 1000 times inference records, most of the inference time cost was 2ms per image, but there were a few that cost 70ms per image.
I had tried to set a fixed card frequency but it didnt seem to work.
So can you help me with that? Thanks!
The text was updated successfully, but these errors were encountered: