You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I tried to use DLA-based model inference on Jetson Agx Orin Developer Kit and the speed was much lower than non-DLA model inference. I tested both the pointpillar vfe model and the model_bn model from the jetson_dla_tutorial repository.
I will show the steps of my work:
I have an additional question: 1. Regarding the gpu+dla model inference, do I need to make any special modifications to the inference code?
System environment
Device: agx orin developer kit
jatback version: 6.0
cuda: 12.2
tensorrt: 8.6.2
Model conversion:
DLA
/usr/src/tensorrt/bin/trtexec --onnx=model_bn.onnx
--shapes=input:8x3x640x640
--saveEngine=model_bn_.engine
--exportProfile=model_bn_.json
--int8 --useDLACore=0 --allowGPUFallback --useSpinWait --separateProfileRun --verbose > model_bn_.log
I tried to use DLA-based model inference on Jetson Agx Orin Developer Kit and the speed was much lower than non-DLA model inference. I tested both the pointpillar vfe model and the model_bn model from the jetson_dla_tutorial repository.
I will show the steps of my work:
I have an additional question: 1. Regarding the gpu+dla model inference, do I need to make any special modifications to the inference code?
System environment
Device: agx orin developer kit
jatback version: 6.0
cuda: 12.2
tensorrt: 8.6.2
Model conversion:
DLA
/usr/src/tensorrt/bin/trtexec --onnx=model_bn.onnx
--shapes=input:8x3x640x640
--saveEngine=model_bn_.engine
--exportProfile=model_bn_.json
--int8 --useDLACore=0 --allowGPUFallback --useSpinWait --separateProfileRun --verbose > model_bn_.log
Non-DLA
/usr/src/tensorrt/bin/trtexec --onnx=model_bn.onnx
--shapes=input:8x3x640x640
--saveEngine=model_bn_.engine
--exportProfile=model_bn_.json
--int8 --useSpinWait --separateProfileRun --verbose > model_bn_.log
Model inference code
int main(int argc, char** argv)
{
// 参数设置
const char* enginePath = "/project/ansy_dla/model_bn_.engine"; // 引擎文件路径
const int batchSize = 8; // 批处理大小
const int inputSize = 3 * 640 * 640; // 输入维度
const int outputSize = 10; // 输出维度
}
Actual reasoning time
dla:
no dla
The text was updated successfully, but these errors were encountered: