Add support for JetPack 6.2 build#3453
Conversation
narendasan
left a comment
There was a problem hiding this comment.
@johnnynunez Thanks for opening this, this is very helpful. We will test it internally and provide additional feedback
| # install pytorch from nvidia jetson distribution: https://developer.download.nvidia.com/compute/redist/jp/v61/pytorch | ||
| python -m pip install torch https://developer.download.nvidia.com/compute/redist/jp/v61/pytorch/torch-2.5.0a0+872d972e41.nv24.08.17622132-cp310-cp310-linux_aarch64.whl | ||
| # install pytorch from nvidia jetson distribution: https://pypi.jetson-ai-lab.dev/jp6/cu126/ | ||
| pip3 install torch torchvision torchaudio --index-url https://pypi.jetson-ai-lab.dev/jp6/cu126/ |
There was a problem hiding this comment.
Looking at the index, there are jp62 builds for PyTorch 2.7.0 and CUDA 12.8? What are the rules behind these builds / Jetson compute stack if you know? My understanding was jp62 was CUDA 12.6 / TensorRT 10.3
There was a problem hiding this comment.
if you look inside index you can see pytorch 2.6.0 stack:
for jetpack 6 is 12.6, jetpack 7 is coming with ubuntu 24.04 so this is why we move everything to ubuntu 24.04
https://pypi.jetson-ai-lab.dev/jp6/cu126
Pytorch index for ubuntu 24.04: https://pypi.jetson-ai-lab.dev/jp6/cu128/+simple/torch/
|
|
||
|
|
||
| Since torch_tensorrt version has dependencies on torch version. torch version supported by JetPack6.1 is from DLFW 24.08/24.09(torch 2.5.0). | ||
| Since torch_tensorrt version has dependencies on torch version. torch version supported by JetPack6.2 is from DLFW 24.08/24.09(torch 2.6.0). |
There was a problem hiding this comment.
@apbose can you update this with appropriate numbers? also replace DLFW with NVIDIA NGC
| Since torch_tensorrt version has dependencies on torch version. torch version supported by JetPack6.2 is from DLFW 24.08/24.09(torch 2.6.0). | ||
|
|
||
| Please make sure to build torch_tensorrt wheel file from source release/2.5 branch | ||
| Please make sure to build torch_tensorrt wheel file from source release/2.6 branch |
There was a problem hiding this comment.
We might need to split the toolchain updates from the docs updates since the toolchain needs to land in the release/2.6 branch.
| @@ -1,4 +1,5 @@ | |||
| setuptools==70.2.0 | |||
| numpy<2.0.0 | |||
| --index-url https://pypi.jetson-ai-lab.dev/jp6/cu126 | |||
There was a problem hiding this comment.
Id like to add this index to the pyproject.toml to support uv as a build tool which I think has the best UX but there's likely some important details we need to think about. cc: @lanluo-nvidia for later
|
how to build this? |
|
https://pypi.jetson-ai-lab.dev |
|
You can use https://pypi.jetson-ai-lab.io/ |
Add support for the jetpack6.2 build.
Currently jetpack 6.2 has:
cuda: 12.6
python: 3.10
tensorrt: 10.3
DLFW: 24.08 (pytorch: 2.6.0)
Jetson now distribute wheels on https://pypi.jetson-ai-lab.dev/
jetpack6.2 wheels: https://pypi.jetson-ai-lab.dev/jp6/cu126