Skip to content

Conversation

alexnorell
Copy link
Contributor

@alexnorell alexnorell commented Oct 18, 2025

Description

Fixes critical issues with the JetPack 6.2 (CUDA 12.6) Dockerfile by compiling ONNX Runtime v1.23.1 from source instead of using pre-built packages that cause memory allocation errors.

Type of change

  • Bug fix (non-breaking change which fixes an issue)

How has this change been tested, please provide a testcase or example of how you tested the change?

Built and tested on Jetson AGX Orin (JetPack 6.2, CUDA 12.6):

  • Container builds successfully
  • ONNX Runtime v1.23.1 with TensorRT + CUDA providers working
  • No memory allocation errors
  • Inference server starts and runs

Any specific deployment considerations

  • Build time: ~45-60 minutes (compiles ONNX Runtime from source)
  • Requires CMake 4.1.2+ for build
  • Uses Jetson's native CUDA 12.6 headers (ARM64-compatible)

Docs

  • Docs updated

@alexnorell alexnorell force-pushed the jetpack-6.2-onnx-runtime-fix branch from ee41f13 to b9b3957 Compare October 18, 2025 01:10
- Compile ONNX Runtime v1.23.1 from source for ARM64 compatibility
- Update CMake to v4.1.2
- Remove x86-specific CCCL headers (use Jetson's native CUDA headers)
- Fix PyPI URL to https://pypi.jetson-ai-lab.io/jp6/cu126
- Add TensorRT and CUDA execution provider support
- Use uv for faster package installation
- Multi-stage build for optimized image size
@alexnorell alexnorell force-pushed the jetpack-6.2-onnx-runtime-fix branch from b9b3957 to 9acad27 Compare October 18, 2025 01:19
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant