Skip to content

lucas/onnx_gpu_backend#24

Merged
Edwardius merged 12 commits intomainfrom
lucas/onnx_gpu_backend
Dec 5, 2025
Merged

lucas/onnx_gpu_backend#24
Edwardius merged 12 commits intomainfrom
lucas/onnx_gpu_backend

Conversation

@lucasreljic
Copy link
Contributor

Adding deep_ort_gpu_backend_plugin to allow for inference with onnxruntime with GPU-based execution providers. Uses the onnxruntime_gpu_vendor package.

  • Tested with object_detection branch on CUDA v12.2
  • Currently, only tested with CUDA as the execution provider; through further debugging, TensorRT might be possible
  • Passes in inputs under CPU memory just like deep_ort_backend_plugin, since onnxruntime converts to GPU memory anyway. There are methods for passing in GPU memory, which could be explored in the future when ROS Nitros is implemented.

Copy link
Contributor

@Edwardius Edwardius left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for working on this! Some comments.

@Edwardius Edwardius self-requested a review December 5, 2025 04:11
Copy link
Contributor

@Edwardius Edwardius left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@lucasreljic tests pass. Works with tensorrt and Cuda

@Edwardius Edwardius marked this pull request as ready for review December 5, 2025 18:49
@Edwardius Edwardius merged commit 3f95427 into main Dec 5, 2025
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants