CUDNN_STATUS_EXECUTION_FAILED issue on Jetson AGX Orin with TensorFlow 2.15 GPU (original) (raw)

Hi,

I’m encountering a runtime GPU inference error when running a frozen TensorFlow model on my Jetson AGX Orin Developer Kit. The error occurs immediately upon the first inference run (sess.run ), with repeated cuDNN registration errors beforehand:

E external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc:9373 Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
E external/local_xla/xla/stream_executor/cuda/cuda_fft.cc:607 Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
E external/local_xla/xla/stream_executor/cuda/cuda_blas.cc:1534 Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered

tensorflow.python.framework.errors_impl.UnknownError: 2 root error(s) found.
  (0) UNKNOWN: CUDNN_STATUS_EXECUTION_FAILED
in external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc(6917): 'status'
 [[{{node graph/fe_shared/conv_down_1_0/Conv2D}}]]
 [[graph/reference_output_disparity/_209]]
  (1) UNKNOWN: CUDNN_STATUS_EXECUTION_FAILED
in external/local_xla/xla/stream_executor/cuda/cuda_dnn.cc(6917): 'status'
 [[{{node graph/fe_shared/conv_down_1_0/Conv2D}}]]
0 successful operations.
0 derived errors ignored.
Original stack trace for 'graph/fe_shared/conv_down_1_0/Conv2D':

The error happens in the Conv2D layer within the imported TensorFlow graph.

Environment Details:
Jetson Platform - NVIDIA Jetson AGX Orin Developer Kit with JetPack 6.1
Python - 3.10.12 (inside a virtual environment)|
TensorFlow Version - 2.15.0+nv24.5 (GPU-enabled build)
TensorFlow Build Info - CUDA 12.2, cuDNN 8
CUDA - 12.2 (nvcc V12.2.140)
cuDNN - 8.9.7.29-1+cuda12.2 (confirmed via dpkg and header file)
GPU Driver - 540.4.0 (checked via nvidia-smi)
OpenCV 4.5.5 (Python pip package),
Other Packages - absl-py 2.3.1, pyyaml installed

No other GPU processes are running during tests (nvidia-smi shows zero GPU utilization

TensorFlow CUDA & GPU Check

python3 -c "import tensorflow as tf; print('CUDA:', tf.test.is_built_with_cuda()); print('GPU Available:', tf.config.list_physical_devices('GPU'))"

Output:

E ... Unable to register cuDNN factory: Attempting to register factory for plugin cuDNN when one has already been registered
E ... Unable to register cuFFT factory: Attempting to register factory for plugin cuFFT when one has already been registered
E ... Unable to register cuBLAS factory: Attempting to register factory for plugin cuBLAS when one has already been registered
CUDA: True
GPU Available: [PhysicalDevice(name='/physical_device:GPU:0', device_type='GPU')]

I’ve also tried the below debugging steps:

Request for Assistance

If any forum members have hit similar errors or can share experience running custom TF frozen graphs with cuDNN on Jetson AGX Orin, your insights will be highly appreciated! Thanks in advance for your help!

Hi @naveenkumar.kamalakannan ,
I believe Jetson AGX Orin team may assist better here.

Thanks