Describe the issue
There must be a way to build onnxruntime with tensorRt without the cuda execution provider and its cuda unused dependencies.
libonnxruntime_providers_cuda.so is big (220MB) and is dragging other big dependencies like libcufft or libcublas that we don't use in inference (another 400MB).
Urgency
non blocking
Target platform
linux
Build script
build.py
Error / output
N/A
Visual Studio Version
N/A
GCC / Compiler Version
gcc11