diff --git a/build.py b/build.py index 915bb38985..1b7d49425b 100755 --- a/build.py +++ b/build.py @@ -675,15 +675,15 @@ def onnxruntime_cmake_args(images, library_paths): ] else: if target_platform() == 'windows': - if 'ort-base' in images: + if 'base' in images: cargs.append( cmake_backend_arg('onnxruntime', 'TRITON_BUILD_CONTAINER', - None, images['ort-base'])) + None, images['base'])) else: - if 'ort-base' in images: + if 'base' in images: cargs.append( cmake_backend_arg('onnxruntime', 'TRITON_BUILD_CONTAINER', - None, images['ort-base'])) + None, images['base'])) else: cargs.append( cmake_backend_arg('onnxruntime', @@ -1085,7 +1085,6 @@ def dockerfile_prepare_container_linux(argmap, backends, enable_gpu, if 'onnxruntime' in backends: df += ''' ENV LD_LIBRARY_PATH /opt/tritonserver/backends/onnxruntime:${LD_LIBRARY_PATH} -ENV LD_LIBRARY_PATH /usr/local/cuda-11.8/lib64:${LD_LIBRARY_PATH} ''' backend_dependencies = "" @@ -1157,10 +1156,6 @@ def dockerfile_prepare_container_linux(argmap, backends, enable_gpu, if enable_gpu: df += install_dcgm_libraries(argmap['DCGM_VERSION'], target_machine) - # This is temporary solution to support 23.01 - df += ''' -RUN apt-get update && apt-get install -y libcufft-11-8 -''' df += ''' # Extra defensive wiring for CUDA Compat lib RUN ln -sf ${_CUDA_COMPAT_PATH}/lib.real ${_CUDA_COMPAT_PATH}/lib \ @@ -2243,12 +2238,9 @@ def enable_all(): fail_if( len(parts) != 2, '--image must specify ,') - # REMOVEME: ONNXRUNTIME 1.13.1 build is failing with cuda 12. Hence, using - # cuda 11.8 container as build image for ORT. fail_if( parts[0] not in [ - 'base', 'gpu-base', 'pytorch', 'tensorflow1', 'tensorflow2', - 'ort-base' + 'base', 'gpu-base', 'pytorch', 'tensorflow1', 'tensorflow2' ], 'unsupported value for --image') log('image "{}": "{}"'.format(parts[0], parts[1])) images[parts[0]] = parts[1]