Skip to content

Commit b9f69b5

Browse files
mc-nvkrishung5
andauthored
Update post-23.11 release (#6653)
* Update README and versions for 2.40.0 / 23.11 (#6544) * Removing path construction to use SymLink alternatives * Update version for PyTorch * Update windows Dockerfile configuration * Update triton version to 23.11 * Update README and versions for 2.40.0 / 23.11 * Fix typo * Ading 'ldconfig' to configure dynamic linking in container (#6602) * Point to tekit_backend (#6616) * Point to tekit_backend * Update version * Revert tekit changes (#6640) --------- Co-authored-by: Kris Hung <[email protected]>
1 parent b96ae5f commit b9f69b5

File tree

26 files changed

+74
-54
lines changed

26 files changed

+74
-54
lines changed

Dockerfile.sdk

+1-1
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@
2929
#
3030

3131
# Base image on the minimum Triton container
32-
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:23.10-py3-min
32+
ARG BASE_IMAGE=nvcr.io/nvidia/tritonserver:23.11-py3-min
3333

3434
ARG TRITON_CLIENT_REPO_SUBDIR=clientrepo
3535
ARG TRITON_COMMON_REPO_TAG=main

Dockerfile.win10.min

+21-5
Original file line numberDiff line numberDiff line change
@@ -83,7 +83,13 @@ ARG BUILD_TOOLS_SOURCE=https://aka.ms/vs/17/release/vs_buildtools.exe
8383
ADD ${BUILD_TOOLS_SOURCE} vs_buildtools.exe
8484
# Install Build Tools with the Microsoft.VisualStudio.Workload.VCTools workload, including recommended.
8585
ARG VS_INSTALL_PATH_WP="C:\BuildTools"
86-
RUN vs_buildtools.exe --quiet --wait --norestart --nocache install --installPath %VS_INSTALL_PATH_WP% --channelUri "C:\tmp\VisualStudio.chman" --installChannelUri "C:\tmp\VisualStudio.chman" --add Microsoft.VisualStudio.Workload.VCTools --includeRecommended --locale "En-us"
86+
RUN vs_buildtools.exe --quiet --wait --norestart --nocache install \
87+
--installPath %VS_INSTALL_PATH_WP% \
88+
--channelUri "C:\tmp\VisualStudio.chman" \
89+
--installChannelUri "C:\tmp\VisualStudio.chman" \
90+
--add Microsoft.VisualStudio.Workload.VCTools \
91+
--includeRecommended \
92+
--locale "En-us"
8793

8894
LABEL BUILDTOOLS_VERSION=${BUILDTOOLS_VERSION}
8995

@@ -97,7 +103,17 @@ RUN git clone --single-branch --depth=1 -b %VCPGK_VERSION% https://github.com/mi
97103
WORKDIR /vcpkg
98104
RUN bootstrap-vcpkg.bat
99105
RUN vcpkg.exe update
100-
RUN vcpkg.exe install openssl:x64-windows openssl-windows:x64-windows rapidjson:x64-windows re2:x64-windows boost-filesystem:x64-windows boost-interprocess:x64-windows boost-stacktrace:x64-windows zlib:x64-windows pthread:x64-windows b64:x64-windows
106+
RUN vcpkg.exe install \
107+
b64:x64-windows \
108+
boost-filesystem:x64-windows \
109+
boost-interprocess:x64-windows \
110+
boost-stacktrace:x64-windows \
111+
openssl-windows:x64-windows \
112+
openssl:x64-windows \
113+
pthread:x64-windows \
114+
rapidjson:x64-windows \
115+
re2:x64-windows \
116+
zlib:x64-windows
101117
RUN vcpkg.exe integrate install
102118

103119
LABEL VCPGK_VERSION=${VCPGK_VERSION}
@@ -108,8 +124,8 @@ WORKDIR /
108124
# Installing CUDA
109125
#
110126
ARG CUDA_MAJOR=12
111-
ARG CUDA_MINOR=2
112-
ARG CUDA_PATCH=1
127+
ARG CUDA_MINOR=3
128+
ARG CUDA_PATCH=0
113129
ARG CUDA_VERSION=${CUDA_MAJOR}.${CUDA_MINOR}.${CUDA_PATCH}
114130
ARG CUDA_PACKAGES="nvcc_${CUDA_MAJOR}.${CUDA_MINOR} \
115131
cudart_${CUDA_MAJOR}.${CUDA_MINOR} \
@@ -157,7 +173,7 @@ LABEL TENSORRT_VERSION="${TENSORRT_VERSION}"
157173
#
158174
# Installing cuDNN
159175
#
160-
ARG CUDNN_VERSION=8.9.5.27
176+
ARG CUDNN_VERSION=8.9.6.50
161177
ARG CUDNN_ZIP=cudnn-windows-x86_64-${CUDNN_VERSION}_cuda12-archive.zip
162178
ARG CUDNN_SOURCE=${CUDNN_ZIP}
163179

README.md

+5-5
Original file line numberDiff line numberDiff line change
@@ -32,8 +32,8 @@
3232

3333
**LATEST RELEASE: You are currently on the main branch which tracks
3434
under-development progress towards the next release. The current release is
35-
version [2.38.0](https://github.com/triton-inference-server/server/tree/r23.09)
36-
and corresponds to the 23.09 container release on
35+
version [2.40.0](https://github.com/triton-inference-server/server/tree/r23.11)
36+
and corresponds to the 23.11 container release on
3737
[NVIDIA GPU Cloud (NGC)](https://catalog.ngc.nvidia.com/orgs/nvidia/containers/tritonserver).**
3838

3939
----
@@ -93,16 +93,16 @@ Inference Server with the
9393

9494
```bash
9595
# Step 1: Create the example model repository
96-
git clone -b r23.10 https://github.com/triton-inference-server/server.git
96+
git clone -b r23.11 https://github.com/triton-inference-server/server.git
9797
cd server/docs/examples
9898
./fetch_models.sh
9999

100100
# Step 2: Launch triton from the NGC Triton container
101-
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:23.10-py3 tritonserver --model-repository=/models
101+
docker run --gpus=1 --rm --net=host -v ${PWD}/model_repository:/models nvcr.io/nvidia/tritonserver:23.11-py3 tritonserver --model-repository=/models
102102

103103
# Step 3: Sending an Inference Request
104104
# In a separate console, launch the image_client example from the NGC Triton SDK container
105-
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:23.10-py3-sdk
105+
docker run -it --rm --net=host nvcr.io/nvidia/tritonserver:23.11-py3-sdk
106106
/workspace/install/bin/image_client -m densenet_onnx -c 3 -s INCEPTION /workspace/images/mug.jpg
107107

108108
# Inference should return the following

build.py

+5-5
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,7 @@
7272
TRITON_VERSION_MAP = {
7373
"2.41.0dev": (
7474
"23.12dev", # triton container
75-
"23.10", # upstream container
75+
"23.11", # upstream container
7676
"1.16.3", # ORT
7777
"2023.0.0", # ORT OpenVINO
7878
"2023.0.0", # Standalone OpenVINO
@@ -1389,10 +1389,10 @@ def add_cpu_libs_to_linux_dockerfile(backends, target_machine):
13891389
COPY --from=min_container /usr/local/cuda/lib64/stubs/libcublasLt.so /usr/local/cuda/lib64/stubs/libcublasLt.so.11
13901390
13911391
RUN mkdir -p /usr/local/cuda/targets/{cuda_arch}-linux/lib
1392-
COPY --from=min_container /usr/local/cuda-12.2/targets/{cuda_arch}-linux/lib/libcudart.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1393-
COPY --from=min_container /usr/local/cuda-12.2/targets/{cuda_arch}-linux/lib/libcupti.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1394-
COPY --from=min_container /usr/local/cuda-12.2/targets/{cuda_arch}-linux/lib/libnvToolsExt.so.1 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1395-
COPY --from=min_container /usr/local/cuda-12.2/targets/{cuda_arch}-linux/lib/libnvJitLink.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1392+
COPY --from=min_container /usr/local/cuda/lib64/libcudart.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1393+
COPY --from=min_container /usr/local/cuda/lib64/libcupti.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1394+
COPY --from=min_container /usr/local/cuda/lib64/libnvToolsExt.so.1 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
1395+
COPY --from=min_container /usr/local/cuda/lib64/libnvJitLink.so.12 /usr/local/cuda/targets/{cuda_arch}-linux/lib/.
13961396
13971397
RUN mkdir -p /opt/hpcx/ucc/lib/ /opt/hpcx/ucx/lib/
13981398
COPY --from=min_container /opt/hpcx/ucc/lib/libucc.so.1 /opt/hpcx/ucc/lib/libucc.so.1

deploy/aws/values.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:23.10-py3
30+
imageName: nvcr.io/nvidia/tritonserver:23.11-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: s3://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/fleetcommand/Chart.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -26,7 +26,7 @@
2626

2727
apiVersion: v1
2828
# appVersion is the Triton version; update when changing release
29-
appVersion: "2.39.0"
29+
appVersion: "2.40.0"
3030
description: Triton Inference Server (Fleet Command)
3131
name: triton-inference-server
3232
# version is the Chart version; update when changing anything in the chart

deploy/fleetcommand/values.yaml

+3-3
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:23.10-py3
30+
imageName: nvcr.io/nvidia/tritonserver:23.11-py3
3131
pullPolicy: IfNotPresent
3232
numGpus: 1
3333
serverCommand: tritonserver
@@ -46,13 +46,13 @@ image:
4646
# Model Control Mode (Optional, default: none)
4747
#
4848
# To set model control mode, uncomment and configure below
49-
# See https://github.com/triton-inference-server/server/blob/r23.10/docs/model_management.md
49+
# See https://github.com/triton-inference-server/server/blob/r23.11/docs/model_management.md
5050
# for more details
5151
#- --model-control-mode=explicit|poll|none
5252
#
5353
# Additional server args
5454
#
55-
# see https://github.com/triton-inference-server/server/blob/r23.10/README.md
55+
# see https://github.com/triton-inference-server/server/blob/r23.11/README.md
5656
# for more details
5757

5858
service:

deploy/gcp/values.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
replicaCount: 1
2828

2929
image:
30-
imageName: nvcr.io/nvidia/tritonserver:23.10-py3
30+
imageName: nvcr.io/nvidia/tritonserver:23.11-py3
3131
pullPolicy: IfNotPresent
3232
modelRepositoryPath: gs://triton-inference-server-repository/model_repository
3333
numGpus: 1

deploy/gke-marketplace-app/benchmark/perf-analyzer-script/triton_client.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ metadata:
3333
namespace: default
3434
spec:
3535
containers:
36-
- image: nvcr.io/nvidia/tritonserver:23.10-py3-sdk
36+
- image: nvcr.io/nvidia/tritonserver:23.11-py3-sdk
3737
imagePullPolicy: Always
3838
name: nv-triton-client
3939
securityContext:

deploy/gke-marketplace-app/server-deployer/build_and_push.sh

+3-3
Original file line numberDiff line numberDiff line change
@@ -27,9 +27,9 @@
2727

2828
export REGISTRY=gcr.io/$(gcloud config get-value project | tr ':' '/')
2929
export APP_NAME=tritonserver
30-
export MAJOR_VERSION=2.39
31-
export MINOR_VERSION=2.39.0
32-
export NGC_VERSION=23.10-py3
30+
export MAJOR_VERSION=2.40
31+
export MINOR_VERSION=2.40.0
32+
export NGC_VERSION=23.11-py3
3333

3434
docker pull nvcr.io/nvidia/$APP_NAME:$NGC_VERSION
3535

deploy/gke-marketplace-app/server-deployer/chart/triton/Chart.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
2626

2727
apiVersion: v1
28-
appVersion: "2.39"
28+
appVersion: "2.40"
2929
description: Triton Inference Server
3030
name: triton-inference-server
31-
version: 2.39.0
31+
version: 2.40.0

deploy/gke-marketplace-app/server-deployer/chart/triton/values.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -32,13 +32,13 @@ tritonProtocol: HTTP
3232
# HPA GPU utilization autoscaling target
3333
HPATargetAverageValue: 85
3434
modelRepositoryPath: gs://triton_sample_models/23_09
35-
publishedVersion: '2.39.0'
35+
publishedVersion: '2.40.0'
3636
gcpMarketplace: true
3737

3838
image:
3939
registry: gcr.io
4040
repository: nvidia-ngc-public/tritonserver
41-
tag: 23.10-py3
41+
tag: 23.11-py3
4242
pullPolicy: IfNotPresent
4343
# modify the model repository here to match your GCP storage bucket
4444
numGpus: 1

deploy/gke-marketplace-app/server-deployer/data-test/schema.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
x-google-marketplace:
2828
schemaVersion: v2
2929
applicationApiVersion: v1beta1
30-
publishedVersion: '2.39.0'
30+
publishedVersion: '2.40.0'
3131
publishedVersionMetadata:
3232
releaseNote: >-
3333
Initial release.

deploy/gke-marketplace-app/server-deployer/schema.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
x-google-marketplace:
2828
schemaVersion: v2
2929
applicationApiVersion: v1beta1
30-
publishedVersion: '2.39.0'
30+
publishedVersion: '2.40.0'
3131
publishedVersionMetadata:
3232
releaseNote: >-
3333
Initial release.

deploy/gke-marketplace-app/trt-engine/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@
3333
```
3434
docker run --gpus all -it --network host \
3535
--shm-size=1g --ulimit memlock=-1 --ulimit stack=67108864 \
36-
-v ~:/scripts nvcr.io/nvidia/tensorrt:23.10-py3
36+
-v ~:/scripts nvcr.io/nvidia/tensorrt:23.11-py3
3737
3838
pip install onnx six torch tf2onnx tensorflow
3939

deploy/k8s-onprem/values.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ tags:
2929
loadBalancing: true
3030

3131
image:
32-
imageName: nvcr.io/nvidia/tritonserver:23.10-py3
32+
imageName: nvcr.io/nvidia/tritonserver:23.11-py3
3333
pullPolicy: IfNotPresent
3434
modelRepositoryServer: < Replace with the IP Address of your file server >
3535
modelRepositoryPath: /srv/models

docs/customization_guide/build.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -173,7 +173,7 @@ $ ./build.py ... --repo-tag=common:<container tag> --repo-tag=core:<container ta
173173

174174
If you are building on a release branch then `<container tag>` will
175175
default to the branch name. For example, if you are building on the
176-
r23.10 branch, `<container tag>` will default to r23.10. If you are
176+
r23.11 branch, `<container tag>` will default to r23.11. If you are
177177
building on any other branch (including the *main* branch) then
178178
`<container tag>` will default to "main". Therefore, you typically do
179179
not need to provide `<container tag>` at all (nor the preceding
@@ -334,8 +334,8 @@ python build.py --cmake-dir=<path/to/repo>/build --build-dir=/tmp/citritonbuild
334334
If you are building on *main* branch then '<container tag>' will
335335
default to "main". If you are building on a release branch then
336336
'<container tag>' will default to the branch name. For example, if you
337-
are building on the r23.10 branch, '<container tag>' will default to
338-
r23.10. Therefore, you typically do not need to provide '<container
337+
are building on the r23.11 branch, '<container tag>' will default to
338+
r23.11. Therefore, you typically do not need to provide '<container
339339
tag>' at all (nor the preceding colon). You can use a different
340340
'<container tag>' for a component to instead use the corresponding
341341
branch/tag in the build. For example, if you have a branch called

docs/customization_guide/compose.md

+7-7
Original file line numberDiff line numberDiff line change
@@ -44,8 +44,8 @@ from source to get more exact customization.
4444
The `compose.py` script can be found in the [server repository](https://github.com/triton-inference-server/server).
4545
Simply clone the repository and run `compose.py` to create a custom container.
4646
Note: Created container version will depend on the branch that was cloned.
47-
For example branch [r23.10](https://github.com/triton-inference-server/server/tree/r23.10)
48-
should be used to create a image based on the NGC 23.10 Triton release.
47+
For example branch [r23.11](https://github.com/triton-inference-server/server/tree/r23.11)
48+
should be used to create a image based on the NGC 23.11 Triton release.
4949

5050
`compose.py` provides `--backend`, `--repoagent` options that allow you to
5151
specify which backends and repository agents to include in the custom image.
@@ -76,19 +76,19 @@ For example, running
7676
```
7777
python3 compose.py --backend tensorflow1 --repoagent checksum
7878
```
79-
on branch [r23.10](https://github.com/triton-inference-server/server/tree/r23.10) pulls:
80-
- `min` container `nvcr.io/nvidia/tritonserver:23.10-py3-min`
81-
- `full` container `nvcr.io/nvidia/tritonserver:23.10-py3`
79+
on branch [r23.11](https://github.com/triton-inference-server/server/tree/r23.11) pulls:
80+
- `min` container `nvcr.io/nvidia/tritonserver:23.11-py3-min`
81+
- `full` container `nvcr.io/nvidia/tritonserver:23.11-py3`
8282

8383
Alternatively, users can specify the version of Triton container to pull from any branch by either:
8484
1. Adding flag `--container-version <container version>` to branch
8585
```
86-
python3 compose.py --backend tensorflow1 --repoagent checksum --container-version 23.10
86+
python3 compose.py --backend tensorflow1 --repoagent checksum --container-version 23.11
8787
```
8888
2. Specifying `--image min,<min container image name> --image full,<full container image name>`.
8989
The user is responsible for specifying compatible `min` and `full` containers.
9090
```
91-
python3 compose.py --backend tensorflow1 --repoagent checksum --image min,nvcr.io/nvidia/tritonserver:23.10-py3-min --image full,nvcr.io/nvidia/tritonserver:23.10-py3
91+
python3 compose.py --backend tensorflow1 --repoagent checksum --image min,nvcr.io/nvidia/tritonserver:23.11-py3-min --image full,nvcr.io/nvidia/tritonserver:23.11-py3
9292
```
9393
Method 1 and 2 will result in the same composed container. Furthermore, `--image` flag overrides the `--container-version` flag when both are specified.
9494

docs/customization_guide/test.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -49,7 +49,7 @@ $ ./gen_qa_custom_ops
4949
```
5050

5151
This will create multiple model repositories in /tmp/<version>/qa_*
52-
(for example /tmp/23.10/qa_model_repository). The TensorRT models
52+
(for example /tmp/23.11/qa_model_repository). The TensorRT models
5353
will be created for the GPU on the system that CUDA considers device 0
5454
(zero). If you have multiple GPUs on your system see the documentation
5555
in the scripts for how to target a specific GPU.

docs/user_guide/custom_operations.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ simple way to ensure you are using the correct version of TensorRT is
6464
to use the [NGC TensorRT
6565
container](https://ngc.nvidia.com/catalog/containers/nvidia:tensorrt)
6666
corresponding to the Triton container. For example, if you are using
67-
the 23.10 version of Triton, use the 23.10 version of the TensorRT
67+
the 23.11 version of Triton, use the 23.11 version of the TensorRT
6868
container.
6969

7070
## TensorFlow
@@ -123,7 +123,7 @@ simple way to ensure you are using the correct version of TensorFlow
123123
is to use the [NGC TensorFlow
124124
container](https://ngc.nvidia.com/catalog/containers/nvidia:tensorflow)
125125
corresponding to the Triton container. For example, if you are using
126-
the 23.10 version of Triton, use the 23.10 version of the TensorFlow
126+
the 23.11 version of Triton, use the 23.11 version of the TensorFlow
127127
container.
128128
129129
## PyTorch
@@ -167,7 +167,7 @@ simple way to ensure you are using the correct version of PyTorch is
167167
to use the [NGC PyTorch
168168
container](https://ngc.nvidia.com/catalog/containers/nvidia:pytorch)
169169
corresponding to the Triton container. For example, if you are using
170-
the 23.10 version of Triton, use the 23.10 version of the PyTorch
170+
the 23.11 version of Triton, use the 23.11 version of the PyTorch
171171
container.
172172
173173
## ONNX

docs/user_guide/performance_tuning.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -235,7 +235,7 @@ with a `tritonserver` binary.
235235

236236
```bash
237237
# Start server container
238-
docker run -ti --rm --gpus=all --network=host -v $PWD:/mnt --name triton-server nvcr.io/nvidia/tritonserver:23.10-py3
238+
docker run -ti --rm --gpus=all --network=host -v $PWD:/mnt --name triton-server nvcr.io/nvidia/tritonserver:23.11-py3
239239

240240
# Start serving your models
241241
tritonserver --model-repository=/mnt/models
@@ -284,7 +284,7 @@ by setting the `-u` flag, such as `perf_analyzer -m densenet_onnx -u
284284

285285
```bash
286286
# Start the SDK container interactively
287-
docker run -ti --rm --gpus=all --network=host -v $PWD:/mnt --name triton-client nvcr.io/nvidia/tritonserver:23.10-py3-sdk
287+
docker run -ti --rm --gpus=all --network=host -v $PWD:/mnt --name triton-client nvcr.io/nvidia/tritonserver:23.11-py3-sdk
288288

289289
# Benchmark model being served from step 3
290290
perf_analyzer -m densenet_onnx --concurrency-range 1:4

qa/L0_infer/test.sh

+2
Original file line numberDiff line numberDiff line change
@@ -38,6 +38,8 @@ if [ ! -z "$TEST_REPO_ARCH" ]; then
3838
REPO_VERSION=${REPO_VERSION}_${TEST_REPO_ARCH}
3939
fi
4040

41+
ldconfig || true
42+
4143
export CUDA_VISIBLE_DEVICES=0
4244

4345
TEST_RESULT_FILE='test_results.txt'

qa/L0_sequence_batcher/test.sh

+2
Original file line numberDiff line numberDiff line change
@@ -42,6 +42,8 @@ TEST_RESULT_FILE='test_results.txt'
4242

4343
# Must run on a single device or else the TRITONSERVER_DELAY_SCHEDULER
4444
# can fail when the requests are distributed to multiple devices.
45+
ldconfig || true
46+
4547
export CUDA_VISIBLE_DEVICES=0
4648

4749
CLIENT_LOG="./client.log"

qa/common/gen_jetson_trt_models

+1-1
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@
3434
# Make all generated files accessible outside of container
3535
umask 0000
3636
# Set the version of the models
37-
TRITON_VERSION=${TRITON_VERSION:=23.10}
37+
TRITON_VERSION=${TRITON_VERSION:=23.11}
3838
# Set the CUDA device to use
3939
CUDA_DEVICE=${RUNNER_ID:=0}
4040
# Set TensorRT image

0 commit comments

Comments
 (0)