You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Method 1 and 2 will result in the same composed container. Furthermore, `--image` flag overrides the `--container-version` flag when both are specified.
97
+
Method 1 and 2 will result in the same composed container. Furthermore,
98
+
`--image` flag overrides the `--container-version` flag when both are specified.
99
+
100
+
Note:
101
+
1. All contents in `/opt/tritonserver` repository of the `min` image will be
102
+
removed to ensure dependencies of the composed image are added properly.
103
+
2. vLLM and TensorRT-LLM backends are currently not supported backends for
104
+
`compose.py`. If you want to build additional backends on top of these backends,
105
+
it would be better to [build it yourself](#build-it-yourself) by using
106
+
`nvcr.io/nvidia/tritonserver:24.03-vllm-python-py3` or
107
+
`nvcr.io/nvidia/tritonserver:24.03-trtllm-python-py3` as a `min` container.
108
+
94
109
95
110
### CPU-only container composition
96
111
97
-
CPU-only containers are not yet available for customization. Please see [build documentation](build.md) for instructions to build a full CPU-only container. When including TensorFlow or PyTorch backends in the composed container, an additional `gpu-min` container is needed
98
-
since this container provided the CUDA stubs and runtime dependencies which are not provided in the CPU only min container.
112
+
CPU-only containers are not yet available for customization. Please see
113
+
[build documentation](build.md) for instructions to build a full CPU-only
114
+
container. When including TensorFlow or PyTorch backends in the composed
115
+
container, an additional `gpu-min` container is needed
116
+
since this container provided the CUDA stubs and runtime dependencies which are
117
+
not provided in the CPU only min container.
99
118
100
119
## Build it yourself
101
120
102
-
If you would like to do what `compose.py` is doing under the hood yourself, you can run `compose.py` with the `--dry-run` option and then modify the `Dockerfile.compose` file to satisfy your needs.
121
+
If you would like to do what `compose.py` is doing under the hood yourself, you
122
+
can run `compose.py` with the `--dry-run` option and then modify the
123
+
`Dockerfile.compose` file to satisfy your needs.
103
124
104
125
105
126
### Triton with Unsupported and Custom Backends
@@ -110,8 +131,8 @@ result of that build should be a directory containing your backend
110
131
shared library and any additional files required by the
111
132
backend. Assuming your backend is called "mybackend" and that the
112
133
directory is "./mybackend", adding the following to the Dockerfile `compose.py`
113
-
created will create a Triton image that contains all the supported Triton backends plus your
114
-
custom backend.
134
+
created will create a Triton image that contains all the supported Triton
0 commit comments