File tree 2 files changed +7
-5
lines changed
2 files changed +7
-5
lines changed Original file line number Diff line number Diff line change @@ -28,7 +28,7 @@ Getting Started
28
28
* :ref: `use_from_pytorch `
29
29
* :ref: `runtime `
30
30
* :ref: `using_dla `
31
- * :ref: `deploy_torch_tensorrt_to_triton `
31
+ * :ref: `serving_torch_tensorrt_with_triton `
32
32
33
33
.. toctree ::
34
34
:caption: Getting Started
@@ -44,7 +44,7 @@ Getting Started
44
44
tutorials/use_from_pytorch
45
45
tutorials/runtime
46
46
tutorials/using_dla
47
- tutorials/deploy_torch_tensorrt_to_triton
47
+ tutorials/serving_torch_tensorrt_with_triton
48
48
49
49
.. toctree ::
50
50
:caption: Notebooks
Original file line number Diff line number Diff line change 1
- Deploying a Torch-TensorRT model (to Triton)
2
- ============================================
1
+ .. _serving_torch_tensorrt_with_triton :
2
+
3
+ Serving a Torch-TensorRT model with Triton
4
+ ==========================================
3
5
4
6
Optimization and deployment go hand in hand in a discussion about Machine
5
7
Learning infrastructure. Once network level optimzation are done
@@ -210,5 +212,5 @@ The output of the same should look like below:
210
212
b'8.234375:11']
211
213
212
214
The output format here is ``<confidence_score>:<classification_index> ``.
213
- To learn how to map these to the label names and more, refer to our
215
+ To learn how to map these to the label names and more, refer to Triton Inference Server's
214
216
`documentation <https://github.com/triton-inference-server/server/blob/main/docs/protocol/extension_classification.md >`__.
You can’t perform that action at this time.
0 commit comments