Skip to content

Commit 85487a1

Browse files
authored
Fixes in request cancellation doc (#6409)
1 parent 9c707e3 commit 85487a1

File tree

1 file changed

+4
-4
lines changed

1 file changed

+4
-4
lines changed

docs/user_guide/request_cancellation.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828

2929
# Request Cancellation
3030

31-
Starting from 23.10, Triton supports handling request cancellation received
31+
Starting from r23.10, Triton supports handling request cancellation received
3232
from the gRPC client or a C API user. Long running inference requests such
3333
as for auto generative large language models may run for an indeterminate
3434
amount of time or indeterminate number of steps. Additionally clients may
@@ -39,7 +39,7 @@ resources.
3939

4040
## Issuing Request Cancellation
4141

42-
### Triton C API
42+
### In-Process C API
4343

4444
[In-Process Triton Server C API](../customization_guide/inference_protocols.md#in-process-triton-server-api) has been enhanced with `TRITONSERVER_InferenceRequestCancel`
4545
and `TRITONSERVER_InferenceRequestIsCancelled` to issue cancellation and query
@@ -77,9 +77,9 @@ detection and handling within Triton core is work in progress.
7777

7878
## Handling in Backend
7979

80-
Upon receiving request cancellation, triton does its best to terminate request
80+
Upon receiving request cancellation, Triton does its best to terminate request
8181
at various points. However, once a request has been given to the backend
82-
for execution, it is upto the individual backends to detect and handle
82+
for execution, it is up to the individual backends to detect and handle
8383
request termination.
8484
Currently, the following backends support early termination:
8585
- [vLLM backend](https://github.com/triton-inference-server/vllm_backend)

0 commit comments

Comments
 (0)