Starting from r23.10, Triton supports handling request cancellation received from the gRPC client or a C API user. Long running inference requests such as for auto generative large language models may run for an indeterminate amount of time or indeterminate number of steps. Additionally clients may enqueue a large number of requests as part of a sequence or request stream and later determine the results are no longer needed. Continuing to process requests whose results are no longer required can significantly impact server resources.
In-Process Triton Server C API has been enhanced with TRITONSERVER_InferenceRequestCancel
and TRITONSERVER_InferenceRequestIsCancelled
to issue cancellation and query
whether cancellation has been issued on an inflight request respectively. Read more
about the APIs in tritonserver.h.
In addition, gRPC endpoint can now detect cancellation from the client and attempt to terminate request. At present, only gRPC python client supports issuing request cancellation to the server endpoint. See request-cancellation for more details on how to issue requests from the client-side. See gRPC guide on RPC cancellation for finer details.
Triton core checks for requests that have been cancelled at some critical points when using dynamic or sequence batching. The checking is also performed between each ensemble steps and terminates further processing if the request is cancelled.
On detecting a cancelled request, Triton core responds with CANCELLED status. If a request is cancelled when using sequence_batching, then all the pending requests in the same sequence will also be cancelled. The sequence is represented by the requests that has identical sequence id.
Note: Currently, Triton core does not detect cancellation status of a request once it is forwarded to rate limiter. Improving the request cancellation detection and handling within Triton core is work in progress.
Upon receiving request cancellation, Triton does its best to terminate request at various points. However, once a request has been given to the backend for execution, it is up to the individual backends to detect and handle request termination. Currently, the following backends support early termination:
Python backend is a special case where we expose the APIs to detect cancellation
status of the request but it is up to the model.py
developer to detect whether
the request is cancelled and terminate further execution.
For the backend developer: The backend APIs have also been enhanced to let the
backend detect whether the request received from Triton core has been cancelled.
See TRITONBACKEND_RequestIsCancelled
and TRITONBACKEND_ResponseFactoryIsCancelled
in tritonbackend.h
for more details. The backend upon detecting request cancellation can stop processing
it any further.
The Python models running behind Python backend can also query the cancellation status
of request and response_sender. See this
section in python backend documentation for more details.