This package provides the reference implementation for the Endpoint Picker (EPP). As demonistrated in the diagram below, it implements the extension protocol, enabling a proxy or gateway to request endpoint hints from an extension, and interacts with the model servers through the defined model server protocol.
An EPP instance handles a single InferencePool
(and so for each InferencePool
, one must create a dedicated EPP deployment), it performs the following core functions:
- Endpoint Selection
- The EPP determines the appropriate Pod endpoint for the load balancer (LB) to route requests.
- It selects from the pool of ready Pods designated by the assigned InferencePool's Selector field.
- Endpoint selection is contingent on the request's ModelName matching an
InferenceModel
that references theInferencePool
. - Requests with unmatched ModelName values trigger an error response to the proxy.
- Traffic Splitting and ModelName Rewriting
- The EPP facilitates controlled rollouts of new adapter versions by implementing traffic splitting between adapters within the same
InferencePool
, as defined by theInferenceModel
. - EPP rewrites the model name in the request to the target model name as defined on the
InferenceModel
object.
- The EPP facilitates controlled rollouts of new adapter versions by implementing traffic splitting between adapters within the same
- Observability
- The EPP generates metrics to enhance observability.
- It reports InferenceModel-level metrics, further broken down by target model.
- Detailed information regarding metrics can be found on the website.
The scheduling package implements request scheduling algorithms for load balancing requests across backend pods in an inference gateway. The scheduler ensures efficient resource utilization while maintaining low latency and prioritizing critical requests. It applies a series of filters based on metrics and heuristics to select the best pod for a given request. The following flow chart summarizes the current scheduling algorithm