Skip to content

Commit a2feb2e

Browse files
authored
typo: add vLLM Prefix Cache & LoRA Adapters links (#1280)
1 parent a634467 commit a2feb2e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

site-src/index.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,7 @@ The following specific terms to this project:
2929
from [Model Serving](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/docs/proposals/003-model-server-protocol/README.md).
3030
- **Metrics and Capabilities**: Data provided by model serving platforms about
3131
performance, availability and capabilities to optimize routing. Includes
32-
things like [Prefix Cache] status or [LoRA Adapters] availability.
32+
things like [Prefix Cache](https://docs.vllm.ai/en/stable/design/v1/prefix_caching.html) status or [LoRA Adapters](https://docs.vllm.ai/en/stable/features/lora.html) availability.
3333
- **Endpoint Picker(EPP)**: An implementation of an `Inference Scheduler` with additional Routing, Flow, and Request Control layers to allow for sophisticated routing strategies. Additional info on the architecture of the EPP [here](https://github.com/kubernetes-sigs/gateway-api-inference-extension/tree/main/docs/proposals/0683-epp-architecture-proposal).
3434

3535
[Inference Gateway]:#concepts-and-definitions

0 commit comments

Comments
 (0)