Skip to content

Commit 0b6b6eb

Browse files
committed
Updates artifacts for v0.1.0 release
Signed-off-by: Kellen Swain <[email protected]>
1 parent 056adfd commit 0b6b6eb

File tree

3 files changed

+3
-3
lines changed

3 files changed

+3
-3
lines changed

pkg/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
2323
1. **Install the Inference Extension CRDs:**
2424

2525
```sh
26-
kubectl apply -k https://github.com/kubernetes-sigs/gateway-api-inference-extension/config/crd
26+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/releases/download/v0.1.0/manifests.yaml
2727
```
2828

2929
1. **Deploy InferenceModel**

pkg/manifests/ext_proc.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ spec:
7171
spec:
7272
containers:
7373
- name: inference-gateway-ext-proc
74-
image: us-central1-docker.pkg.dev/k8s-staging-images/gateway-api-inference-extension/epp:main
74+
image: us-central1-docker.pkg.dev/k8s-staging-images/gateway-api-inference-extension/epp:v0.1.0
7575
args:
7676
- -poolName
7777
- "vllm-llama2-7b-pool"

pkg/manifests/vllm/deployment.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ spec:
1414
spec:
1515
containers:
1616
- name: lora
17-
image: "vllm/vllm-openai:latest"
17+
image: "vllm/vllm-openai:0.7.1"
1818
imagePullPolicy: Always
1919
command: ["python3", "-m", "vllm.entrypoints.openai.api_server"]
2020
args:

0 commit comments

Comments
 (0)