Skip to content

Commit e42113f

Browse files
committed
Move pkg/manifests -> config/manifests
1 parent cab2472 commit e42113f

11 files changed

+14
-14
lines changed
File renamed without changes.
File renamed without changes.
File renamed without changes.
File renamed without changes.

hack/release-quickstart.sh

+4-4
Original file line numberDiff line numberDiff line change
@@ -36,9 +36,9 @@ sed -i.bak -E "s|(releases/download/)v[0-9]+\.[0-9]+\.0-rc\.?[0-9]+|\1${RELEASE_
3636
sed -i.bak "s|kubectl apply -k https://github.com/kubernetes-sigs/gateway-api-inference-extension/config/crd|kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/releases/download/${RELEASE_TAG}/manifests.yaml|g" "$README"
3737

3838
# -----------------------------------------------------------------------------
39-
# Update pkg/manifests/ext_proc.yaml
39+
# Update config/manifests/ext_proc.yaml
4040
# -----------------------------------------------------------------------------
41-
EXT_PROC="pkg/manifests/ext_proc.yaml"
41+
EXT_PROC="config/manifests/ext_proc.yaml"
4242
echo "Updating ${EXT_PROC} ..."
4343

4444
# Update the EPP container tag.
@@ -51,9 +51,9 @@ sed -i.bak '/us-central1-docker.pkg.dev\/k8s-staging-images\/gateway-api-inferen
5151
sed -i.bak -E "s|us-central1-docker\.pkg\.dev/k8s-staging-images|registry.k8s.io|g" "$EXT_PROC"
5252

5353
# -----------------------------------------------------------------------------
54-
# Update pkg/manifests/vllm/deployment.yaml
54+
# Update config/manifests/vllm/deployment.yaml
5555
# -----------------------------------------------------------------------------
56-
VLLM_DEPLOY="pkg/manifests/vllm/deployment.yaml"
56+
VLLM_DEPLOY="config/manifests/vllm/deployment.yaml"
5757
echo "Updating ${VLLM_DEPLOY} ..."
5858

5959
# Update the vLLM image version

site-src/guides/index.md

+8-8
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
1717
Deploy a sample vLLM deployment with the proper protocol to work with the LLM Instance Gateway.
1818
```bash
1919
kubectl create secret generic hf-token --from-literal=token=$HF_TOKEN # Your Hugging Face Token with access to Llama2
20-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/vllm/deployment.yaml
20+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/vllm/deployment.yaml
2121
```
2222

2323
### Install the Inference Extension CRDs
@@ -31,22 +31,22 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
3131
Deploy the sample InferenceModel which is configured to load balance traffic between the `tweet-summary-0` and `tweet-summary-1`
3232
[LoRA adapters](https://docs.vllm.ai/en/latest/features/lora.html) of the sample model server.
3333
```bash
34-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/inferencemodel.yaml
34+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/inferencemodel.yaml
3535
```
3636

3737
### Update Envoy Gateway Config to enable Patch Policy**
3838

3939
Our custom LLM Gateway ext-proc is patched into the existing envoy gateway via `EnvoyPatchPolicy`. To enable this feature, we must extend the Envoy Gateway config map. To do this, simply run:
4040
```bash
41-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/gateway/enable_patch_policy.yaml
41+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/gateway/enable_patch_policy.yaml
4242
kubectl rollout restart deployment envoy-gateway -n envoy-gateway-system
4343
```
4444
Additionally, if you would like to enable the admin interface, you can uncomment the admin lines and run this again.
4545

4646
### Deploy Gateway
4747

4848
```bash
49-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/gateway/gateway.yaml
49+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/gateway/gateway.yaml
5050
```
5151
> **_NOTE:_** This file couples together the gateway infra and the HTTPRoute infra for a convenient, quick startup. Creating additional/different InferencePools on the same gateway will require an additional set of: `Backend`, `HTTPRoute`, the resources included in the `./manifests/gateway/ext-proc.yaml` file, and an additional `./manifests/gateway/patch_policy.yaml` file. ***Should you choose to experiment, familiarity with xDS and Envoy are very useful.***
5252
@@ -59,13 +59,13 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
5959
### Deploy the Inference Extension and InferencePool
6060

6161
```bash
62-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/ext_proc.yaml
62+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/ext_proc.yaml
6363
```
6464
### Deploy Envoy Gateway Custom Policies
6565

6666
```bash
67-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/gateway/extension_policy.yaml
68-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/gateway/patch_policy.yaml
67+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/gateway/extension_policy.yaml
68+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/gateway/patch_policy.yaml
6969
```
7070
> **_NOTE:_** This is also per InferencePool, and will need to be configured to support the new pool should you wish to experiment further.
7171
@@ -74,7 +74,7 @@ This quickstart guide is intended for engineers familiar with k8s and model serv
7474
For high-traffic benchmarking you can apply this manifest to avoid any defaults that can cause timeouts/errors.
7575

7676
```bash
77-
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/pkg/manifests/gateway/traffic_policy.yaml
77+
kubectl apply -f https://github.com/kubernetes-sigs/gateway-api-inference-extension/raw/main/config/manifests/gateway/traffic_policy.yaml
7878
```
7979

8080
### Try it out

test/e2e/e2e_suite_test.go

+2-2
Original file line numberDiff line numberDiff line change
@@ -69,15 +69,15 @@ const (
6969
// clientManifest is the manifest for the client test resources.
7070
clientManifest = "../testdata/client.yaml"
7171
// modelServerManifest is the manifest for the model server test resources.
72-
modelServerManifest = "../../pkg/manifests/vllm/deployment.yaml"
72+
modelServerManifest = "../../config/manifests/vllm/deployment.yaml"
7373
// modelServerSecretManifest is the manifest for the model server secret resource.
7474
modelServerSecretManifest = "../testdata/model-secret.yaml"
7575
// inferPoolManifest is the manifest for the inference pool CRD.
7676
inferPoolManifest = "../../config/crd/bases/inference.networking.x-k8s.io_inferencepools.yaml"
7777
// inferModelManifest is the manifest for the inference model CRD.
7878
inferModelManifest = "../../config/crd/bases/inference.networking.x-k8s.io_inferencemodels.yaml"
7979
// inferExtManifest is the manifest for the inference extension test resources.
80-
inferExtManifest = "../../pkg/manifests/ext_proc.yaml"
80+
inferExtManifest = "../../config/manifests/ext_proc.yaml"
8181
// envoyManifest is the manifest for the envoy proxy test resources.
8282
envoyManifest = "../testdata/envoy.yaml"
8383
)

0 commit comments

Comments
 (0)