-
Notifications
You must be signed in to change notification settings - Fork 4.7k
/
Copy pathhelpers.go
333 lines (293 loc) · 15.5 KB
/
helpers.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
package machine_config
import (
"context"
"encoding/json"
"fmt"
"math/rand"
"time"
osconfigv1 "github.com/openshift/api/config/v1"
machinev1beta1 "github.com/openshift/api/machine/v1beta1"
mcfgv1 "github.com/openshift/api/machineconfiguration/v1"
opv1 "github.com/openshift/api/operator/v1"
machineclient "github.com/openshift/client-go/machine/clientset/versioned"
machineconfigclient "github.com/openshift/client-go/machineconfiguration/clientset/versioned"
mcopclient "github.com/openshift/client-go/operator/clientset/versioned"
o "github.com/onsi/gomega"
exutil "github.com/openshift/origin/test/extended/util"
corev1 "k8s.io/api/core/v1"
apierrors "k8s.io/apimachinery/pkg/api/errors"
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
"k8s.io/client-go/kubernetes"
"k8s.io/kubernetes/test/e2e/framework"
e2eskipper "k8s.io/kubernetes/test/e2e/framework/skipper"
"k8s.io/utils/ptr"
"sigs.k8s.io/yaml"
)
const (
mcoNamespace = "openshift-machine-config-operator"
mapiNamespace = "openshift-machine-api"
mapiMachinesetQualifiedName = "machinesets.machine.openshift.io"
cmName = "coreos-bootimages"
mapiMasterMachineLabelSelector = "machine.openshift.io/cluster-api-machine-role=master"
mapiMachineSetArchAnnotationKey = "capacity.cluster-autoscaler.kubernetes.io/labels"
)
// skipUnlessTargetPlatform skips the test if it is running on the target platform
func skipUnlessTargetPlatform(oc *exutil.CLI, platformType osconfigv1.PlatformType) {
infra, err := oc.AdminConfigClient().ConfigV1().Infrastructures().Get(context.Background(), "cluster", metav1.GetOptions{})
o.Expect(err).NotTo(o.HaveOccurred())
if infra.Status.PlatformStatus.Type != platformType {
e2eskipper.Skipf("This test only applies to %s platform", platformType)
}
}
// skipUnlessFunctionalMachineAPI skips the test if the cluster is using Machine API
func skipUnlessFunctionalMachineAPI(oc *exutil.CLI) {
machineClient, err := machineclient.NewForConfig(oc.KubeFramework().ClientConfig())
o.Expect(err).ToNot(o.HaveOccurred())
machines, err := machineClient.MachineV1beta1().Machines(mapiNamespace).List(context.Background(), metav1.ListOptions{LabelSelector: mapiMasterMachineLabelSelector})
// the machine API can be unavailable resulting in a 404 or an empty list
if err != nil {
if !apierrors.IsNotFound(err) {
o.Expect(err).ToNot(o.HaveOccurred())
}
e2eskipper.Skipf("haven't found machines resources on the cluster, this test can be run on a platform that supports functional MachineAPI")
return
}
if len(machines.Items) == 0 {
e2eskipper.Skipf("got an empty list of machines resources from the cluster, this test can be run on a platform that supports functional MachineAPI")
return
}
// we expect just a single machine to be in the Running state
for _, machine := range machines.Items {
phase := ptr.Deref(machine.Status.Phase, "")
if phase == "Running" {
return
}
}
e2eskipper.Skipf("haven't found a machine in running state, this test can be run on a platform that supports functional MachineAPI")
}
// skipOnSingleNodeTopology skips the test if the cluster is using single-node topology
func skipOnSingleNodeTopology(oc *exutil.CLI) {
infra, err := oc.AdminConfigClient().ConfigV1().Infrastructures().Get(context.Background(), "cluster", metav1.GetOptions{})
o.Expect(err).NotTo(o.HaveOccurred())
if infra.Status.ControlPlaneTopology == osconfigv1.SingleReplicaTopologyMode {
e2eskipper.Skipf("This test does not apply to single-node topologies")
}
}
// getRandomMachineSet picks a random machineset present on the cluster
func getRandomMachineSet(machineClient *machineclient.Clientset) machinev1beta1.MachineSet {
machineSets, err := machineClient.MachineV1beta1().MachineSets("openshift-machine-api").List(context.TODO(), metav1.ListOptions{})
o.Expect(err).NotTo(o.HaveOccurred())
rnd := rand.New(rand.NewSource(time.Now().UnixNano()))
machineSetUnderTest := machineSets.Items[rnd.Intn(len(machineSets.Items))]
return machineSetUnderTest
}
// verifyMachineSetUpdate verifies that the the boot image values of a MachineSet are reconciled correctly
func verifyMachineSetUpdate(oc *exutil.CLI, machineSet machinev1beta1.MachineSet, updateExpected bool) {
newProviderSpecPatch, originalProviderSpecPatch, newBootImage, originalBootImage := createFakeUpdatePatch(oc, machineSet)
err := oc.Run("patch").Args(mapiMachinesetQualifiedName, machineSet.Name, "-p", fmt.Sprintf(`{"spec":{"template":{"spec":{"providerSpec":{"value":%s}}}}}`, newProviderSpecPatch), "-n", mapiNamespace, "--type=merge").Execute()
o.Expect(err).NotTo(o.HaveOccurred())
defer func() {
// Restore machineSet to original boot image as the machineset may be used by other test variants, regardless of success/fail
err = oc.Run("patch").Args(mapiMachinesetQualifiedName, machineSet.Name, "-p", fmt.Sprintf(`{"spec":{"template":{"spec":{"providerSpec":{"value":%s}}}}}`, originalProviderSpecPatch), "-n", mapiNamespace, "--type=merge").Execute()
o.Expect(err).NotTo(o.HaveOccurred())
framework.Logf("Restored build name in the machineset %s from \"%s\" to \"%s\"", machineSet.Name, newBootImage, originalBootImage)
}()
// Ensure boot image controller is not progressing
framework.Logf("Waiting until the boot image controller is not progressing...")
WaitForBootImageControllerToComplete(oc)
// Fetch the providerSpec of the machineset under test again
providerSpecDisks, err := oc.Run("get").Args(mapiMachinesetQualifiedName, machineSet.Name, "-o", "template", "--template=`{{.spec.template.spec.providerSpec.value}}`", "-n", mapiNamespace).Output()
o.Expect(err).NotTo(o.HaveOccurred())
// Verify that the machineset has the expected boot image values
if updateExpected {
o.Expect(providerSpecDisks).ShouldNot(o.ContainSubstring(newBootImage))
} else {
o.Expect(providerSpecDisks).Should(o.ContainSubstring(newBootImage))
}
}
// unmarshalProviderSpec unmarshals the machineset's provider spec into
// a ProviderSpec object. Returns an error if providerSpec field is nil,
// or the unmarshal fails
func unmarshalProviderSpec(ms *machinev1beta1.MachineSet, providerSpec interface{}) error {
if ms.Spec.Template.Spec.ProviderSpec.Value == nil {
return fmt.Errorf("providerSpec field was empty")
}
if err := yaml.Unmarshal(ms.Spec.Template.Spec.ProviderSpec.Value.Raw, &providerSpec); err != nil {
return fmt.Errorf("unmarshal into providerSpec failed %w", err)
}
return nil
}
// marshalProviderSpec marshals the ProviderSpec object into a MachineSet object.
// Returns an error if ProviderSpec or MachineSet is nil, or if the marshal fails
func marshalProviderSpec(providerSpec interface{}) (string, error) {
if providerSpec == nil {
return "", fmt.Errorf("ProviderSpec object was nil")
}
rawBytes, err := json.Marshal(providerSpec)
if err != nil {
return "", fmt.Errorf("marshalling providerSpec failed: %w", err)
}
return string(rawBytes), nil
}
// createFakeUpdatePatch creates an update patch for the MachineSet object based on the platform
func createFakeUpdatePatch(oc *exutil.CLI, machineSet machinev1beta1.MachineSet) (string, string, string, string) {
infra, err := oc.AdminConfigClient().ConfigV1().Infrastructures().Get(context.Background(), "cluster", metav1.GetOptions{})
o.Expect(err).NotTo(o.HaveOccurred())
switch infra.Status.PlatformStatus.Type {
case osconfigv1.AWSPlatformType:
return generateAWSProviderSpecPatch(machineSet)
case osconfigv1.GCPPlatformType:
return generateGCPProviderSpecPatch(machineSet)
default:
exutil.FatalErr(fmt.Errorf("unexpected platform type; should not be here"))
return "", "", "", ""
}
}
// generateAWSProviderSpecPatch generates a fake update patch for the AWS MachineSet
func generateAWSProviderSpecPatch(machineSet machinev1beta1.MachineSet) (string, string, string, string) {
providerSpec := new(machinev1beta1.AWSMachineProviderConfig)
err := unmarshalProviderSpec(&machineSet, providerSpec)
o.Expect(err).NotTo(o.HaveOccurred())
// Modify the boot image to a "fake" value
originalBootImage := *providerSpec.AMI.ID
newBootImage := originalBootImage + "-fake-update"
newProviderSpec := providerSpec.DeepCopy()
newProviderSpec.AMI.ID = &newBootImage
newProviderSpecPatch, err := marshalProviderSpec(newProviderSpec)
o.Expect(err).NotTo(o.HaveOccurred())
originalProviderSpecPatch, err := marshalProviderSpec(providerSpec)
o.Expect(err).NotTo(o.HaveOccurred())
return newProviderSpecPatch, originalProviderSpecPatch, newBootImage, originalBootImage
}
// generateGCPProviderSpecPatch generates a fake update patch for the GCP MachineSet
func generateGCPProviderSpecPatch(machineSet machinev1beta1.MachineSet) (string, string, string, string) {
providerSpec := new(machinev1beta1.GCPMachineProviderSpec)
err := unmarshalProviderSpec(&machineSet, providerSpec)
o.Expect(err).NotTo(o.HaveOccurred())
// Modify the boot image to a "fake" value
originalBootImage := providerSpec.Disks[0].Image
newBootImage := "projects/centos-cloud/global/images/family/centos-stream-9"
newProviderSpec := providerSpec.DeepCopy()
for idx := range newProviderSpec.Disks {
newProviderSpec.Disks[idx].Image = newBootImage
}
newProviderSpecPatch, err := marshalProviderSpec(newProviderSpec)
o.Expect(err).NotTo(o.HaveOccurred())
originalProviderSpecPatch, err := marshalProviderSpec(providerSpec)
o.Expect(err).NotTo(o.HaveOccurred())
return newProviderSpecPatch, originalProviderSpecPatch, newBootImage, originalBootImage
}
// WaitForBootImageControllerToComplete waits until the boot image controller is no longer progressing
func WaitForBootImageControllerToComplete(oc *exutil.CLI) {
machineConfigurationClient, err := mcopclient.NewForConfig(oc.KubeFramework().ClientConfig())
o.Expect(err).NotTo(o.HaveOccurred())
// This has a MustPassRepeatedly(3) to ensure there isn't a false positive by checking the
// controller status too quickly after applying the fixture/labels.
o.Eventually(func() bool {
mcop, err := machineConfigurationClient.OperatorV1().MachineConfigurations().Get(context.TODO(), "cluster", metav1.GetOptions{})
if err != nil {
framework.Logf("Failed to grab machineconfiguration object, error :%v", err)
return false
}
return IsMachineConfigurationConditionFalse(mcop.Status.Conditions, opv1.MachineConfigurationBootImageUpdateProgressing)
}, 3*time.Minute, 5*time.Second).MustPassRepeatedly(3).Should(o.BeTrue())
}
// WaitForMachineConfigurationStatus waits until the MCO syncs the operator status to the latest spec
func WaitForMachineConfigurationStatusUpdate(oc *exutil.CLI) {
machineConfigurationClient, err := mcopclient.NewForConfig(oc.KubeFramework().ClientConfig())
o.Expect(err).NotTo(o.HaveOccurred())
// This has a MustPassRepeatedly(3) to ensure there isn't a false positive by checking the
// status too quickly after applying the fixture.
o.Eventually(func() bool {
mcop, err := machineConfigurationClient.OperatorV1().MachineConfigurations().Get(context.TODO(), "cluster", metav1.GetOptions{})
if err != nil {
framework.Logf("Failed to grab machineconfiguration object, error :%v", err)
return false
}
return mcop.Generation == mcop.Status.ObservedGeneration
}, 3*time.Minute, 1*time.Second).MustPassRepeatedly(3).Should(o.BeTrue())
}
// IsMachineConfigPoolConditionTrue returns true when the conditionType is present and set to `ConditionTrue`
func IsMachineConfigPoolConditionTrue(conditions []mcfgv1.MachineConfigPoolCondition, conditionType mcfgv1.MachineConfigPoolConditionType) bool {
for _, condition := range conditions {
if condition.Type == conditionType {
return condition.Status == corev1.ConditionTrue
}
}
return false
}
// IsMachineConfigurationConditionFalse returns false when the conditionType is present and set to `ConditionFalse`
func IsMachineConfigurationConditionFalse(conditions []metav1.Condition, conditionType string) bool {
for _, condition := range conditions {
if condition.Type == conditionType {
return condition.Status == metav1.ConditionFalse
}
}
return false
}
// IsClusterOperatorConditionTrue returns true when the conditionType is present and set to `configv1.ConditionTrue`
func IsClusterOperatorConditionTrue(conditions []osconfigv1.ClusterOperatorStatusCondition, conditionType osconfigv1.ClusterStatusConditionType) bool {
return IsClusterOperatorConditionPresentAndEqual(conditions, conditionType, osconfigv1.ConditionTrue)
}
// IsClusterOperatorConditionFalse returns true when the conditionType is present and set to `configv1.ConditionFalse`
func IsClusterOperatorConditionFalse(conditions []osconfigv1.ClusterOperatorStatusCondition, conditionType osconfigv1.ClusterStatusConditionType) bool {
return IsClusterOperatorConditionPresentAndEqual(conditions, conditionType, osconfigv1.ConditionFalse)
}
// IsClusterOperatorConditionPresentAndEqual returns true when conditionType is present and equal to status.
func IsClusterOperatorConditionPresentAndEqual(conditions []osconfigv1.ClusterOperatorStatusCondition, conditionType osconfigv1.ClusterStatusConditionType, status osconfigv1.ConditionStatus) bool {
for _, condition := range conditions {
if condition.Type == conditionType {
return condition.Status == status
}
}
return false
}
// FindClusterOperatorStatusCondition finds the conditionType in conditions.
func FindClusterOperatorStatusCondition(conditions []osconfigv1.ClusterOperatorStatusCondition, conditionType osconfigv1.ClusterStatusConditionType) *osconfigv1.ClusterOperatorStatusCondition {
for i := range conditions {
if conditions[i].Type == conditionType {
return &conditions[i]
}
}
return nil
}
// WaitForOneMasterNodeToBeReady waits until atleast one master node has completed an update
func WaitForOneMasterNodeToBeReady(oc *exutil.CLI) error {
machineConfigClient, err := machineconfigclient.NewForConfig(oc.KubeFramework().ClientConfig())
o.Expect(err).NotTo(o.HaveOccurred())
o.Eventually(func() bool {
mcp, err := machineConfigClient.MachineconfigurationV1().MachineConfigPools().Get(context.TODO(), "master", metav1.GetOptions{})
if err != nil {
framework.Logf("Failed to grab machineconfigpools, error :%v", err)
return false
}
// Check if the pool has atleast one updated node(mid-upgrade), or if the pool has completed the upgrade to the new config(the additional check for spec==status here is
// to ensure we are not checking an older "Updated" condition and the MCP fields haven't caught up yet
if (IsMachineConfigPoolConditionTrue(mcp.Status.Conditions, mcfgv1.MachineConfigPoolUpdating) && mcp.Status.UpdatedMachineCount > 0) ||
(IsMachineConfigPoolConditionTrue(mcp.Status.Conditions, mcfgv1.MachineConfigPoolUpdated) && (mcp.Spec.Configuration.Name == mcp.Status.Configuration.Name)) {
return true
}
framework.Logf("Waiting for atleast one ready control-plane node")
return false
}, 5*time.Minute, 10*time.Second).Should(o.BeTrue())
return nil
}
// Applies a boot image fixture and waits for the MCO to reconcile the status
func ApplyBootImageFixture(oc *exutil.CLI, fixture string) {
err := oc.Run("apply").Args("-f", fixture).Execute()
o.Expect(err).NotTo(o.HaveOccurred())
// Ensure status accounts for the fixture that was applied
WaitForMachineConfigurationStatusUpdate(oc)
}
// Get nodes from a Pool
func getNodesForPool(ctx context.Context, oc *exutil.CLI, kubeClient *kubernetes.Clientset, pool *mcfgv1.MachineConfigPool) (*corev1.NodeList, error) {
selector, err := metav1.LabelSelectorAsSelector(pool.Spec.NodeSelector)
if err != nil {
return nil, fmt.Errorf("invalid label selector: %w", err)
}
nodes, err := kubeClient.CoreV1().Nodes().List(ctx, metav1.ListOptions{LabelSelector: selector.String()})
if err != nil {
return nil, fmt.Errorf("couldnt get nodes for mcp: %w", err)
}
return nodes, nil
}