6
6
"path/filepath"
7
7
"time"
8
8
9
- mcClient "github.com/openshift/client-go/machineconfiguration/clientset/versioned"
9
+ machineconfigclient "github.com/openshift/client-go/machineconfiguration/clientset/versioned"
10
10
exutil "github.com/openshift/origin/test/extended/util"
11
11
12
12
g "github.com/onsi/ginkgo/v2"
@@ -34,12 +34,6 @@ var _ = g.Describe("[sig-mco][OCPFeatureGate:MachineConfigNode][Serial]", func()
34
34
oc = exutil .NewCLIWithoutNamespace ("machine-config" )
35
35
)
36
36
37
- // TODO: Update to properly cleanup after tests
38
- // g.AfterAll(func(ctx context.Context) {
39
- // // clean up the created custom MCP
40
- // CleanupCustomMCP(oc)
41
- // })
42
-
43
37
g .It ("Should have MCN properties matching associated node properties [apigroup:machineconfiguration.openshift.io]" , func () {
44
38
ValidateMCNProperties (oc , infraMCPFixture )
45
39
})
@@ -52,7 +46,7 @@ var _ = g.Describe("[sig-mco][OCPFeatureGate:MachineConfigNode][Serial]", func()
52
46
// `ValidateMCNProperties` checks that MCN properties match the corresponding node properties
53
47
func ValidateMCNProperties (oc * exutil.CLI , fixture string ) {
54
48
// Create client set for test
55
- clientSet , clientErr := mcClient .NewForConfig (oc .KubeFramework ().ClientConfig ())
49
+ clientSet , clientErr := machineconfigclient .NewForConfig (oc .KubeFramework ().ClientConfig ())
56
50
o .Expect (clientErr ).NotTo (o .HaveOccurred ())
57
51
58
52
// Grab a random node from each default pool
@@ -99,8 +93,29 @@ func ValidateMCNProperties(oc *exutil.CLI, fixture string) {
99
93
labelErr := oc .Run ("label" ).Args (fmt .Sprintf ("node/%s" , workerNode .Name ), fmt .Sprintf ("node-role.kubernetes.io/%s=" , custom )).Execute ()
100
94
o .Expect (labelErr ).NotTo (o .HaveOccurred ())
101
95
96
+ defer func () {
97
+ // Get starting state of default worker MCP
98
+ workerMcp , err := clientSet .MachineconfigurationV1 ().MachineConfigPools ().Get (context .TODO (), worker , metav1.GetOptions {})
99
+ o .Expect (err ).NotTo (o .HaveOccurred ())
100
+ workerMcpReadyMachines := workerMcp .Status .ReadyMachineCount
101
+
102
+ // Unlabel node
103
+ framework .Logf ("Removing label node-role.kubernetes.io/%v from node %v" , custom , workerNode .Name )
104
+ unlabelErr := oc .Run ("label" ).Args (fmt .Sprintf ("node/%s" , workerNode .Name ), fmt .Sprintf ("node-role.kubernetes.io/%s-" , custom )).Execute ()
105
+ o .Expect (unlabelErr ).NotTo (o .HaveOccurred ())
106
+
107
+ // Wait for infra pool to report no nodes & for worker MCP to be ready
108
+ WaitForMCPToBeReady (oc , clientSet , custom , 0 )
109
+ WaitForMCPToBeReady (oc , clientSet , worker , workerMcpReadyMachines + 1 )
110
+
111
+ // Delete custom MCP
112
+ framework .Logf ("Deleting MCP %v" , custom )
113
+ deleteMCPErr := oc .Run ("delete" ).Args ("mcp" , custom ).Execute ()
114
+ o .Expect (deleteMCPErr ).NotTo (o .HaveOccurred ())
115
+ }()
116
+
102
117
// Wait for the custom pool to be updated with the node ready
103
- WaitForMCPToBeReady (oc , custom )
118
+ WaitForMCPToBeReady (oc , clientSet , custom , 1 )
104
119
105
120
// Get node desired and current config versions
106
121
customNodes , customNodeErr := GetNodesByRole (oc , custom )
@@ -134,7 +149,7 @@ func ValidateMCNProperties(oc *exutil.CLI, fixture string) {
134
149
// `ValidateMCNConditionTransitions` check that Conditions properly update on a node update
135
150
func ValidateMCNConditionTransitions (oc * exutil.CLI , fixture string ) {
136
151
// Create client set for test
137
- clientSet , clientErr := mcClient .NewForConfig (oc .KubeFramework ().ClientConfig ())
152
+ clientSet , clientErr := machineconfigclient .NewForConfig (oc .KubeFramework ().ClientConfig ())
138
153
o .Expect (clientErr ).NotTo (o .HaveOccurred ())
139
154
140
155
// Apply MC targeting worker pool
@@ -207,33 +222,3 @@ func ValidateMCNConditionTransitions(oc *exutil.CLI, fixture string) {
207
222
framework .Logf ("Checking all conditions other than 'Updated' are False." )
208
223
o .Expect (confirmUpdatedMCNStatus (clientSet , workerNode .Name )).Should (o .BeTrue ())
209
224
}
210
-
211
- // TODO: test this cleanup works when running full test
212
- // `CleanupCustomMCP` deletes the custom MCP for the MCN tests
213
- func CleanupCustomMCP (oc * exutil.CLI ) {
214
- // TODO: add length check to see if any nodes are labeled with custom role
215
- // TODO: add check if mcp exists before trying to delete it
216
-
217
- // Remove custom role from nodes
218
- customNodes , customNodeErr := GetNodesByRole (oc , custom )
219
- o .Expect (customNodeErr ).NotTo (o .HaveOccurred ())
220
- for _ , node := range customNodes {
221
- framework .Logf ("Unlabeling node %v" , node .Name )
222
- unlabelErr := oc .Run ("label" ).Args (fmt .Sprintf ("node/%s" , node .Name ), fmt .Sprintf ("node-role.kubernetes.io/%s-" , custom )).Execute ()
223
- o .Expect (unlabelErr ).NotTo (o .HaveOccurred ())
224
- }
225
-
226
- // Wait for worker MCP to be updated
227
- // TODO: fix this since it seemes to not wait long enough to actually catch the mcp needing an update and being updated
228
- // TODO: maybe check the node annotations instead?
229
- // TODO: Maybe update WaitForMCPToBeReady to take an int again but have it be a number representing the previous number of machines in the pool? so that the updated can also chek if ready machine count is greater than the previous count.
230
- framework .Logf ("Waiting for worker MCP to re-sync." )
231
- WaitForMCPToBeReady (oc , worker )
232
-
233
- // Delete custom MCP
234
- framework .Logf ("Deleting MCP %v" , custom )
235
- deleteMCPErr := oc .Run ("delete" ).Args ("mcp" , custom ).Execute ()
236
- o .Expect (deleteMCPErr ).NotTo (o .HaveOccurred ())
237
-
238
- framework .Logf ("Custom MCP %v has been cleaned up." , custom )
239
- }
0 commit comments