Skip to content

Commit 4bc0be6

Browse files
author
OpenShift Bot
authored
Merge pull request #12650 from danwinship/multicast-fixes-and-annotation
Merged by openshift-bot
2 parents 6ceb412 + 8ba9e66 commit 4bc0be6

12 files changed

+221
-339
lines changed

pkg/sdn/api/plugin.go

+3
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,9 @@ const (
1717
// HostSubnet annotations. (Note: should be "hostsubnet.network.openshift.io/", but the incorrect name is now part of the API.)
1818
AssignHostSubnetAnnotation = "pod.network.openshift.io/assign-subnet"
1919
FixedVNIDHostAnnotation = "pod.network.openshift.io/fixed-vnid-host"
20+
21+
// NetNamespace annotations
22+
MulticastEnabledAnnotation = "netnamespace.network.openshift.io/multicast-enabled"
2023
)
2124

2225
func IsOpenShiftNetworkPlugin(pluginName string) bool {

pkg/sdn/plugin/controller.go

+8-2
Original file line numberDiff line numberDiff line change
@@ -313,10 +313,16 @@ func (plugin *OsdnNode) SetupSDN() (bool, error) {
313313
// eg, "table=100, reg0=${tenant_id}, priority=2, ip, nw_dst=${external_cidr}, actions=drop
314314
otx.AddFlow("table=100, priority=0, actions=output:2")
315315

316-
// Table 110: multicast delivery from local pods to the VXLAN
316+
// Table 110: outbound multicast filtering, updated by updateLocalMulticastFlows() in pod.go
317+
// eg, "table=110, priority=100, reg0=${tenant_id}, actions=goto_table:111
317318
otx.AddFlow("table=110, priority=0, actions=drop")
318319

319-
// Table 120: multicast delivery to local pods (either from VXLAN or local pods)
320+
// Table 111: multicast delivery from local pods to the VXLAN; only one rule, updated by updateVXLANMulticastRules() in subnets.go
321+
// eg, "table=111, priority=100, actions=move:NXM_NX_REG0[]->NXM_NX_TUN_ID[0..31],set_field:${remote_node_ip_1}->tun_dst,output:1,set_field:${remote_node_ip_2}->tun_dst,output:1,goto_table:120"
322+
otx.AddFlow("table=111, priority=0, actions=drop")
323+
324+
// Table 120: multicast delivery to local pods (either from VXLAN or local pods); updated by updateLocalMulticastFlows() in pod.go
325+
// eg, "table=120, priority=100, reg0=${tenant_id}, actions=output:${ovs_port_1},output:${ovs_port_2}"
320326
otx.AddFlow("table=120, priority=0, actions=drop")
321327

322328
err = otx.EndTransaction()

pkg/sdn/plugin/multitenant.go

+34-21
Original file line numberDiff line numberDiff line change
@@ -64,35 +64,41 @@ func (mp *multiTenantPlugin) updatePodNetwork(namespace string, oldNetID, netID
6464
services = &kapi.ServiceList{}
6565
}
6666

67-
movedVNIDRefs := 0
67+
if oldNetID != netID {
68+
movedVNIDRefs := 0
69+
70+
// Update OF rules for the existing/old pods in the namespace
71+
for _, pod := range pods {
72+
err = mp.node.UpdatePod(pod)
73+
if err == nil {
74+
movedVNIDRefs++
75+
} else {
76+
glog.Errorf("Could not update pod %q in namespace %q: %v", pod.Name, namespace, err)
77+
}
78+
}
79+
80+
// Update OF rules for the old services in the namespace
81+
for _, svc := range services.Items {
82+
if !kapi.IsServiceIPSet(&svc) {
83+
continue
84+
}
6885

69-
// Update OF rules for the existing/old pods in the namespace
70-
for _, pod := range pods {
71-
err = mp.node.UpdatePod(pod)
72-
if err == nil {
86+
mp.node.DeleteServiceRules(&svc)
87+
mp.node.AddServiceRules(&svc, netID)
7388
movedVNIDRefs++
74-
} else {
75-
glog.Errorf("Could not update pod %q in namespace %q: %v", pod.Name, namespace, err)
7689
}
77-
}
7890

79-
// Update OF rules for the old services in the namespace
80-
for _, svc := range services.Items {
81-
if !kapi.IsServiceIPSet(&svc) {
82-
continue
91+
if movedVNIDRefs > 0 {
92+
mp.moveVNIDRefs(movedVNIDRefs, oldNetID, netID)
8393
}
8494

85-
mp.node.DeleteServiceRules(&svc)
86-
mp.node.AddServiceRules(&svc, netID)
87-
movedVNIDRefs++
88-
}
89-
90-
if movedVNIDRefs > 0 {
91-
mp.moveVNIDRefs(movedVNIDRefs, oldNetID, netID)
95+
// Update namespace references in egress firewall rules
96+
mp.node.UpdateEgressNetworkPolicyVNID(namespace, oldNetID, netID)
9297
}
9398

94-
// Update namespace references in egress firewall rules
95-
mp.node.UpdateEgressNetworkPolicyVNID(namespace, oldNetID, netID)
99+
// Update local multicast rules
100+
mp.node.podManager.UpdateLocalMulticastRules(oldNetID)
101+
mp.node.podManager.UpdateLocalMulticastRules(netID)
96102
}
97103

98104
func (mp *multiTenantPlugin) AddNetNamespace(netns *osapi.NetNamespace) {
@@ -115,6 +121,13 @@ func (mp *multiTenantPlugin) GetNamespaces(vnid uint32) []string {
115121
return mp.vnids.GetNamespaces(vnid)
116122
}
117123

124+
func (mp *multiTenantPlugin) GetMulticastEnabled(vnid uint32) bool {
125+
if vnid == osapi.GlobalVNID {
126+
return false
127+
}
128+
return mp.vnids.GetMulticastEnabled(vnid)
129+
}
130+
118131
func (mp *multiTenantPlugin) RefVNID(vnid uint32) {
119132
if vnid == 0 {
120133
return

pkg/sdn/plugin/networkpolicy.go

+9-1
Original file line numberDiff line numberDiff line change
@@ -160,7 +160,11 @@ func (np *networkPolicyPlugin) AddNetNamespace(netns *osapi.NetNamespace) {
160160
}
161161

162162
func (np *networkPolicyPlugin) UpdateNetNamespace(netns *osapi.NetNamespace, oldNetID uint32) {
163-
glog.Warning("Got UpdateNetNamespace for namespace %s (%d) while using %s plugin", netns.NetName, netns.NetID, osapi.NetworkPolicyPluginName)
163+
if netns.NetID != oldNetID {
164+
glog.Warning("Got VNID change for namespace %s while using %s plugin", netns.NetName, osapi.NetworkPolicyPluginName)
165+
}
166+
167+
np.node.podManager.UpdateLocalMulticastRules(netns.NetID)
164168
}
165169

166170
func (np *networkPolicyPlugin) DeleteNetNamespace(netns *osapi.NetNamespace) {
@@ -178,6 +182,10 @@ func (np *networkPolicyPlugin) GetNamespaces(vnid uint32) []string {
178182
return np.vnids.GetNamespaces(vnid)
179183
}
180184

185+
func (np *networkPolicyPlugin) GetMulticastEnabled(vnid uint32) bool {
186+
return np.vnids.GetMulticastEnabled(vnid)
187+
}
188+
181189
func (np *networkPolicyPlugin) syncNamespace(npns *npNamespace) {
182190
inUse := npns.refs > 0
183191
if !inUse && !npns.inUse {

pkg/sdn/plugin/node.go

+1
Original file line numberDiff line numberDiff line change
@@ -40,6 +40,7 @@ type osdnPolicy interface {
4040

4141
GetVNID(namespace string) (uint32, error)
4242
GetNamespaces(vnid uint32) []string
43+
GetMulticastEnabled(vnid uint32) bool
4344

4445
RefVNID(vnid uint32)
4546
UnrefVNID(vnid uint32)

pkg/sdn/plugin/pod.go

+91-94
Original file line numberDiff line numberDiff line change
@@ -5,6 +5,7 @@ import (
55
"fmt"
66
"net"
77
"sort"
8+
"sync"
89

910
"github.com/openshift/origin/pkg/sdn/plugin/cniserver"
1011
"github.com/openshift/origin/pkg/util/netutils"
@@ -21,7 +22,7 @@ import (
2122

2223
type podHandler interface {
2324
setup(req *cniserver.PodRequest) (*cnitypes.Result, *runningPod, error)
24-
update(req *cniserver.PodRequest) (*runningPod, error)
25+
update(req *cniserver.PodRequest) (uint32, error)
2526
teardown(req *cniserver.PodRequest) error
2627
}
2728

@@ -38,7 +39,8 @@ type podManager struct {
3839
// Request queue for pod operations incoming from the CNIServer
3940
requests chan (*cniserver.PodRequest)
4041
// Tracks pod :: IP address for hostport handling
41-
runningPods map[string]*runningPod
42+
runningPods map[string]*runningPod
43+
runningPodsLock sync.Mutex
4244

4345
// Live pod setup/teardown stuff not used in testing code
4446
kClient *kclientset.Clientset
@@ -99,7 +101,7 @@ func getIPAMConfig(clusterNetwork *net.IPNet, localSubnet string) ([]byte, error
99101
IPAM *hostLocalIPAM `json:"ipam"`
100102
}
101103

102-
mcaddr := net.ParseIP("224.0.0.0")
104+
_, mcnet, _ := net.ParseCIDR("224.0.0.0/3")
103105
return json.Marshal(&cniNetworkConfig{
104106
Name: "openshift-sdn",
105107
Type: "openshift-sdn",
@@ -111,19 +113,20 @@ func getIPAMConfig(clusterNetwork *net.IPNet, localSubnet string) ([]byte, error
111113
},
112114
Routes: []cnitypes.Route{
113115
{
116+
// Default route
114117
Dst: net.IPNet{
115118
IP: net.IPv4zero,
116119
Mask: net.IPMask(net.IPv4zero),
117120
},
118121
GW: netutils.GenerateDefaultGateway(nodeNet),
119122
},
120-
{Dst: *clusterNetwork},
123+
{
124+
// Cluster network
125+
Dst: *clusterNetwork,
126+
},
121127
{
122128
// Multicast
123-
Dst: net.IPNet{
124-
IP: mcaddr,
125-
Mask: net.IPMask(mcaddr),
126-
},
129+
Dst: *mcnet,
127130
},
128131
},
129132
},
@@ -179,112 +182,106 @@ func (m *podManager) handleCNIRequest(request *cniserver.PodRequest) ([]byte, er
179182
return result.Response, result.Err
180183
}
181184

182-
type runningPodsSlice []*runningPod
183-
184-
func (l runningPodsSlice) Len() int { return len(l) }
185-
func (l runningPodsSlice) Less(i, j int) bool { return l[i].ofport < l[j].ofport }
186-
func (l runningPodsSlice) Swap(i, j int) { l[i], l[j] = l[j], l[i] }
187-
188-
// FIXME: instead of calculating all this ourselves, figure out a way to pass
189-
// the old VNID through the Update() call (or get it from somewhere else).
190-
func updateMulticastFlows(runningPods map[string]*runningPod, ovs *ovs.Interface, podKey string, changedPod *runningPod) error {
191-
// FIXME: prevents TestPodUpdate() from crashing. (We separately test this function anyway.)
192-
if ovs == nil {
193-
return nil
194-
}
195-
196-
// Build map of pods by their VNID, excluding the changed pod
197-
podsByVNID := make(map[uint32]runningPodsSlice)
198-
for key, runningPod := range runningPods {
199-
if key != podKey {
200-
podsByVNID[runningPod.vnid] = append(podsByVNID[runningPod.vnid], runningPod)
185+
func localMulticastOutputs(runningPods map[string]*runningPod, vnid uint32) string {
186+
var ofports []int
187+
for _, pod := range runningPods {
188+
if pod.vnid == vnid {
189+
ofports = append(ofports, pod.ofport)
201190
}
202191
}
192+
if len(ofports) == 0 {
193+
return ""
194+
}
203195

204-
// Figure out what two VNIDs changed so we can update only those two flows
205-
changedVNIDs := make([]uint32, 0)
206-
oldPod, exists := runningPods[podKey]
207-
if changedPod != nil {
208-
podsByVNID[changedPod.vnid] = append(podsByVNID[changedPod.vnid], changedPod)
209-
changedVNIDs = append(changedVNIDs, changedPod.vnid)
210-
if exists {
211-
// VNID changed
212-
changedVNIDs = append(changedVNIDs, oldPod.vnid)
196+
sort.Ints(ofports)
197+
outputs := ""
198+
for _, ofport := range ofports {
199+
if len(outputs) > 0 {
200+
outputs += ","
213201
}
214-
} else if exists {
215-
// Pod deleted
216-
changedVNIDs = append(changedVNIDs, oldPod.vnid)
202+
outputs += fmt.Sprintf("output:%d", ofport)
217203
}
204+
return outputs
205+
}
218206

219-
if len(changedVNIDs) == 0 {
220-
// Shouldn't happen, but whatever
221-
return fmt.Errorf("Multicast update requested but not required!")
207+
func (m *podManager) updateLocalMulticastRulesWithLock(vnid uint32) {
208+
var outputs string
209+
otx := m.ovs.NewTransaction()
210+
if m.policy.GetMulticastEnabled(vnid) {
211+
outputs = localMulticastOutputs(m.runningPods, vnid)
212+
otx.AddFlow("table=110, reg0=%d, actions=goto_table:111", vnid)
213+
} else {
214+
otx.DeleteFlows("table=110, reg0=%d", vnid)
222215
}
223-
224-
otx := ovs.NewTransaction()
225-
for _, vnid := range changedVNIDs {
226-
// Sort pod array to ensure consistent ordering for testcases and readability
227-
pods := podsByVNID[vnid]
228-
sort.Sort(pods)
229-
230-
// build up list of ports on this VNID
231-
outputs := ""
232-
for _, pod := range pods {
233-
if len(outputs) > 0 {
234-
outputs += ","
235-
}
236-
outputs += fmt.Sprintf("output:%d", pod.ofport)
237-
}
238-
239-
// Update or delete the flows for the vnid
240-
if len(outputs) > 0 {
241-
otx.AddFlow("table=120, priority=100, reg0=%d, actions=%s", vnid, outputs)
242-
} else {
243-
otx.DeleteFlows("table=120, reg0=%d", vnid)
244-
}
216+
if len(outputs) > 0 {
217+
otx.AddFlow("table=120, priority=100, reg0=%d, actions=%s", vnid, outputs)
218+
} else {
219+
otx.DeleteFlows("table=120, reg0=%d", vnid)
220+
}
221+
if err := otx.EndTransaction(); err != nil {
222+
glog.Errorf("Error updating OVS multicast flows for VNID %d: %v", vnid, err)
245223
}
246-
return otx.EndTransaction()
224+
}
225+
226+
// Update multicast OVS rules for the given vnid
227+
func (m *podManager) UpdateLocalMulticastRules(vnid uint32) {
228+
m.runningPodsLock.Lock()
229+
defer m.runningPodsLock.Unlock()
230+
m.updateLocalMulticastRulesWithLock(vnid)
247231
}
248232

249233
// Process all CNI requests from the request queue serially. Our OVS interaction
250234
// and scripts currently cannot run in parallel, and doing so greatly complicates
251235
// setup/teardown logic
252236
func (m *podManager) processCNIRequests() {
253237
for request := range m.requests {
254-
pk := getPodKey(request)
255-
256-
var pod *runningPod
257-
var ipamResult *cnitypes.Result
258-
259238
glog.V(5).Infof("Processing pod network request %v", request)
260-
result := &cniserver.PodResult{}
261-
switch request.Command {
262-
case cniserver.CNI_ADD:
263-
ipamResult, pod, result.Err = m.podHandler.setup(request)
264-
if ipamResult != nil {
265-
result.Response, result.Err = json.Marshal(ipamResult)
239+
result := m.processRequest(request)
240+
glog.V(5).Infof("Processed pod network request %v, result %s err %v", request, string(result.Response), result.Err)
241+
request.Result <- result
242+
}
243+
panic("stopped processing CNI pod requests!")
244+
}
245+
246+
func (m *podManager) processRequest(request *cniserver.PodRequest) *cniserver.PodResult {
247+
m.runningPodsLock.Lock()
248+
defer m.runningPodsLock.Unlock()
249+
250+
pk := getPodKey(request)
251+
result := &cniserver.PodResult{}
252+
switch request.Command {
253+
case cniserver.CNI_ADD:
254+
ipamResult, runningPod, err := m.podHandler.setup(request)
255+
if ipamResult != nil {
256+
result.Response, err = json.Marshal(ipamResult)
257+
if result.Err == nil {
258+
m.runningPods[pk] = runningPod
259+
if m.ovs != nil {
260+
m.updateLocalMulticastRulesWithLock(runningPod.vnid)
261+
}
266262
}
267-
case cniserver.CNI_UPDATE:
268-
pod, result.Err = m.podHandler.update(request)
269-
case cniserver.CNI_DEL:
270-
result.Err = m.podHandler.teardown(request)
271-
default:
272-
result.Err = fmt.Errorf("unhandled CNI request %v", request.Command)
273263
}
274-
275-
if result.Err == nil {
276-
if err := updateMulticastFlows(m.runningPods, m.ovs, pk, pod); err != nil {
277-
glog.Warningf("Failed to update multicast flows: %v", err)
264+
if err != nil {
265+
result.Err = err
266+
}
267+
case cniserver.CNI_UPDATE:
268+
vnid, err := m.podHandler.update(request)
269+
if err == nil {
270+
if runningPod, exists := m.runningPods[pk]; exists {
271+
runningPod.vnid = vnid
278272
}
279-
if pod != nil {
280-
m.runningPods[pk] = pod
281-
} else {
282-
delete(m.runningPods, pk)
273+
}
274+
result.Err = err
275+
case cniserver.CNI_DEL:
276+
if runningPod, exists := m.runningPods[pk]; exists {
277+
delete(m.runningPods, pk)
278+
if m.ovs != nil {
279+
m.updateLocalMulticastRulesWithLock(runningPod.vnid)
283280
}
284281
}
285-
286-
glog.V(5).Infof("Processed pod network request %v, result %s err %v", request, string(result.Response), result.Err)
287-
request.Result <- result
282+
result.Err = m.podHandler.teardown(request)
283+
default:
284+
result.Err = fmt.Errorf("unhandled CNI request %v", request.Command)
288285
}
289-
panic("stopped processing CNI pod requests!")
286+
return result
290287
}

0 commit comments

Comments
 (0)