mogaika / kube-vxlan-controller

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Kube VXLAN Controller

License: Apache-2.0 GitHub Release Badge

Any pod in a Kubernetes cluster can become a VXLAN overlay network member by having a VXLAN type network interface set up and L2 peer forwarding entries specified. This can be done automatically on a pod creation by using the Kube VXLAN Controller. A pod could be configured to have any number of VXLAN interfaces, i.e. to be a member of any number of VXLAN Segments.

Deployment

The controller monitors pods using the Kubernetes API and could run as a standalone application provided with a desired cluster API access. But the most simple way is to run it as a part of the Kubernetes cluster itself, for example as a Kubernetes deployment. This repository provides a Bundle Manifest to create such a deployment and related workloads:

$ kubectl apply -f https://raw.githubusercontent.com/openvnf/kube-vxlan-controller/master/kubernetes/bundle.yaml

Usage

To make a pod VXLAN enabled it should answer the following conditions:

  • have a "vxlan.openvnf.org" label set to "true"
  • have a "vxlan.openvnf.org/networks" annotation describing VXLAN networks
  • run a Kube VXLAN Controller Agent init container with the security context "NET_ADMIN" capability
  • run a Kube VXLAN Controller Agent sidecar container with the security context "NET_ADMIN" capability.

These conditions could be described in a single manifest:

spec:
  template:
    metadata:
      labels:
        vxlan.openvnf.org: "true"
      annotations:
        vxlan.openvnf.org/networks: vxeth0, vxeth1
    spec:
      initContainers:
      - name: vxlan-controller-agent-init
        image: openvnf/kube-vxlan-controller-agent
        securityContext:
          capabilities:
            add: ["NET_ADMIN"]
      containers:
      - name: vxlan-controller-agent
        image: openvnf/kube-vxlan-controller-agent
        securityContext:
          capabilities:
            add: ["NET_ADMIN"]

This can be saved into a file (for example "patch.yaml") and applied against a running deployment by patching it:

$ kubectl patch deployment <name> -p "$(cat patch.yaml)"

Or could be merged into a deployment manifest before creating it.

In this example "vxeth0" and "vxeth1" are the list of VXLAN network names that will be created in a pod. In this case the network interfaces created in a pod will have the same names as the specfied network names, i.e "vxeth0" and "vxeth1". This could be configured via "Network Options" (see below).

VXLAN Network Identifier

According to VXLAN specification during the setup process a VXLAN should be provided with a Segment ID or "VXLAN Network Identifier (VNI)". The controller does that automatically using the ID specified in the network options (see "Network Options" below).

Network Options

Network options define how the controller behaves setting up a particular network interface. Options could be defined in networks defining configmap or in a network defining annotation. Options defined in the configmap apply to all the VXLAN enabled pods in a cluster. Options defined in a pod annotation overrides that for the pod.

Example of the options defined in the configmap:

data:
  vxeth0: id=1000 dev=tun0
  vxeth1: id=1001 up

Example of the options defined in a pod annotation:

annotations:
  vxlan.openvnf.org/networks: |
    vxeth0
      ip=192.168.10.1/24
      route=192.168.10.0/24:192.168.100.1
    vxeth1
      up=false
annotations:
  vxlan.openvnf.org/networks: vxeth0 dev=tun0, vxeth1

The only mandatory option for now is "id" (VNI) and should be explicitly defined in either configmap or annotation.

The manifest used in the "Deployment" section defines "kube-vxlan-controller" configmap with some options for networks named "vxlan[0-3]" including "id". Therefore, the example above with "vxeth0" and "vxeth1" networks works without complains about non-existing id.

type

Network interface type that will be created in a pod. Currently the "vxlan" one is only tested and supported and a default value, therefore usually should not be specified.

id

Defines network identified (VNI in case of VXLAN). Mandatory, no default value. Usually should be specified in the configmap. Specify it in a pod annotation if you know what you do only.

name

Name of the actual network interface that will be created. Will be set to a network name if not specified.

dev

A pod network device used to create a network interface (default: "eth0").

up

Defines whether a created interface should be set up. Can be set to "true" or "false". Specifying without any value implies "true". If specified globaly in the configmap, could be overriden with "up=false".

ip

Defines an IP address that will be assigned to a created network interface.

route

Defines a routing table rule that controller can create after an interface setup. The following example describes how to create a rule for routing to the subnet "192.168.2.0/24" via the "192.168.1.1" gateway:

route=192.168.2.0/24:192.168.1.1

Controller Workflow

The controller is subscribed to the pod events using the Pod Watch API. On the "pod added" event the controller is looking for the networks annotation and sets up networks according to it (and according to the networks configmap) using the Agent init container. Thus the other init containers available in a pod can already work with the interfaces.

Once the network interfaces are created and configured according to the options, the controller sends a TERM signal to the main process of the Agent to let it terminate so that the pod could proceed with its creation.

Once a pod is running the sidecar Agent container is used to configure fdb entries to set up configured networks peers forwarding. If added or removed pod is a member of a certain network, the controller makes sure all the pods in this network get the fdb entries table updated.

The controller uses the "Pod Exec API" to execute commands in a pod via Agent container.

Example

A basic use case example demonstrates how to send packets from a pod "a" to a pod "b" provided that the pods are in the different VXLAN networks. This would require a gateway pod being a member of the both networks and the routes set up on "a" and "b". Use Example Manifest to create example workloads:

$ kubectl create -f https://raw.githubusercontent.com/openvnf/kube-vxlan-controller/master/kubernetes/example.yaml

This creates deployments and the corresponding pods "a", "b" and "gw" with the VXLAN networks and the corresponding network interfaces configured the following way:

  • pod "a":
    • network: "vxeth1", IP: "192.168.11.2/29"
    • route: "192.168.12.0/29 via 192.168.11.1"
  • pod "b":
    • network: "vxeth2", IP: "192.168.12.2/29"
    • route: "192.168.11.0/29 via 192.168.12.1"
  • pod "gw":
    • network: "vxeth1", IP: "192.168.11.1/29"
    • network: "vxeth2", IP: "192.168.12.1/29"

To check that example works as expected we can ping one pod's VXLAN network interface from another pod and vice versa:

$ POD_A=$(kubectl get po -l run=a -o jsonpath={.items[*].metadata.name})
$ POD_B=$(kubectl get po -l run=b -o jsonpath={.items[*].metadata.name})

$ kubectl exec -it $POD_A -c a ping 192.168.12.2
PING 192.168.12.2 (192.168.12.2): 56 data bytes
64 bytes from 192.168.12.2: seq=0 ttl=63 time=0.082 ms

$ kubectl exec -it $POD_B -c b ping 192.168.11.2
PING 192.168.11.2 (192.168.11.2): 56 data bytes
64 bytes from 192.168.11.2: seq=0 ttl=63 time=0.107 ms

Troubleshooting

If a pod is not get provisioned with a VXLAN interface, or pods are not pingable within a VXLAN network, there are several things to check:

Label and Annotation

The label and annotation are named according to the current configuration. To check the configuration:

$ kubectl -n kube-system get cm kube-vxlan-controller-config -o jsonpath="{.data.config}"
[{'kube-vxlan-controller', [
    {db_file, "/usr/share/kube-vxlan-controller/db"},
    {selector, "vxlan.openvnf.org"},
    {annotation, "vxlan.openvnf.org/networks"}
]}].

In this case the label should be "vxlan.openvnf.org", the annotation — "vxlan.openvnf.org/networks".

Containers

Both init agent and runtime agent containers are present in a pod and named this way:

  • init agent: "vxlan-controller-agent-init"
  • runtime agent: "vxlan-controller-agent"

Both containers should have "NET_ADMIN" capability.

Logs

Check the controller logs to see if a pod was processed:

$ POD=$(kubectl -n kube-system get po -l run=kube-vxlan-controller -o jsonpath="{.items[*].metadata.name}")
$ kubectl -n kube-system logs $POD

The logs should contain the pod name and records of attempts of creating VXLAN interfaces and provisioning FDB table.

Inspect

The VXLAN networks managed by the controller can be inspected:

$ POD=$(kubectl -n kube-system get po -l run=kube-vxlan-controller -o jsonpath="{.items[*].metadata.name}")
$ kubectl -n kube-system exec $POD kube-vxlan-controller inspect nets <Nets>

Example:

$ kubectl -n kube-system exec $POD kube-vxlan-controller inspect nets vxeth1 vxeth2
[vxeth1]
 pod: default/a-5cd8c6ccb8-kkvwt 10.234.98.195
 net: dev:eth0 id:101 ip:192.168.11.2/29 name:vxeth1 route:192.168.12.0/29:192.168.11.1 type:vxlan up:true
 fdb: 00:00:00:00:00:00 dst 10.234.72.134 self permanent

 pod: default/gw-f87979b47-5p57h 10.234.72.134
 net: dev:eth0 id:101 ip:192.168.11.1/29 name:vxeth1 type:vxlan up:true
 fdb: 00:00:00:00:00:00 dst 10.234.98.195 self permanent

[vxeth2]
 pod: default/b-cf894dbfd-t657f 10.234.101.131
 net: dev:eth0 id:102 ip:192.168.12.2/29 name:vxeth2 route:192.168.11.0/29:192.168.12.1 type:vxlan up:true
 fdb: 00:00:00:00:00:00 dst 10.234.72.134 self permanent

 pod: default/gw-f87979b47-5p57h 10.234.72.134
 net: dev:eth0 id:102 ip:192.168.12.1/29 name:vxeth2 type:vxlan up:true
 fdb: 00:00:00:00:00:00 dst 10.234.101.131 self permanent

The important part to check here is that all the desired pods are in the desired network, and "fdb" field of a particular pod has a corresponding record for each pod (by its IP address) that it supposed to be connected to.

If expected pods are missing in the corresponding "fdb" fields it worth trying to restart the pod where the records are missing.

List Pods

To list all the pods the controller is aware of:

$ POD=$(kubectl -n kube-system get po -l run=kube-vxlan-controller -o jsonpath="{.items[*].metadata.name}")
$ kubectl -n kube-system exec $POD kube-vxlan-controller list pods
default/a-5cd8c6ccb8-kkvwt vxeth1 192.168.11.2/29
default/b-cf894dbfd-t657f vxeth2 192.168.12.2/29
default/gw-f87979b47-5p57h vxeth1 192.168.11.1/29
default/gw-f87979b47-5p57h vxeth2 192.168.12.1/29

Pod name prefix only can be used to filter:

$ kubectl -n kube-system exec $POD kube-vxlan-controller list pods gw
default/gw-f87979b47-5p57h vxeth1 192.168.11.1/29
default/gw-f87979b47-5p57h vxeth2 192.168.12.1/29

Pods Connectivity

The interpod connectivity is fundamental requirement for VXLAN network to work. Each IP address listed in the "fdb" field of a given pod (see Inspect) should be pingable from that pod. If that is not the case the issue should be resolved first.

Stuck on a pod

If the last log record of the controller does not look similar to this:

Watching pods (selector: vxlan.openvnf.org) from version: 181701

and does not advance, that might be a sign of the controller stuck during execution a command on a stuck pod. Most likely the last log record contains the pod name. In this case the pod should be deleted. If the controller log records do not still advance, the controller should be restarted. Either with

$ kill -TERM 1

on the controller pod or by deleting the pod.

License

Copyright 2018-2019 Travelping GmbH

Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at

http://www.apache.org/licenses/LICENSE-2.0

Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.

About

License:Apache License 2.0


Languages

Language:Erlang 82.8%Language:Makefile 9.9%Language:Shell 6.2%Language:Dockerfile 1.2%