[COE-36] Get CoE as-is manually running on K8s an OpenStack (RDO) Cloud Created: 05/Dec/18  Updated: 19/Dec/18  Resolved: 19/Dec/18

Status: Resolved
Project: coe
Component/s: None
Affects Version/s: None
Fix Version/s: None

Type: Story Priority: Medium
Reporter: Michael Vorburger Assignee: Michael Vorburger
Resolution: Done Votes: 0
Labels: None
Remaining Estimate: Not Specified
Time Spent: Not Specified
Original Estimate: Not Specified

Issue Links:
Blocks
blocks COE-10 COE Deployment and Packaging for Kube... Open
blocks COE-37 Get CoE fully containerized running o... Open
blocks COE-39 Get CoE as-is manually running on Ope... Open
is blocked by NETVIRT-1532 NFE in CoeUtils.buildElanInstance() Resolved

 Description   

https://github.com/vorburger/opendaylight-coe-kubernetes-openshift ...

The goal of this first step is to Manually “cobble together” COE run on K8s on RDO Cloud.



 Comments   
Comment by Michael Vorburger [ 05/Dec/18 ]

fkautz & shague I'm currently stuck at this, when can you help me get past this:

[fedora@coe-node ~]$ sudo ovs-vsctl show
e29c4851-cc9d-4833-ba3a-906d87fd72b7
    Manager "tcp:192.168.0.5:6640"
        is_connected: true
    Bridge br-int
        Controller "tcp:192.168.0.5:6653"
            is_connected: true
        Port "vethb206b14b"
            Interface "vethb206b14b"
                error: "could not open network device vethb206b14b (No such device)"
        Port "veth741cafb2"
            Interface "veth741cafb2"
        Port "veth9e3d3ff1"
            Interface "veth9e3d3ff1"
                error: "could not open network device veth9e3d3ff1 (No such device)"
        Port "vethae78875f"
            Interface "vethae78875f"
        Port "enp0s9"
            Interface "enp0s9"
                type: system
                error: "could not open network device enp0s9 (No such device)"
        Port br-int
            Interface br-int
                type: internal
    ovs_version: "2.8.1"
Comment by Michael Vorburger [ 05/Dec/18 ]

Above is from the node (minion), this is the master:

[fedora@coe-master ~]$ sudo ovs-vsctl show
0b2f6636-f32f-4bc0-a6f9-651b61563aae
    Manager "tcp:192.168.0.5:6640"
        is_connected: true
    Bridge br-int
        Controller "tcp:192.168.0.5:6653"
            is_connected: true
        Port "veth3c325425"
            Interface "veth3c325425"
        Port "veth6c86e2c9"
            Interface "veth6c86e2c9"
        Port br-int
            Interface br-int
                type: internal
    ovs_version: "2.8.1"

 

Comment by Michael Vorburger [ 05/Dec/18 ]

Doing the following:

[fedora@coe-node ~]$ sudo ovs-vsctl set O . other_config:local_ip=192.168.0.8

[fedora@coe-master ~]$ sudo ovs-vsctl set O . other_config:local_ip=192.168.0.5

added this tunb5dbe23b682 which k.faseela says is good news:

[fedora@coe-node ~]$ sudo ovs-vsctl show
e29c4851-cc9d-4833-ba3a-906d87fd72b7
    Manager "tcp:192.168.0.5:6640"
        is_connected: true
    Bridge br-int
        Controller "tcp:192.168.0.5:6653"
            is_connected: true
        Port "vethb206b14b"
            Interface "vethb206b14b"
                error: "could not open network device vethb206b14b (No such device)"
        Port "veth741cafb2"
            Interface "veth741cafb2"
                error: "could not open network device veth741cafb2 (No such device)"
        Port "veth9e3d3ff1"
            Interface "veth9e3d3ff1"
                error: "could not open network device veth9e3d3ff1 (No such device)"
        Port "tunb5dbe23b682"
            Interface "tunb5dbe23b682"
                type: vxlan
                options: {key=flow, local_ip="192.168.0.8", remote_ip="192.168.0.5"}
        Port "vethae78875f"
            Interface "vethae78875f"
                error: "could not open network device vethae78875f (No such device)"
        Port "enp0s9"
            Interface "enp0s9"
                type: system
                error: "could not open network device enp0s9 (No such device)"
        Port br-int
            Interface br-int
                type: internal
    ovs_version: "2.8.1"

but the veth errors are still there. I'm also hitting COE-42 and NETVIRT-1532 ...

Comment by Faseela K [ 05/Dec/18 ]

These veth ports error won't cause any issue for you, they are actually stale veth ports that remain. I have seen the same issue, but have not clearly seen when they start showing up, but ofcourse this is a bug.

Comment by Michael Vorburger [ 05/Dec/18 ]

OK then I'll ignore that issue (veth "could not open network device") here, and just gocus on tgetting the ping test working, and we can look at this problem later; I've created COE-43 so that we don't forget it.

Comment by Michael Vorburger [ 10/Dec/18 ]

I'm currently blocked by the various isuses linked to COE-10, and need help from CoE experts to proceed.

Comment by Faseela K [ 10/Dec/18 ]

vorburger : I just now noticed that COE CSIT is broken on master, with the same error as you have reported  stable/fluorine works fine though. Could you please checkout fluorine ODL, as well as build fluorine cni plugin and watcher, and test this out?

Comment by Michael Vorburger [ 10/Dec/18 ]

Sure I could try that (probably only on Wednesday). So the basic CoE use case already worked in Fluorine?

Comment by Faseela K [ 10/Dec/18 ]

Everything was working fine all the way from stable/oxygen.. Something is broken recently.. I will take a look

Comment by Michael Vorburger [ 10/Dec/18 ]

I'll first try out the patch on NETVIRT-1533 to see if that helps; if not, I'll attempt FluorineOxygen instead.

Comment by Michael Vorburger [ 17/Dec/18 ]

The inter-container ping on a single node VM meanwhile works in my (scipted) test environment on the RDO Cloud.

I'll set up a 2nd node and test pings between busybox containers on different nodes next.

Comment by Michael Vorburger [ 18/Dec/18 ]

https://lists.opendaylight.org/pipermail/coe-dev/2018-December/000449.html

Comment by Michael Vorburger [ 19/Dec/18 ]

It works now, see coe-dev list reply. (This did the trick.)

With this, https://github.com/vorburger/opendaylight-coe-kubernetes-openshift lets you create a K8s cluster with CoE SDN on OpenStack. The next step is to containerizing odlovs-cni, watcher & ODL and reducing the number of manual steps and moving pieces.

I've used 1 master and 2 worker nodes VMs (like CSIT), but my scripts would letus create 99 works nodes as well, which could make this a useful set-up for CoE scale testing in the future.

Considering this task Done.

Comment by Michael Vorburger [ 19/Dec/18 ]

The various issues I've seen here are now linked to COE-46 in addition to COE-10.

Generated at Wed Feb 07 19:15:27 UTC 2024 using Jira 8.20.10#820010-sha1:ace47f9899e9ee25d7157d59aa17ab06aee30d3d.