[COE-36] Get CoE as-is manually running on K8s an OpenStack (RDO) Cloud Created: 05/Dec/18 Updated: 19/Dec/18 Resolved: 19/Dec/18 |
|
| Status: | Resolved |
| Project: | coe |
| Component/s: | None |
| Affects Version/s: | None |
| Fix Version/s: | None |
| Type: | Story | Priority: | Medium |
| Reporter: | Michael Vorburger | Assignee: | Michael Vorburger |
| Resolution: | Done | Votes: | 0 |
| Labels: | None | ||
| Remaining Estimate: | Not Specified | ||
| Time Spent: | Not Specified | ||
| Original Estimate: | Not Specified | ||
| Issue Links: |
|
||||||||||||||||||||
| Description |
|
https://github.com/vorburger/opendaylight-coe-kubernetes-openshift ... The goal of this first step is to Manually “cobble together” COE run on K8s on RDO Cloud. |
| Comments |
| Comment by Michael Vorburger [ 05/Dec/18 ] |
|
fkautz & shague I'm currently stuck at this, when can you help me get past this: [fedora@coe-node ~]$ sudo ovs-vsctl show e29c4851-cc9d-4833-ba3a-906d87fd72b7 Manager "tcp:192.168.0.5:6640" is_connected: true Bridge br-int Controller "tcp:192.168.0.5:6653" is_connected: true Port "vethb206b14b" Interface "vethb206b14b" error: "could not open network device vethb206b14b (No such device)" Port "veth741cafb2" Interface "veth741cafb2" Port "veth9e3d3ff1" Interface "veth9e3d3ff1" error: "could not open network device veth9e3d3ff1 (No such device)" Port "vethae78875f" Interface "vethae78875f" Port "enp0s9" Interface "enp0s9" type: system error: "could not open network device enp0s9 (No such device)" Port br-int Interface br-int type: internal ovs_version: "2.8.1" |
| Comment by Michael Vorburger [ 05/Dec/18 ] |
|
Above is from the node (minion), this is the master:
[fedora@coe-master ~]$ sudo ovs-vsctl show
0b2f6636-f32f-4bc0-a6f9-651b61563aae
Manager "tcp:192.168.0.5:6640"
is_connected: true
Bridge br-int
Controller "tcp:192.168.0.5:6653"
is_connected: true
Port "veth3c325425"
Interface "veth3c325425"
Port "veth6c86e2c9"
Interface "veth6c86e2c9"
Port br-int
Interface br-int
type: internal
ovs_version: "2.8.1"
|
| Comment by Michael Vorburger [ 05/Dec/18 ] |
|
Doing the following: [fedora@coe-node ~]$ sudo ovs-vsctl set O . other_config:local_ip=192.168.0.8 [fedora@coe-master ~]$ sudo ovs-vsctl set O . other_config:local_ip=192.168.0.5 added this tunb5dbe23b682 which k.faseela says is good news: [fedora@coe-node ~]$ sudo ovs-vsctl show
e29c4851-cc9d-4833-ba3a-906d87fd72b7
Manager "tcp:192.168.0.5:6640"
is_connected: true
Bridge br-int
Controller "tcp:192.168.0.5:6653"
is_connected: true
Port "vethb206b14b"
Interface "vethb206b14b"
error: "could not open network device vethb206b14b (No such device)"
Port "veth741cafb2"
Interface "veth741cafb2"
error: "could not open network device veth741cafb2 (No such device)"
Port "veth9e3d3ff1"
Interface "veth9e3d3ff1"
error: "could not open network device veth9e3d3ff1 (No such device)"
Port "tunb5dbe23b682"
Interface "tunb5dbe23b682"
type: vxlan
options: {key=flow, local_ip="192.168.0.8", remote_ip="192.168.0.5"}
Port "vethae78875f"
Interface "vethae78875f"
error: "could not open network device vethae78875f (No such device)"
Port "enp0s9"
Interface "enp0s9"
type: system
error: "could not open network device enp0s9 (No such device)"
Port br-int
Interface br-int
type: internal
ovs_version: "2.8.1"
but the veth errors are still there. I'm also hitting COE-42 and |
| Comment by Faseela K [ 05/Dec/18 ] |
|
These veth ports error won't cause any issue for you, they are actually stale veth ports that remain. I have seen the same issue, but have not clearly seen when they start showing up, but ofcourse this is a bug. |
| Comment by Michael Vorburger [ 05/Dec/18 ] |
|
OK then I'll ignore that issue (veth "could not open network device") here, and just gocus on tgetting the ping test working, and we can look at this problem later; I've created |
| Comment by Michael Vorburger [ 10/Dec/18 ] |
|
I'm currently blocked by the various isuses linked to COE-10, and need help from CoE experts to proceed. |
| Comment by Faseela K [ 10/Dec/18 ] |
|
vorburger : I just now noticed that COE CSIT is broken on master, with the same error as you have reported |
| Comment by Michael Vorburger [ 10/Dec/18 ] |
|
Sure I could try that (probably only on Wednesday). So the basic CoE use case already worked in Fluorine? |
| Comment by Faseela K [ 10/Dec/18 ] |
|
Everything was working fine all the way from stable/oxygen.. Something is broken recently.. I will take a look |
| Comment by Michael Vorburger [ 10/Dec/18 ] |
|
I'll first try out the patch on |
| Comment by Michael Vorburger [ 17/Dec/18 ] |
|
The inter-container ping on a single node VM meanwhile works in my (scipted) test environment on the RDO Cloud. I'll set up a 2nd node and test pings between busybox containers on different nodes next. |
| Comment by Michael Vorburger [ 18/Dec/18 ] |
|
https://lists.opendaylight.org/pipermail/coe-dev/2018-December/000449.html |
| Comment by Michael Vorburger [ 19/Dec/18 ] |
|
It works now, see coe-dev list reply. (This did the trick.) With this, https://github.com/vorburger/opendaylight-coe-kubernetes-openshift lets you create a K8s cluster with CoE SDN on OpenStack. The next step is to containerizing odlovs-cni, watcher & ODL and reducing the number of manual steps and moving pieces. I've used 1 master and 2 worker nodes VMs (like CSIT), but my scripts would letus create 99 works nodes as well, which could make this a useful set-up for CoE scale testing in the future. Considering this task Done. |
| Comment by Michael Vorburger [ 19/Dec/18 ] |
|
The various issues I've seen here are now linked to COE-46 in addition to COE-10. |