cloudstack-users mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Alessandro Caviglione <c.alessan...@gmail.com>
Subject Very BIG mess with networking
Date Mon, 24 Jun 2019 23:15:19 GMT
Hi guys,
I'm experiencing a big issue with networking.
First of all, we're running CS 4.11.2 on XS 6.5 with Advanced Networking.
Our XS Pool network conf was:
- eth0 Management
- eth1 empty
- eth2 Public
- eth3 empty
- eth4 + eth5: bond LACP for GuestVM

During our last maintenance last week, we decided to create a new bond
active-passive for Management (eth0 + eth1) and a new bond for Public
(eth2 + eth3)
In addition, we would change GuestVM bond from LACP to active-active.
So, we created from pool master a new bond and put eth0 + eth1 interface in.

MGT_NET_UUID=$(xe network-create name-label=Management)
PMI_PIF_UUID=$(xe pif-list host-uuid=xxx management=true params=uuid | awk
'{ print $5 }')
MGT_PIF0_UUID=$(xe pif-list host-uuid=xxx device=eth0 params=uuid | awk '{
print $5 }')
MGT_PIF1_UUID=$(xe pif-list host-uuid=xxx device=eth1 params=uuid | awk '{
print $5 }')
xe bond-create network-uuid=$MGT_NET_UUID
pif-uuids=$PMI_PIF_UUID,$MGT_PIF1_UUID mode=active-backup

I used the same method to create new bond for Public network (obviously
changing nics).

To change bond mode for GuestVM network I've used:

xe pif-list device=bond0 VLAN=-1
xe pif-param-set uuid=<Bond0 UUID> other-config:bond-mode=balance-slb

I've repeated Public and GuestVM commands on each of the three hosts in the
pool, for Management I've done it only on Pool Master.
After that I've restarted toolstack and (after the issue I'll explain) also
reboot every host.
However, this is the result:
- Existing VMs runs fine and I can stop, start, migrate
- New VM that require new network will fail
- restart network with clean will fail and makes network and instance
unavailable

This is the CS log:

2019-06-25 00:14:35,751 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Created VM
dbb045ef-5072-96ef-fbb0-1d7e3af0a0ea for r-899-VM
2019-06-25 00:14:35,756 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) PV args are
%template=domP%name=r-899-VM%eth2ip=31.44.38.53%eth2mask=255.255.255.240%gateway=31.44.38.49%eth0ip=10.122.12.1%eth0mask=255.255.255.0%domain=
tet.com
%cidrsize=24%dhcprange=10.122.12.1%eth1ip=169.254.3.241%eth1mask=255.255.0.0%type=router%disable_rp_filter=true%dns1=8.8.8.8%dns2=8.8.4.4
2019-06-25 00:14:35,757 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) HVM args are  template=domP
name=r-899-VM eth2ip=31.44.38.53 eth2mask=255.255.255.240
gateway=31.44.38.49 eth0ip=10.122.12.1 eth0mask=255.255.255.0 domain=tet.com
cidrsize=24 dhcprange=10.122.12.1 eth1ip=169.254.3.241 eth1mask=255.255.0.0
type=router disable_rp_filter=true dns1=8.8.8.8 dns2=8.8.4.4
2019-06-25 00:14:35,790 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) VBD
ec5e1e54-5902-cbcb-a6a8-abec0479d27c created for
com.cloud.agent.api.to.DiskTO@36117847
2019-06-25 00:14:35,790 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Creating VIF for r-899-VM on
nic [Nic:Public-31.44.38.53-vlan://untagged]
2019-06-25 00:14:35,792 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Looking for network named
Public
2019-06-25 00:14:35,793 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Found more than one network
with the name Public
2019-06-25 00:14:35,802 DEBUG [c.c.h.x.r.XsLocalNetwork]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Found a network called
Public on host=192.168.200.39;
 Network=9fa48b75-d68e-feaf-2eb4-8a7340f8c89b;
pif=ca4c1679-fa36-bc93-37de-28a74ddc4f2c
2019-06-25 00:14:35,807 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Created a vif
dfaab3d7-7921-e4d5-ba27-537e8d549a5c on 2
2019-06-25 00:14:35,807 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Creating VIF for r-899-VM on
nic [Nic:Guest-10.122.12.1-vlan://384]
2019-06-25 00:14:35,809 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Looking for network named
GuestVM
2019-06-25 00:14:35,825 DEBUG [c.c.h.x.r.XsLocalNetwork]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Found a network called
GuestVM on host=192.168.200.39;
 Network=300e55f0-88ff-a460-e498-e75424bc292a;
pif=b67841c5-6361-0dbf-a63d-a3e9c1b9f2fc
2019-06-25 00:14:35,826 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Creating VLAN 384 on host
192.168.200.39 on device bond0
2019-06-25 00:14:36,467 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) VLAN is created for 384.
The uuid is e34dc684-8a87-7ef6-5a49-8214011f8c3c
2019-06-25 00:14:36,480 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Created a vif
94558211-6bc6-ce64-9535-6d424b2b072c on 0
2019-06-25 00:14:36,481 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Creating VIF for r-899-VM on
nic [Nic:Control-169.254.3.241-null]
2019-06-25 00:14:36,531 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) already have a vif on dom0
for link local network
2019-06-25 00:14:36,675 DEBUG [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Created a vif
1ee58631-8524-ffa8-7ef7-e0acde48449f on 1
2019-06-25 00:14:37,688 WARN  [c.c.h.x.r.CitrixResourceBase]
(DirectAgent-35:ctx-c5073156) (logid:0d9e7907) Task failed! Task record:
              uuid: 3177dcfc-ade9-3079-00f3-191efa1f90b2
           nameLabel: Async.VM.start_on
     nameDescription:
   allowedOperations: []
   currentOperations: {}
             created: Tue Jun 25 00:14:42 CEST 2019
            finished: Tue Jun 25 00:14:42 CEST 2019
              status: failure
          residentOn: com.xensource.xenapi.Host@85c62ee8
            progress: 1.0
                type: <none/>
              result:
           errorInfo: [HOST_CANNOT_ATTACH_NETWORK,
OpaqueRef:a323ff49-04bb-3e0f-50e3-b1bc7ef31630,
OpaqueRef:f9e9b47a-28c3-5031-a8fb-5e103c970a8a]
         otherConfig: {}
           subtaskOf: com.xensource.xenapi.Task@aaf13f6f
            subtasks: []

 And this is the xensource.log

 INET 0.0.0.0:80|VBD.create R:47cd5359e955|audit] VBD.create: VM =
'dc9b4f1a-960c-8f02-a34c-32ad20e053f6 (r-899-VM)'; VDI =
'f80241ff-2a00-4565-bfa4-a980f1462f3e'
 INET 0.0.0.0:80|VBD.create R:47cd5359e955|xapi] Checking whether there's a
migrate in progress...
 INET 0.0.0.0:80|VBD.create R:47cd5359e955|xapi] VBD.create (device = 0;
uuid = f2b9e26f-66d2-d939-e2a1-04af3ae41ac9; ref =
OpaqueRef:48deb7cb-2d3b-8370-107a-68ba25213c3e)
 INET 0.0.0.0:80|VIF.create R:ac85e8d39cd8|audit] VIF.create: VM =
'dc9b4f1a-960c-8f02-a34c-32ad20e053f6 (r-899-VM)'; network =
'9fa48b75-d68e-feaf-2eb4-8a7340f8c89b'
 INET 0.0.0.0:80|VIF.create R:ac85e8d39cd8|xapi] VIF.create running
 INET 0.0.0.0:80|VIF.create R:ac85e8d39cd8|xapi] Found mac_seed on VM:
supplied MAC parameter = '1e:00:19:00:00:4e'
 INET 0.0.0.0:80|VIF.create R:ac85e8d39cd8|xapi] VIF
ref='OpaqueRef:96cfcfaa-cf0f-e8a2-6ff5-4029815999b1' created (VM =
'OpaqueRef:d3de28ef-ef26-02f7-46aa-ec01fe5c035e'; MAC address =
'1e:00:19:00:00:4e')
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|audit] VLAN.create: network =
'2ce5f897-376f-4562-46c5-6f161106584b'; VLAN tag = 363
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|xapi] Session.create
trackid=70355f146f1dd9d4873b9fdde8d1b8eb pool=true uname= originator=
is_local_superuser=true auth_user_sid=
parent=trackid=ab7ab58a3d7585f75b89abdab8725787
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|mscgen] xapi=>xapi
[label="(XML)"];
 UNIX /var/xapi/xapi||dummytaskhelper] task dispatch:session.get_uuid
D:9eca0ccc2d01 created by task R:6046e28d22a8
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|mscgen] xapi=>dst_xapi
[label="(XML)"];
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|xmlrpc_client] stunnel pid:
10191 (cached = true) connected to 192.168.200.36:443
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|xmlrpc_client]
with_recorded_stunnelpid
task_opt=OpaqueRef:6046e28d-22a8-72de-9c3e-b08b87bb1ca6 s_pid=10191
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|xmlrpc_client] stunnel pid:
10191 (cached = true) returned stunnel to cache
 INET 0.0.0.0:80|local logout in message forwarder D:457fa1a34193|xapi]
Session.destroy trackid=70355f146f1dd9d4873b9fdde8d1b8eb
 INET 0.0.0.0:80|VLAN.create R:6046e28d22a8|taskhelper] the status of
R:6046e28d22a8 is: success; cannot set it to `success
 INET 0.0.0.0:80|VIF.create R:320e6896912a|audit] VIF.create: VM =
'dc9b4f1a-960c-8f02-a34c-32ad20e053f6 (r-899-VM)'; network =
'2ce5f897-376f-4562-46c5-6f161106584b'
 INET 0.0.0.0:80|VIF.create R:320e6896912a|xapi] VIF.create running
 INET 0.0.0.0:80|VIF.create R:320e6896912a|xapi] Found mac_seed on VM:
supplied MAC parameter = '02:00:19:9e:00:02'
 INET 0.0.0.0:80|VIF.create R:320e6896912a|xapi] VIF
ref='OpaqueRef:d4db98cf-6c5c-bb9f-999a-291c33e5c0cd' created (VM =
'OpaqueRef:d3de28ef-ef26-02f7-46aa-ec01fe5c035e'; MAC address =
'02:00:19:9e:00:02')
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|audit] Host.call_plugin
host = '18d712b3-f64a-4178-9404-144f4f8fce2f (LIONARCH)'; plugin = 'vmops';
fn = 'setLinkLocalIP'; args = [ brName: xapi4 ]
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|xapi] Session.create
trackid=f3e9f44a74e7dad26498e75c6de5eeba pool=true uname= originator=
is_local_superuser=true auth_user_sid=
parent=trackid=ab7ab58a3d7585f75b89abdab8725787
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|mscgen] xapi=>xapi
[label="(XML)"];
 UNIX /var/xapi/xapi||dummytaskhelper] task dispatch:session.get_uuid
D:30eddea1c3fd created by task R:620d4a6a391b
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|mscgen] xapi=>dst_xapi
[label="(XML)"];
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|xmlrpc_client] stunnel
pid: 10224 (cached = true) connected to 192.168.200.36:443
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|xmlrpc_client]
with_recorded_stunnelpid
task_opt=OpaqueRef:620d4a6a-391b-e414-e4f3-40ea69e89d7a s_pid=10224
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|xmlrpc_client] stunnel
pid: 10224 (cached = true) returned stunnel to cache
 INET 0.0.0.0:80|local logout in message forwarder D:f89f813a5123|xapi]
Session.destroy trackid=f3e9f44a74e7dad26498e75c6de5eeba
 INET 0.0.0.0:80|host.call_plugin R:620d4a6a391b|taskhelper] the status of
R:620d4a6a391b is: success; cannot set it to `success
 INET 0.0.0.0:80|VIF.create R:1685ff9e765e|audit] VIF.create: VM =
'dc9b4f1a-960c-8f02-a34c-32ad20e053f6 (r-899-VM)'; network =
'e667320f-0e48-4cc6-6329-d723846cf8be'
 INET 0.0.0.0:80|VIF.create R:1685ff9e765e|xapi] VIF.create running
 INET 0.0.0.0:80|VIF.create R:1685ff9e765e|xapi] Found mac_seed on VM:
supplied MAC parameter = '0e:00:a9:fe:01:73'
 INET 0.0.0.0:80|VIF.create R:1685ff9e765e|xapi] VIF
ref='OpaqueRef:36a3163e-2a8e-fda1-7e13-0d6802b8ba12' created (VM =
'OpaqueRef:d3de28ef-ef26-02f7-46aa-ec01fe5c035e'; MAC address =
'0e:00:a9:fe:01:73')
|Async.VM.start_on R:96fa354a6a0d|dispatcher] spawning a new thread to
handle the current task (trackid=ab7ab58a3d7585f75b89abdab8725787)
|Async.VM.start_on R:96fa354a6a0d|audit] VM.start_on: VM =
'dc9b4f1a-960c-8f02-a34c-32ad20e053f6 (r-899-VM)'; host
'18d712b3-f64a-4178-9404-144f4f8fce2f (LIONARCH)'
|Async.VM.start_on R:96fa354a6a0d|xapi] No operations are valid because
current-operations = [ OpaqueRef:96fa354a-6a0d-da55-f1d4-15eaa51d6640 ->
attach ]
es
|Async.VM.start_on R:96fa354a6a0d|xapi] The VM's BIOS strings were not yet
filled in. The VM is now made BIOS-generic.
|Async.VM.start_on R:96fa354a6a0d|xapi] Checking whether VM
OpaqueRef:d3de28ef-ef26-02f7-46aa-ec01fe5c035e can run on host
OpaqueRef:f0e56fe4-004c-31af-4767-531e22622c1e
|Async.VM.start_on R:96fa354a6a0d|backtrace] Raised at
xapi_network_attach_helpers.ml:50.8-90 -> list.ml:69.12-15 ->
xapi_vm_helpers.ml:379.4-111
|Async.VM.start_on R:96fa354a6a0d|xapi] Caught exception while checking if
network OpaqueRef:f9e9b47a-28c3-5031-a8fb-5e103c970a8a could be attached on
host OpaqueRef:f0e56fe4-004c-31af-4767-531e22622c1e:CANNOT_PLUG_BOND_SLAVE:
[ OpaqueRef:b9238dea-802c-d8da-a7b7-ee5bc642a615 ]
|Async.VM.start_on R:96fa354a6a0d|xapi] Raised at xapi_vm_helpers.ml:390.10-134
-> list.ml:69.12-15 -> xapi_vm_helpers.ml:507.1-47 ->
message_forwarding.ml:932.5-85 -> threadext.ml:20.20-24 ->
threadext.ml:20.62-65
-> message_forwarding.ml:40.25-57 -> message_forwarding.ml:1262.9-276 ->
pervasiveext.ml:22.2-9
|Async.VM.start_on R:96fa354a6a0d|xapi] Raised at pervasiveext.ml:26.22-25
-> pervasiveext.ml:22.2-9
|Async.VM.start_on R:96fa354a6a0d|xapi] Raised at pervasiveext.ml:26.22-25
-> pervasiveext.ml:22.2-9
|Async.VM.start_on R:96fa354a6a0d|backtrace] Raised at pervasiveext.ml:26.22-25
-> message_forwarding.ml:1248.3-1023 -> rbac.ml:229.16-23
|Async.VM.start_on R:96fa354a6a0d|backtrace] Raised at rbac.ml:238.10-15 ->
server_helpers.ml:79.11-41
|Async.VM.start_on R:96fa354a6a0d|dispatcher] Server_helpers.exec
exception_handler: Got exception HOST_CANNOT_ATTACH_NETWORK: [
OpaqueRef:f0e56fe4-004c-31af-4767-531e22622c1e;
OpaqueRef:f9e9b47a-28c3-5031-a8fb-5e103c970a8a ]

Please, PLEASE, let me know that someone knows how to fix it!

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message