Return-Path: X-Original-To: apmail-incubator-cloudstack-users-archive@minotaur.apache.org Delivered-To: apmail-incubator-cloudstack-users-archive@minotaur.apache.org Received: from mail.apache.org (hermes.apache.org [140.211.11.3]) by minotaur.apache.org (Postfix) with SMTP id CD525D88A for ; Wed, 5 Sep 2012 21:22:40 +0000 (UTC) Received: (qmail 27201 invoked by uid 500); 5 Sep 2012 21:22:40 -0000 Delivered-To: apmail-incubator-cloudstack-users-archive@incubator.apache.org Received: (qmail 27175 invoked by uid 500); 5 Sep 2012 21:22:40 -0000 Mailing-List: contact cloudstack-users-help@incubator.apache.org; run by ezmlm Precedence: bulk List-Help: List-Unsubscribe: List-Post: List-Id: Reply-To: cloudstack-users@incubator.apache.org Delivered-To: mailing list cloudstack-users@incubator.apache.org Received: (qmail 27167 invoked by uid 99); 5 Sep 2012 21:22:40 -0000 Received: from nike.apache.org (HELO nike.apache.org) (192.87.106.230) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 05 Sep 2012 21:22:40 +0000 X-ASF-Spam-Status: No, hits=-0.0 required=5.0 tests=RCVD_IN_DNSWL_NONE,SPF_PASS X-Spam-Check-By: apache.org Received-SPF: pass (nike.apache.org: local policy) Received: from [208.97.132.119] (HELO homiemail-a63.g.dreamhost.com) (208.97.132.119) by apache.org (qpsmtpd/0.29) with ESMTP; Wed, 05 Sep 2012 21:22:33 +0000 Received: from homiemail-a63.g.dreamhost.com (localhost [127.0.0.1]) by homiemail-a63.g.dreamhost.com (Postfix) with ESMTP id 2631B2F401D for ; Wed, 5 Sep 2012 14:22:11 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha1; c=relaxed; d=nfinausa.com; h=message-id :date:from:mime-version:to:subject:references:in-reply-to :content-type:content-transfer-encoding; s=nfinausa.com; bh=Qn3C TsUE84lZ2q9NjWuRVi65g/c=; b=peUdY2zQ3WFGpuXPk1K3KIg0mj1nvk3JCzPD bzgxdWlSQJgqKH4B/R4G8QzDH6aS9JqFZsRuuOahYXJPO3fnV8RmGT1XM2jv5l70 aofPEtaFjJhK0ntyDXlKNF/o0BATbFFte8NZm8gYloJWJvuZZmhYinQYPzDUEFVQ aqGeN3c= Received: from [172.16.1.99] (adsl-98-90-202-241.mob.bellsouth.net [98.90.202.241]) (using TLSv1 with cipher DHE-RSA-AES256-SHA (256/256 bits)) (No client certificate requested) (Authenticated sender: nik.martin@nfinausa.com) by homiemail-a63.g.dreamhost.com (Postfix) with ESMTPSA id 3F7C12F4071 for ; Wed, 5 Sep 2012 14:22:09 -0700 (PDT) Message-ID: <5047C27F.4050602@nfinausa.com> Date: Wed, 05 Sep 2012 16:22:07 -0500 From: Nik Martin Organization: Nfina Technologies, Inc. User-Agent: Mozilla/5.0 (X11; Linux i686; rv:15.0) Gecko/20120827 Thunderbird/15.0 MIME-Version: 1.0 To: cloudstack-users@incubator.apache.org Subject: Re: Cannot use NIC bond on guest/public networks References: <504795CD.3090305@nfinausa.com> In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit On 09/05/2012 01:30 PM, Anthony Xu wrote: > Hi Nik, > > You need to move ip configuration from eth* to bond* device. > > Check NIC Bonding for XenServer (Optional) in http://support.citrix.com/servlet/KbServlet/download/31038-102-685337/CloudPlatform3.0.3_3.0.4InstallGuide.pdf > > > Anthony > Thank you. Xencenter makes it LOOK like it renames the slave interfaces, when in fact they are still named "guest_pub". Xencenter shows them as "guest_pub(slave)". I can get VMs added and started properly now, but I still cannot communicate with xenserver 6.02 hosts with bonded management interfaces. This is a xenserver/openvswitch issue though. Regards, Nik > >> -----Original Message----- >> From: Nik Martin [mailto:nik.martin@nfinausa.com] >> Sent: Wednesday, September 05, 2012 11:11 AM >> To: cloudstack-users@incubator.apache.org >> Subject: Cannot use NIC bond on guest/public networks >> >> I'm having an issue using bonded interfaces on gguest and public >> networks. Details: >> >> 3 HVs, Xenserver 6.02 all patches applied, open vswitch backend, sixe >> nics in each, configured as: >> eth0 management >> eth1+3 bonded, tagged as guest_pub >> eth2, unused(management backup) >> eth4+5 bonded, tagged as storage, which is iSCSI based >> >> Each bond slave connects to pair of L3 switches that are properly >> stacked >> >> If I unbond bond 1+3, vms are fine, guest and public traffic flows fine. >> If I stop cloud-management service, open xencenter and bond eth1+3, >> then name the resulting network "guest_pub", it now carries the label >> for my network tag for guest and public traffic. >> >> When I restart cloud-management, as soon as my system VMs try to start, >> they start bouncing wildly in xencenter. The VIF is created, and looks >> like "VLAN-SOMEUUID-15, where 15 is the vlan for my public network. >> The >> VIF never gets assigned to the bond. If I look in the cloudstack log, >> I >> see: >> >> (DirectAgent-97:null) PV args are -- quiet >> console=hvc0%template=domP%type=consoleproxy%host=172.16.5.2%port=8250% >> name=v-2- >> VM%premium=true%zone=1%pod=1%guid=Proxy.2%proxy_vm=2%disable_rp_filter= >> true%eth2ip=172.16.15.118%eth2mask=255.255.255.0%gateway=172.16.15.1%et >> h0ip=169.254.2.227%eth0mask=255.255.0.0%eth1ip=172.16.5.34%eth1mask=255 >> .255.255.0%mgmtcidr=172.16.5.0/24%localgw=172.16.5.1%internaldns1=8.8.8 >> .8%internaldns2=8.8.4.4%dns1=8.8.8.8%dns2=8.8.4.4 >> 2012-09-05 12:54:31,762 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) VBD 14e0d8bc-8b5b-2316-cd1d-07c04389e242 created >> for Vol[2|ROOT|a77d8900-aa57-4ce9-9c73-6db35d985f48|2147483648] >> 2012-09-05 12:54:31,824 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Creating VIF for v-2-VM on nic >> [Nic:Public-172.16.15.118-vlan://15] >> 2012-09-05 12:54:31,824 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Looking for network named guest_pub >> 2012-09-05 12:54:31,829 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Found more than one network with the name >> guest_pub >> 2012-09-05 12:54:31,845 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Found a network called guest_pub on >> host=172.16.5.5; Network=f008aa20-1ea1-f6e5-c881-08911e3549e5; >> pif=64d27657-d3be-df29-d365-0baa2dab5303 >> 2012-09-05 12:54:31,855 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Creating VLAN 15 on host 172.16.5.5 on device >> eth1 >> 2012-09-05 12:54:31,885 WARN [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Catch Exception: class >> com.xensource.xenapi.Types$XenAPIException due to >> CANNOT_ADD_VLAN_TO_BOND_SLAVEOpaqueRef:70faa00e-7444-ee41-3cf3- >> 7d151b6fd561 >> CANNOT_ADD_VLAN_TO_BOND_SLAVEOpaqueRef:70faa00e-7444-ee41-3cf3- >> 7d151b6fd561 >> at com.xensource.xenapi.Types.checkResponse(Types.java:1731) >> at com.xensource.xenapi.Connection.dispatch(Connection.java:372) >> at >> com.cloud.hypervisor.xen.resource.XenServerConnectionPool$XenServerConn >> ection.dispatch(XenServerConnectionPool.java:905) >> at com.xensource.xenapi.VLAN.create(VLAN.java:349) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.enableVlanNetwork( >> CitrixResourceBase.java:3646) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.getNetwork(CitrixR >> esourceBase.java:635) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.createVif(CitrixRe >> sourceBase.java:671) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.execute(CitrixReso >> urceBase.java:1104) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.executeRequest(Cit >> rixResourceBase.java:466) >> at >> com.cloud.hypervisor.xen.resource.XenServer56Resource.executeRequest(Xe >> nServer56Resource.java:69) >> at >> com.cloud.agent.manager.DirectAgentAttache$Task.run(DirectAgentAttache. >> java:187) >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> at >> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >> at java.util.concurrent.FutureTask.run(FutureTask.java:166) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ac >> cess$101(ScheduledThreadPoolExecutor.java:165) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ru >> n(ScheduledThreadPoolExecutor.java:266) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >> va:1110) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >> ava:603) >> at java.lang.Thread.run(Thread.java:679) >> 2012-09-05 12:54:31,886 WARN [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Unable to start v-2-VM due to >> CANNOT_ADD_VLAN_TO_BOND_SLAVEOpaqueRef:70faa00e-7444-ee41-3cf3- >> 7d151b6fd561 >> at com.xensource.xenapi.Types.checkResponse(Types.java:1731) >> at com.xensource.xenapi.Connection.dispatch(Connection.java:372) >> at >> com.cloud.hypervisor.xen.resource.XenServerConnectionPool$XenServerConn >> ection.dispatch(XenServerConnectionPool.java:905) >> at com.xensource.xenapi.VLAN.create(VLAN.java:349) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.enableVlanNetwork( >> CitrixResourceBase.java:3646) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.getNetwork(CitrixR >> esourceBase.java:635) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.createVif(CitrixRe >> sourceBase.java:671) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.execute(CitrixReso >> urceBase.java:1104) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.executeRequest(Cit >> rixResourceBase.java:466) >> at >> com.cloud.hypervisor.xen.resource.XenServer56Resource.executeRequest(Xe >> nServer56Resource.java:69) >> at >> com.cloud.agent.manager.DirectAgentAttache$Task.run(DirectAgentAttache. >> java:187) >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> at >> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >> at java.util.concurrent.FutureTask.run(FutureTask.java:166) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ac >> cess$101(ScheduledThreadPoolExecutor.java:165) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ru >> n(ScheduledThreadPoolExecutor.java:266) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >> va:1110) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >> ava:603) >> at java.lang.Thread.run(Thread.java:679) >> 2012-09-05 12:54:31,977 WARN [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Unable to clean up VBD due to >> You gave an invalid object reference. The object may have recently >> been >> deleted. The class parameter gives the type of reference given, and >> the >> handle parameter echoes the bad value given. >> at com.xensource.xenapi.Types.checkResponse(Types.java:211) >> at com.xensource.xenapi.Connection.dispatch(Connection.java:372) >> at >> com.cloud.hypervisor.xen.resource.XenServerConnectionPool$XenServerConn >> ection.dispatch(XenServerConnectionPool.java:905) >> at com.xensource.xenapi.VBD.unplug(VBD.java:1058) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.handleVmStartFailu >> re(CitrixResourceBase.java:929) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.execute(CitrixReso >> urceBase.java:1167) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.executeRequest(Cit >> rixResourceBase.java:466) >> at >> com.cloud.hypervisor.xen.resource.XenServer56Resource.executeRequest(Xe >> nServer56Resource.java:69) >> at >> com.cloud.agent.manager.DirectAgentAttache$Task.run(DirectAgentAttache. >> java:187) >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> at >> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >> at java.util.concurrent.FutureTask.run(FutureTask.java:166) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ac >> cess$101(ScheduledThreadPoolExecutor.java:165) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ru >> n(ScheduledThreadPoolExecutor.java:266) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >> va:1110) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >> ava:603) >> at java.lang.Thread.run(Thread.java:679) >> 2012-09-05 12:54:31,996 WARN [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) Unable to clean up VBD due to >> You gave an invalid object reference. The object may have recently >> been >> deleted. The class parameter gives the type of reference given, and >> the >> handle parameter echoes the bad value given. >> at com.xensource.xenapi.Types.checkResponse(Types.java:211) >> at com.xensource.xenapi.Connection.dispatch(Connection.java:372) >> at >> com.cloud.hypervisor.xen.resource.XenServerConnectionPool$XenServerConn >> ection.dispatch(XenServerConnectionPool.java:905) >> at com.xensource.xenapi.VBD.unplug(VBD.java:1058) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.handleVmStartFailu >> re(CitrixResourceBase.java:929) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.execute(CitrixReso >> urceBase.java:1167) >> at >> com.cloud.hypervisor.xen.resource.CitrixResourceBase.executeRequest(Cit >> rixResourceBase.java:466) >> at >> com.cloud.hypervisor.xen.resource.XenServer56Resource.executeRequest(Xe >> nServer56Resource.java:69) >> at >> com.cloud.agent.manager.DirectAgentAttache$Task.run(DirectAgentAttache. >> java:187) >> at >> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >> at >> java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:334) >> at java.util.concurrent.FutureTask.run(FutureTask.java:166) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ac >> cess$101(ScheduledThreadPoolExecutor.java:165) >> at >> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.ru >> n(ScheduledThreadPoolExecutor.java:266) >> at >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.ja >> va:1110) >> at >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.j >> ava:603) >> at java.lang.Thread.run(Thread.java:679) >> 2012-09-05 12:54:31,996 DEBUG [xen.resource.CitrixResourceBase] >> (DirectAgent-97:null) The VM is in stopped state, detected problem >> during startup : v-2-VM >> 2012-09-05 12:54:31,996 DEBUG [agent.manager.DirectAgentAttache] >> (DirectAgent-97:null) Seq 4-204669124: Cancelling because one of the >> answers is false and it is stop on error. >> 2012-09-05 12:54:31,996 DEBUG [agent.manager.DirectAgentAttache] >> (DirectAgent-97:null) Seq 4-204669124: Response Received: >> 2012-09-05 12:54:31,997 DEBUG [agent.transport.Request] >> (DirectAgent-97:null) Seq 4-204669124: Processing: { Ans: , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 110, >> [{"StartAnswer":{"vm":{"id":2,"name":"v-2- >> VM","bootloader":"PyGrub","type":"ConsoleProxy","cpus":1,"speed":500,"m >> inRam":1073741824,"maxRam":1073741824,"arch":"x86_64","os":"Debian >> GNU/Linux 5.0 (32-bit)","bootArgs":" template=domP type=consoleproxy >> host=172.16.5.2 port=8250 name=v-2-VM premium=true zone=1 pod=1 >> guid=Proxy.2 proxy_vm=2 disable_rp_filter=true eth2ip=172.16.15.118 >> eth2mask=255.255.255.0 gateway=172.16.15.1 eth0ip=169.254.2.227 >> eth0mask=255.255.0.0 eth1ip=172.16.5.34 eth1mask=255.255.255.0 >> mgmtcidr=172.16.5.0/24 localgw=172.16.5.1 internaldns1=8.8.8.8 >> internaldns2=8.8.4.4 dns1=8.8.8.8 >> dns2=8.8.4.4","rebootOnCrash":false,"enableHA":false,"limitCpuUse":fals >> e,"vncPassword":"9d8805d384dc10ce","params":{},"disks":[{"id":2,"name": >> "ROOT-2","mountPoint":"/iqn.2012- >> 01.com.nfinausa:san01/1","path":"a77d8900-aa57-4ce9-9c73- >> 6db35d985f48","size":2147483648,"type":"ROOT","storagePoolType":"IscsiL >> UN","storagePoolUuid":"776299a8-a770-30f7-9f6b- >> 8c422790f5b7","deviceId":0}],"nics":[{"deviceId":2,"networkRateMbps":- >> 1,"defaultNic":true,"ip":"172.16.15.118","netmask":"255.255.255.0","gat >> eway":"172.16.15.1","mac":"06:18:16:00:00:27","dns1":"8.8.8.8","dns2":" >> 8.8.4.4","broadcastType":"Vlan","type":"Public","broadcastUri":"vlan:// >> 15","isolationUri":"vlan://15","isSecurityGroupEnabled":false,"name":"g >> uest_pub"},{"deviceId":0,"networkRateMbps":- >> 1,"defaultNic":false,"ip":"169.254.2.227","netmask":"255.255.0.0","gate >> way":"169.254.0.1","mac":"0e:00:a9:fe:02:e3","broadcastType":"LinkLocal >> ","type":"Control","isSecurityGroupEnabled":false},{"deviceId":1,"netwo >> rkRateMbps":-1,"defaultNic":f >> alse,"ip" >> :"172.16.5.34","netmask":"255.255.255.0","gateway":"172.16.5.1","mac":" >> 06:58:58:00:00:0f","broadcastType":"Native","type":"Management","isSecu >> rityGroupEnabled":false,"name":"management"}]},"result":false,"details" >> :"Unable >> to start v-2-VM due to ","wait":0}}] } >> 2012-09-05 12:54:31,997 WARN [cloud.vm.VirtualMachineManagerImpl] >> (DirectAgent-97:null) Cleanup failed due to Unable to start v-2-VM due >> to >> 2012-09-05 12:54:31,997 DEBUG [agent.transport.Request] >> (consoleproxy-1:null) Seq 4-204669124: Received: { Ans: , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 110, { StartAnswer } } >> 2012-09-05 12:54:31,997 DEBUG [agent.manager.AgentAttache] >> (DirectAgent-97:null) Seq 4-204669124: No more commands found >> 2012-09-05 12:54:31,997 WARN [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Cleanup failed due to Unable to start v-2-VM due >> to >> 2012-09-05 12:54:32,002 INFO [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Unable to start VM on Host[-4-Routing] due to >> Unable to start v-2-VM due to >> 2012-09-05 12:54:32,005 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Cleaning up resources for the vm >> VM[ConsoleProxy|v-2-VM] in Starting state >> 2012-09-05 12:54:32,006 DEBUG [agent.transport.Request] >> (consoleproxy-1:null) Seq 4-204669125: Sending { Cmd , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 100111, >> [{"StopCommand":{"isProxy":false,"vmName":"v-2-VM","wait":0}}] } >> 2012-09-05 12:54:32,006 DEBUG [agent.transport.Request] >> (consoleproxy-1:null) Seq 4-204669125: Executing: { Cmd , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 100111, >> [{"StopCommand":{"isProxy":false,"vmName":"v-2-VM","wait":0}}] } >> 2012-09-05 12:54:32,006 DEBUG [agent.manager.DirectAgentAttache] >> (DirectAgent-104:null) Seq 4-204669125: Executing request >> 2012-09-05 12:54:32,097 INFO [xen.resource.CitrixResourceBase] >> (DirectAgent-104:null) VM does not exist on >> XenServer0c091805-ebac-4279-9899-52018163a557 >> 2012-09-05 12:54:32,097 DEBUG [agent.manager.DirectAgentAttache] >> (DirectAgent-104:null) Seq 4-204669125: Response Received: >> 2012-09-05 12:54:32,098 DEBUG [agent.transport.Request] >> (DirectAgent-104:null) Seq 4-204669125: Processing: { Ans: , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 110, >> [{"StopAnswer":{"vncPort":0,"bytesSent":0,"bytesReceived":0,"result":tr >> ue,"details":"VM >> does not exist","wait":0}}] } >> 2012-09-05 12:54:32,098 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (DirectAgent-104:null) Cleanup succeeded. Details VM does not exist >> 2012-09-05 12:54:32,098 DEBUG [agent.transport.Request] >> (consoleproxy-1:null) Seq 4-204669125: Received: { Ans: , MgmtId: >> 130577622632, via: 4, Ver: v1, Flags: 110, { StopAnswer } } >> 2012-09-05 12:54:32,098 DEBUG [agent.manager.AgentAttache] >> (DirectAgent-104:null) Seq 4-204669125: No more commands found >> 2012-09-05 12:54:32,098 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Cleanup succeeded. Details VM does not exist >> 2012-09-05 12:54:32,119 DEBUG [dc.dao.DataCenterIpAddressDaoImpl] >> (consoleproxy-1:null) Releasing ip address for >> reservationId=636f327b-5183-46bd-adf2-57bf0d5fae80, instance=7 >> 2012-09-05 12:54:32,125 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Successfully cleanued up resources for the vm >> VM[ConsoleProxy|v-2-VM] in Starting state >> 2012-09-05 12:54:32,126 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Root volume is ready, need to place VM in >> volume's >> cluster >> 2012-09-05 12:54:32,126 DEBUG [cloud.vm.VirtualMachineManagerImpl] >> (consoleproxy-1:null) Vol[2|vm=2|ROOT] is READY, changing deployment >> plan to use this pool's dcId: 1 , podId: 1 , and clusterId: 1 >> 2012-09-05 12:54:32,127 DEBUG [cloud.deploy.FirstFitPlanner] >> (consoleproxy-1:null) DeploymentPlanner allocation algorithm: random >> 2012-09-05 12:54:32,127 DEBUG [cloud.deploy.FirstFitPlanner] >> (consoleproxy-1:null) Trying to allocate a host and storage pools from >> dc:1, pod:1,cluster:1, requested cpu: 500, requested ram: 1073741824 >> >> >> >> -- >> Regards, >> >> Nik >> >> Nik Martin >> VP Business Development >> Nfina Technologies, Inc. >> +1.251.243.0043 x1003 >> Relentless Reliability -- Regards, Nik Nik Martin VP Business Development Nfina Technologies, Inc. +1.251.243.0043 x1003 Relentless Reliability