cloudstack-users mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Indra Pramana <in...@sg.or.id>
Subject Re: Problem in adding Ceph RBD as primary storage for CloudStack 4.1.0
Date Mon, 15 Jul 2013 07:19:08 GMT
Hi Prasanna and all,

I managed to create a VM instance using the NFS primary storage (not using
the RBD primary storage due to the bug we discussed on another thread).

However, there is no public IP being assigned to the VM instance (unlike
the system VMs which are getting public IPs). How do I assign public IP to
the instance?

===
NIC 1 (Default)
Network Name    Indra
IP Address    10.19.8.186
Type    Isolated
Gateway    10.19.0.1
Netmask    255.255.240.0
Is Default    Yes
===

The network was created under the
"DefaultIsolatedNetworkOfferingWithSourceNatService" network offering. I am
not able to create the network using different network offerings, despite
the other offerings are available under Service Offerings > Network
Offerings and the state of those network offerings are enabled. Is it
related to the way how I configured the network bridges and configuration
earlier during zone creation, anything that I might have missed out?

I also tried to "view console" to the VM (via the web GUI) and not able to
do so. Earlier I got the "Unable to start console session as connection is
refused by the machine you are accessing" error message, which I have fixed
by changing the /etc/libvirt/qemu.conf to enable this line:

vnc_listen = "0.0.0.0"

because initially the VNC only listens on localhost (127.0.0.1) by default.
Now I am no longer getting the error message, but the console window still
failed to load after some time, with Firefox browser error message:

===
The connection was reset

The connection to the server was reset while the page was loading.
===

How can I further troubleshoot on this? I tried to SSH into the console
proxy VM and can't find anything useful on the logs there.

Looking forward to your reply, thank you.

Cheers.



On Mon, Jul 15, 2013 at 12:38 PM, Indra Pramana <indra@sg.or.id> wrote:

> Hi Prasanna,
>
> On Mon, Jul 15, 2013 at 12:20 PM, Prasanna Santhanam <tsp@apache.org>wrote:
>
>> On Sat, Jul 13, 2013 at 02:25:35PM +0800, Indra Pramana wrote:
>> > Hi Prasanna,
>> >
>> > Good day to you, and thank you for your e-mail.
>> >
>> Were you able to get beyond this error?
>>
>
> Yes. I managed to get around the network bridges issue by setting up this
> on my KVM host:
>
> ===
> auto eth0.5
> iface eth0.5 inet static
>         address X.X.X.22
>         netmask 255.255.255.240
>         network X.X.X.16
>         broadcast X.X.X.31
>         gateway X.X.X.17
>         # dns-* options are implemented by the resolvconf package, if
> installed
>         dns-nameservers 8.8.8.8 8.8.4.4
>         dns-search xxx.com
>
> # Public network
> auto cloudbr0
> iface cloudbr0 inet manual
>     bridge_ports eth1.201
>     bridge_fd 5
>     bridge_stp off
>     bridge_maxwait 1
>
> # Guest network
> auto cloudbr1
> iface cloudbr1 inet manual
>     bridge_ports eth1.41
>     bridge_fd 5
>     bridge_stp off
>     bridge_maxwait 1
>
> # Management network
> auto cloudbr2
> iface cloudbr2 inet manual
>     bridge_ports eth0.6
>     bridge_fd 5
>     bridge_stp off
>     bridge_maxwait 1
>
> # The 10G network interface
> auto eth2
> iface eth2 inet static
>         address 10.237.11.22
>         netmask 255.255.255.0
>         network 10.237.11.0
>         broadcast 10.237.11.255
> ===
>
> On the Cloudstack GUI, I created 3 physical networks:
>
> cloudbr0 - public network
> cloudbr1 - guest network
> cloudbr2 - management (and storage) network
>
> However, when I tried to setup the public network using VLAN 201, it
> failed because CloudStack tried to create a new bridge (i.e. breth1-201)
> instead of using the existing bridge I have set up for public network
> (cloudbr mapped to eth1.201). So I set up the public network using a
> different VLAN (VLAN 21) and change the settings on my router's gateway to
> use VLAN 21 instead of VLAN 201.
>
> ===
> root@hv-kvm-02:~# brctl show
> bridge name     bridge id               STP enabled     interfaces
> breth1-21               8000.002590c29d7d       no              eth1.21
>                                                         vnet2
>                                                         vnet5
> cloud0          8000.fe00a9fe0165       no              vnet0
>                                                         vnet3
> cloudbr0                8000.002590c29d7d       no              eth1.201
> cloudbr1                8000.002590c29d7d       no              eth1.41
> cloudbr2                8000.002590c29d7c       no              eth0.6
>                                                         vnet1
>                                                         vnet4
>                                                         vnet6
> virbr0          8000.000000000000       yes
> ===
>
> The above setup works, but I am not too sure whether this is the correct
> way on how to do it. Can you advise?
>
> Looking forward to your reply, thank you.
>
> Cheers.
>
>
>
>
>
>> > Yesterday I troubleshooted further and found out that the mapping of the
>> > bridge is incorrect. For example, eth1 of the SSVM, which is supposed
>> to be
>> > private network interface for communication with management server, is
>> > being mapped to cloudbr0 bridge which is used for public. While eth2 of
>> the
>> > SSVM, which is supposed to be public, is being mapped to cloudbr1 bridge
>> > which is supposed for guest traffic. I checked with brctl show on the
>> KVM
>> > host and traced the MAC address of the SSVM's eth interfaces.
>> >
>> > I have tried different type of configuration on the KVM hosts'
>> > /etc/network/interfaces (I am using Ubuntu), different ways on how to
>> > configure the advanced network zone (physical network interface and KVM
>> > label for each traffic), but I am still not able to get it working.
>> >
>> > Can please advise what is the best way to set-up the network bridges?
>>
>> Use brctl to add bridges. That's how I've done it.
>>
>> >
>> > My /etc/network/interfaces on KVM host:
>> >
>> > http://pastebin.com/nx8xJ1L2
>> >
>> > I used 3 physical NICs:
>> >
>> > eth0 --> for management and secondary storage traffic
>> > eth1 --> for public and guest traffic
>> > eth2 --> for primary storage traffic to the Ceph RBD (separate NIC not
>> > configured by CloudStack)
>> >
>> > On the Cloudstack GUI, while creating the zone, I used 3 physical
>> networks:
>> >
>> > eth0.6 --> Management traffic (KVM label: eth0.6)
>> > cloudbr0 --> Public traffic (KVM label: cloudbr0)
>> > cloudbr1 --> Guest traffic (KVM label: cloudbr1)
>> >
>> Is eth0.6 on a tagged VLAN?
>>
>> > I didn't specify storage traffic since I presume it will use the
>> management
>> > VLAN.
>>
>> That's correct.
>>
>> >
>> > However, I always failed to add the KVM host using the above
>> configuration.
>> > I tried to run cloudstack-setup-agent manually on the KVM host and
>> getting
>> > this error message:
>> >
>> > root@hv-kvm-02:/etc/cloudstack/agent# cloudstack-setup-agent -m
>> x.x.x.18 -z
>> > 3 -p 3 -c 3 -g 00dd5dba-7419-3689-bc51-1671035c0d8f -a --pubNic=cloudbr0
>> > --prvNic=eth0.6 --guestNic=cloudbr1
>> > Starting to configure your system:
>> > Configure Apparmor ...        [OK]
>> > Configure Network ...         [Failed]
>> > eth0.6 is not a bridge
>> > Try to restore your system:
>> > Restore Apparmor ...          [OK]
>> > Restore Network ...           [OK]
>> >
>> > When I tried to create another bridge called cloudbr2 with bridge_ports
>> to
>> > eth0.6, the KVM host's network didn't work.
>> >
>> > Can advise what is the best practice on configuring the network bridges
>> on
>> > the advanced network zone setup?
>> >
>> > Looking forward to your reply, thank you.
>> >
>> > Cheers.
>> >
>> >
>> >
>> > On Sat, Jul 13, 2013 at 12:28 PM, Prasanna Santhanam <tsp@apache.org>
>> wrote:
>> >
>> > > See Inline,
>> > >
>> > > On Fri, Jul 12, 2013 at 05:12:25PM +0800, Indra Pramana wrote:
>> > > > Hi Wido,
>> > > >
>> > > > Noted, can't wait for 4.2 to be released. :)
>> > > >
>> > > > Dear Prasanna, Wido and all,
>> > > >
>> > > > I just realised that while the system VMs are running, they are
>> still not
>> > > > accessible through the public IPs assigned to them. I have been
>> waiting
>> > > for
>> > > > the SSVM to download the default CentOS template and it doesn't
>> appear on
>> > > > the template list.
>> > > >
>> > > > I tried to SSH into the SSVM via the local link address from the KVM
>> > > host,
>> > > > and running the health check /usr/local/cloud/systemvm/ssvm-check.sh
>> > > shows
>> > > > that the VM cannot reach anywhere. It cannot reach the public DNS
>> server
>> > > (I
>> > > > used Google's 8.8.8.8), cannot reach the management server, and
>> cannot
>> > > even
>> > > > reach the public IP gateway.
>> > >
>> > > Ok - we're not there yet. But there's a few troubleshooting steps you
>> > > can try:
>> > >
>> > > 1. Which of the checks fail apart from the DNS ping?
>> > > 2. Are you able to have those tests pass from the KVM host itself?
>> > > 3. Can you paste the output of $route -n on your SSVM and on your
>> > > host?
>> > >
>> > > >
>> > > > Is it due to misconfiguration of the KVM network bridges? How can I
>> see
>> > > the
>> > > > mapping between the NIC interfaces of the SSVM (eth0, eth1, eth2 and
>> > > eth3)
>> > > > and the actual physical NIC interfaces on the KVM hosts (eth0) and
>> the
>> > > > network bridges (cloudbr0, cloudbr1)? Any logs I can verify to
>> ensure
>> > > that
>> > > > the VLAN and network bridging is working?
>> > > >
>> > >
>> > > You can see this from the dumpxml.
>> > >
>> > > $virsh list
>> > > (gives you all the  domains. find the ssvm for example)
>> > >
>> > > $virsh dumpxml my-domain > mydomain.xml
>> > >
>> > > > Appreciate any advice.
>> > > >
>> > > > Thank you.
>> > > >
>> > > >
>> > > >
>> > > > On Fri, Jul 12, 2013 at 4:19 PM, Wido den Hollander <wido@widodh.nl
>> >
>> > > wrote:
>> > > >
>> > > > > On 07/12/2013 10:14 AM, Indra Pramana wrote:
>> > > > >
>> > > > >> Hi Prasanna,
>> > > > >>
>> > > > >> I managed to fix the problem, thanks for your advice to turn the
>> agent
>> > > > >> log level to debug:
>> > > > >>
>> > > > >> https://cwiki.apache.org/**confluence/display/CLOUDSTACK/**
>> > > > >> KVM+agent+debug<
>> > >
>> https://cwiki.apache.org/confluence/display/CLOUDSTACK/KVM+agent+debug>
>> > > > >>
>> > > > >>  From the log, I found out that the agent on the KVM host tried
>> to NFS
>> > > > >> mount directly to 103.25.200.19:/mnt/vol1/sec-**
>> > > > >> storage/template/tmpl/1/3,
>> > > > >> which was not allowed by the NFS server due to its default
>> > > configuration
>> > > > >> to only allow to mount to /mnt/vol1/sec-storage (the root of the
>> NFS
>> > > > >> share).
>> > > > >>
>> > > > >>
>> > > > > Ah, that's odd!
>> > > > >
>> > > > > Btw, in 4.2 you'll be able to deploy SSVMs on RBD as well, so that
>> > > > > limitation will be gone.
>> > > > >
>> > > > > Wido
>> > > > >
>> > > > >  After I changed the NFS server configuration to allow mount to
>> all
>> > > > >> sub-directories, re-export the NFS and voila, the system was
>> able to
>> > > > >> download the template and now both the system VMs (CPVM and
>> SSVM) are
>> > > > >> running!
>> > > > >>
>> > > > >> Many thanks for your help! :)
>> > > > >>
>> > > > >> Cheers.
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >> On Fri, Jul 12, 2013 at 3:31 PM, Indra Pramana <indra@sg.or.id
>> > > > >> <mailto:indra@sg.or.id>> wrote:
>> > > > >>
>> > > > >>     Hi Prasanna,
>> > > > >>
>> > > > >>     Good day to you, and thank you for your e-mail.
>> > > > >>
>> > > > >>     Yes, the cloudstack-agent service is running on both the KVM
>> > > hosts.
>> > > > >>     There is no "cloud" user being created though, when I
>> installed
>> > > the
>> > > > >>     agent. I installed the agent as root.
>> > > > >>
>> > > > >>     root@hv-kvm-01:/home/indra# service cloudstack-agent status
>> > > > >>       * cloud-agent is running
>> > > > >>
>> > > > >>     root@hv-kvm-01:/home/indra# su - cloud
>> > > > >>     Unknown id: cloud
>> > > > >>
>> > > > >>     Please advise how can I resolve this problem, shall I create
>> the
>> > > > >>     Unix "cloud" user manually? Basically I follow this
>> instruction to
>> > > > >>     prepare the KVM host and install the CloudStack agent:
>> > > > >>
>> > > > >>
>> http://cloudstack.apache.org/**docs/en-US/Apache_CloudStack/**
>> > > > >>
>> 4.1.0/html/Installation_Guide/**hypervisor-kvm-install-flow.**html<
>> > >
>> http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.1.0/html/Installation_Guide/hypervisor-kvm-install-flow.html
>> > > >
>> > > > >>
>> > > > >>     with this instruction from Wido on how to prepare libvirt
>> with
>> > > Ceph
>> > > > >>     RBD storage pool support:
>> > > > >>
>> > > > >>     http://blog.widodh.nl/2013/06/**a-quick-note-on-running-**
>> > > > >> cloudstack-with-rbd-on-ubuntu-**12-04/<
>> > >
>> http://blog.widodh.nl/2013/06/a-quick-note-on-running-cloudstack-with-rbd-on-ubuntu-12-04/
>> > > >
>> > > > >>
>> > > > >>     I also have checked /var/log/cloud/agent/agent.log and I
>> don't see
>> > > > >>     any error messages, except this error message which will
>> show up
>> > > > >>     every time I restart the agent:
>> > > > >>
>> > > > >>     2013-07-12 15:22:47,454 ERROR
>> > > [cloud.resource.**ServerResourceBase]
>> > > > >>     (main:null) Nics are not configured!
>> > > > >>     2013-07-12 15:22:47,459 INFO
>> > >  [cloud.resource.**ServerResourceBase]
>> > > > >>     (main:null) Designating private to be nic eth0.5
>> > > > >>
>> > > > >>     More logs can be found here: http://pastebin.com/yeNmCt7S
>> > > > >>
>> > > > >>     I have configured the network bridges on the NIC interface
>> as per
>> > > > >>     this instruction:
>> > > > >>
>> > > > >>
>> http://cloudstack.apache.org/**docs/en-US/Apache_CloudStack/**
>> > > > >> 4.1.0/html/Installation_Guide/**hypervisor-kvm-install-flow.**
>> > > > >> html#hypervisor-host-install-**network<
>> > >
>> http://cloudstack.apache.org/docs/en-US/Apache_CloudStack/4.1.0/html/Installation_Guide/hypervisor-kvm-install-flow.html#hypervisor-host-install-network
>> > > >
>> > > > >>
>> > > > >>     On the zone, I used advanced network configuration with just
>> one
>> > > > >>     physical network for management, public and guest/private. I
>> > > didn't
>> > > > >>     include storage, which by default the traffic will use the
>> > > > >>     management VLAN network.
>> > > > >>
>> > > > >>     Please advise if there's anything else I might have been
>> missing.
>> > > > >>
>> > > > >>     Looking forward to your reply, thank you.
>> > > > >>
>> > > > >>     Cheers.
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>     On Fri, Jul 12, 2013 at 2:56 PM, Prasanna Santhanam <
>> > > tsp@apache.org
>> > > > >>     <mailto:tsp@apache.org>> wrote:
>> > > > >>
>> > > > >>         Indeed, cloudstack will go through the allocation to
>> startup
>> > > the
>> > > > >>         system VMs too. So that process is failing to recognize
>> the
>> > > volume
>> > > > >>         (.qcow2) present on your NFS storage.
>> > > > >>
>> > > > >>         Can you check if your cloudstack agent service is
>> running with
>> > > > >>         the KVM
>> > > > >>         host? And it should've created the user cloud. $id cloud
>> to
>> > > check
>> > > > >> if
>> > > > >>         there is a user.
>> > > > >>
>> > > > >>         Did you see what's happening in the agent logs? These are
>> > > under
>> > > > >>         /var/log/cloud/ on your host when the systemVMs are
>> coming up.
>> > > > >>         If the
>> > > > >>         logs are not showing any useful information you can turn
>> on
>> > > debug
>> > > > >>         level for more verbosity.
>> > > > >>
>> > > > >>         See here: https://cwiki.apache.org/**confluence/x/FgPMAQ
>> <
>> > > https://cwiki.apache.org/confluence/x/FgPMAQ>
>> > > > >>
>> > > > >>         On Fri, Jul 12, 2013 at 02:40:35PM +0800, Indra Pramana
>> wrote:
>> > > > >>          > Hi Prasanna,
>> > > > >>          >
>> > > > >>          > Good day to you, and thank you for your e-mail.
>> > > > >>          >
>> > > > >>          > Yes, when I export the NFS, I set the permission so
>> that
>> > > > >>         normal user will
>> > > > >>          > be able to have read/write access to the files
>> > > > >> (no_root_squash).
>> > > > >>          >
>> > > > >>          > I have tested and I can have read/write access from
>> my KVM
>> > > > >>         hosts using
>> > > > >>          > normal user. BTW, there's no "cloud" user on the
>> hosts, I
>> > > > >>         believe it's not
>> > > > >>          > created during cloudstack-agent installation?
>> > > > >>          >
>> > > > >>          > In any case, do you think the template issue and the
>> > > storage
>> > > > >> pool
>> > > > >>          > allocation issue might be related, or are they two
>> > > different
>> > > > >>         problems
>> > > > >>          > altogether?
>> > > > >>          >
>> > > > >>          > Looking forward to your reply, thank you.
>> > > > >>          >
>> > > > >>          > Cheers.
>> > > > >>          >
>> > > > >>          >
>> > > > >>          >
>> > > > >>          > On Fri, Jul 12, 2013 at 2:26 PM, Prasanna Santhanam
>> > > > >>         <tsp@apache.org <mailto:tsp@apache.org>> wrote:
>> > > > >>          >
>> > > > >>          > > Can you access the file as user root? Or user
>> cloud? The
>> > > > >>         cloudstack
>> > > > >>          > > agent on your KVM host runs as user cloud and the
>> NFS
>> > > > >>         permissions
>> > > > >>          > > might be disallowing the volume (.qcow2) from being
>> > > accessed.
>> > > > >>          > >
>> > > > >>          > > On Fri, Jul 12, 2013 at 02:16:41PM +0800, Indra
>> Pramana
>> > > > >> wrote:
>> > > > >>          > > > Hi Prasanna,
>> > > > >>          > > >
>> > > > >>          > > > Good day to you, and thank you for your e-mail.
>> > > > >>          > > >
>> > > > >>          > > > Yes, the file exists. I can access the file from
>> the
>> > > > >>         management server
>> > > > >>          > > and
>> > > > >>          > > > the two hypervisors hosts if I mount manually.
>> > > > >>          > > >
>> > > > >>          > > > [root@cs-nas-01
>> /mnt/vol1/sec-storage/**template/tmpl/1/3]#
>> > > > >> ls
>> > > > >>          > > > -la
>> > > > >>          > > > total
>> > > > >>          > > > 1418787
>> > > > >>          > > > drwxr-xr-x  2 root  wheel          4 Jul 11 20:21
>> > > > >>          > > > .
>> > > > >>          > > > drwxr-xr-x  3 root  wheel          3 Jul 11 20:17
>> > > > >>          > > > ..
>> > > > >>          > > > -rw-r--r--  1 root  wheel  725811200 Jul 11 20:21
>> > > > >>          > > > 425b9e5a-fbc7-4637-a33a-**fe9d0ed4fa98.qcow2
>> > > > >>          > > > -rw-r--r--  1 root  wheel        295 Jul 11 20:21
>> > > > >>          > > > template.properties
>> > > > >>          > > > [root@cs-nas-01 /mnt/vol1/sec-storage/**
>> > > > >> template/tmpl/1/3]#
>> > > > >>          > > > pwd
>> > > > >>          > > > /mnt/vol1/sec-storage/**template/tmpl/1/3
>> > > > >>          > > >
>> > > > >>          > > >
>> > > > >>          > > > Any advise?
>> > > > >>          > > >
>> > > > >>          > > > Looking forward to your reply, thank you.
>> > > > >>          > > >
>> > > > >>          > > > Cheers.
>> > > > >>          > > >
>> > > > >>          > > >
>> > > > >>          > > >
>> > > > >>          > > > On Fri, Jul 12, 2013 at 2:07 PM, Prasanna
>> Santhanam
>> > > > >>         <tsp@apache.org <mailto:tsp@apache.org>>
>> > > > >>
>> > > > >>          > > wrote:
>> > > > >>          > > >
>> > > > >>          > > > > Can you check whether there is a file at:
>> > > > >>          > > > >
>> > > > >>         nfs://
>> 103.25.200.19/mnt/vol1/**sec-storage/template/tmpl/1/3/
>> > > <http://103.25.200.19/mnt/vol1/sec-storage/template/tmpl/1/3/>
>> > > > >>         <
>> > > http://103.25.200.19/mnt/**vol1/sec-storage/template/**tmpl/1/3/<
>> > > http://103.25.200.19/mnt/vol1/sec-storage/template/tmpl/1/3/>
>> > > > >> >
>> > > > >>
>> > > > >>          > > > >
>> > > > >>          > > > > On Fri, Jul 12, 2013 at 01:59:34PM +0800, Indra
>> > > Pramana
>> > > > >>         wrote:
>> > > > >>          > > > > > Hi Prasanna,
>> > > > >>          > > > > >
>> > > > >>          > > > > > Thanks for your e-mail.
>> > > > >>          > > > > >
>> > > > >>          > > > > > I have tried restarting the management
>> server, and
>> > > > >>         the problem still
>> > > > >>          > > > > > persists. I even tried to re-do the
>> installation
>> > > and
>> > > > >>         configuration
>> > > > >>          > > again
>> > > > >>          > > > > > from scratch last night, but the problem still
>> > > there.
>> > > > >>          > > > > >
>> > > > >>          > > > > > I also noted that on the beginning of the
>> logs, I
>> > > > >>         found some error
>> > > > >>          > > > > messages
>> > > > >>          > > > > > saying that the template cannot be downloaded
>> to
>> > > the
>> > > > >>         pool. See this
>> > > > >>          > > logs:
>> > > > >>          > > > > >
>> > > > >>          > > > > > http://pastebin.com/BY1AVJ08
>> > > > >>          > > > > >
>> > > > >>          > > > > > It says it failed because cannot get volume
>> from
>> > > the
>> > > > >>         pool. Could it
>> > > > >>          > > be
>> > > > >>          > > > > > related, i.e. the absence of the template
>> caused
>> > > the
>> > > > >>         system VMs
>> > > > >>          > > cannot be
>> > > > >>          > > > > > created and started?
>> > > > >>          > > > > >
>> > > > >>          > > > > > I have ensured that I downloaded the system VM
>> > > > >>         template using
>> > > > >>          > > > > > cloud-install-sys-tmplt and verified that the
>> > > > >>         template is already
>> > > > >>          > > there
>> > > > >>          > > > > in
>> > > > >>          > > > > > the secondary storage server.
>> > > > >>          > > > > >
>> > > > >>          > > > > > Any advice is appreciated.
>> > > > >>          > > > > >
>> > > > >>          > > > > > Looking forward to your reply, thank you.
>> > > > >>          > > > > >
>> > > > >>          > > > > > Cheers.
>> > > > >>          > > > > >
>> > > > >>          > > > > >
>> > > > >>          > > > > >
>> > > > >>          > > > > > On Fri, Jul 12, 2013 at 1:21 PM, Prasanna
>> Santhanam
>> > > > >>         <tsp@apache.org <mailto:tsp@apache.org>>
>> > > > >>
>> > > > >>          > > > > wrote:
>> > > > >>          > > > > >
>> > > > >>          > > > > > > It looks like a previous attempt to start
>> the
>> > > > >>         systemVMs has failed
>> > > > >>          > > > > > > putting the nfs storage in the avoid set.
>> Did you
>> > > > >>         try restarting
>> > > > >>          > > your
>> > > > >>          > > > > > > management server?
>> > > > >>          > > > > > >
>> > > > >>          > > > > > > This line leads me to the above mentioned:
>> > > > >>          > > > > > > 2013-07-12 13:10:48,236 DEBUG
>> > > > >>          > > > > > >
>> > > [storage.allocator.**AbstractStoragePoolAllocator]
>> > > > >>          > > (secstorage-1:null)
>> > > > >>          > > > > > > StoragePool is in avoid set, skipping this
>> pool
>> > > > >>          > > > > > >
>> > > > >>          > > > > > >
>> > > > >>          > > > > > > On Fri, Jul 12, 2013 at 01:16:53PM +0800,
>> Indra
>> > > > >>         Pramana wrote:
>> > > > >>          > > > > > > > Dear Wido and all,
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > I have managed to get the hosts, primary
>> and
>> > > > >>         secondary storage
>> > > > >>          > > > > running.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > - 2 KVM hypervisor hosts
>> > > > >>          > > > > > > > - One RBD primary storage
>> > > > >>          > > > > > > > - One NFS primary storage (for system VMs,
>> > > since
>> > > > >>         I understand
>> > > > >>          > > that
>> > > > >>          > > > > system
>> > > > >>          > > > > > > > VMs cannot use RBD)
>> > > > >>          > > > > > > > - One NFS secondary storage
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > However, now I am having problem with the
>> > > system
>> > > > >>         VMs: CPVM and
>> > > > >>          > > SSVM,
>> > > > >>          > > > > > > unable
>> > > > >>          > > > > > > > to start.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > Excerpt from management-server.log file is
>> > > here:
>> > > > >>          > > > > > > > http://pastebin.com/ENkpCALY
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > It seems that the VMs were not able to be
>> > > created
>> > > > >>         because unable
>> > > > >>          > > to
>> > > > >>          > > > > find
>> > > > >>          > > > > > > > suitable StoragePools.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > I understand that system VMs will be
>> using the
>> > > > >>         NFS primary
>> > > > >>          > > storage
>> > > > >>          > > > > > > instead
>> > > > >>          > > > > > > > of RBD, so I have confirmed that I am
>> able to
>> > > > >>         mount the primary
>> > > > >>          > > > > storage
>> > > > >>          > > > > > > via
>> > > > >>          > > > > > > > NFS and have read and write access, from
>> both
>> > > the
>> > > > >>         hypervisor and
>> > > > >>          > > the
>> > > > >>          > > > > > > > management server.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > Any advise how can I resolve the problem
>> to
>> > > make
>> > > > >>         both the system
>> > > > >>          > > VMs
>> > > > >>          > > > > > > > created and started?
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > Looking forward to your reply, thank you.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > Cheers.
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > On Fri, Jul 12, 2013 at 9:43 AM, Indra
>> Pramana
>> > > > >>         <indra@sg.or.id <mailto:indra@sg.or.id>>
>> > > > >>
>> > > > >>          > > > > wrote:
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > > Hi Wido,
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > > Thanks for the advice, I'm now able to
>> add
>> > > the
>> > > > >>         RBD pool as
>> > > > >>          > > primary
>> > > > >>          > > > > > > storage.
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > > Many thanks! :)
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > > Cheers.
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > > On Thursday, July 11, 2013, Wido den
>> > > Hollander
>> > > > >>         wrote:
>> > > > >>          > > > > > > > >
>> > > > >>          > > > > > > > >> Hi,
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> On 07/10/2013 03:42 PM, Chip Childers
>> wrote:
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >>> Cc'ing Wido, our resident Ceph
>> expert. ;-)
>> > > > >>          > > > > > > > >>>
>> > > > >>          > > > > > > > >>>
>> > > > >>          > > > > > > > >> Hehe ;)
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >>  On Wed, Jul 10, 2013 at 05:45:25PM
>> +0800,
>> > > > >>         Indra Pramana
>> > > > >>          > > wrote:
>> > > > >>          > > > > > > > >>>
>> > > > >>          > > > > > > > >>>> Dear all,
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> I am installing CloudStack 4.1.0
>> (upgraded
>> > > > >>         from 4.0.2) and I
>> > > > >>          > > > > also
>> > > > >>          > > > > > > have a
>> > > > >>          > > > > > > > >>>> Ceph cluster running. However, I am
>> having
>> > > > >>         issues in adding
>> > > > >>          > > the
>> > > > >>          > > > > RBD
>> > > > >>          > > > > > > as
>> > > > >>          > > > > > > > >>>> primary storage. Tried to follow the
>> > > > >>         instruction here, but
>> > > > >>          > > > > unable to
>> > > > >>          > > > > > > > >>>> make
>> > > > >>          > > > > > > > >>>> it work:
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>>
>> > > > >>         http://ceph.com/docs/master/****rbd/rbd-cloudstack/<
>> > > http://ceph.com/docs/master/**rbd/rbd-cloudstack/>
>> > > > >> <
>> > > > >>          > > > > > >
>> > > http://ceph.com/docs/master/**rbd/rbd-cloudstack/<
>> > > http://ceph.com/docs/master/rbd/rbd-cloudstack/>
>> > > > >> >
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> I have setup a pool on the Ceph
>> cluster.
>> > > The
>> > > > >>         status of the
>> > > > >>          > > > > cluster
>> > > > >>          > > > > > > is
>> > > > >>          > > > > > > > >>>> healthy. Since I am using Ubuntu
>> 12.04.2
>> > > LTS
>> > > > >>         (Precise) for
>> > > > >>          > > the
>> > > > >>          > > > > > > > >>>> hypervisors,
>> > > > >>          > > > > > > > >>>> I also have compiled libvirt
>> manually to
>> > > > >>         ensure that the
>> > > > >>          > > version
>> > > > >>          > > > > > > 0.9.13
>> > > > >>          > > > > > > > >>>> is
>> > > > >>          > > > > > > > >>>> installed (previously it's 0.9.8).
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >> You can also use the Ubuntu Cloud
>> Archive, I
>> > > > >>         still need to
>> > > > >>          > > get the
>> > > > >>          > > > > > > docs
>> > > > >>          > > > > > > > >> updated for that.
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> I described the process in a blogpost:
>> > > > >>          > > > > > > http://blog.widodh.nl/2013/06/****<
>> > > http://blog.widodh.nl/2013/06/**>
>> > > > >>          > > > > > > > >>
>> > > > >>          > >
>> > > > >>
>> a-quick-note-on-running-****cloudstack-with-rbd-on-ubuntu-**
>> > > > >> **12-04/<
>> > > > >>          > > > > > >
>> > > > >>          > > > >
>> > > > >>          > >
>> > > > >>
>> http://blog.widodh.nl/2013/06/**a-quick-note-on-running-**
>> > > > >> cloudstack-with-rbd-on-ubuntu-**12-04/<
>> > >
>> http://blog.widodh.nl/2013/06/a-quick-note-on-running-cloudstack-with-rbd-on-ubuntu-12-04/
>> > > >
>> > > > >>          > > > > > > >
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >>  indra@hv-kvm-01:~/rbd$ ceph
>> > > > >>          > > > > > > > >>>> ceph> health
>> > > > >>          > > > > > > > >>>> HEALTH_OK
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> indra@hv-kvm-01:~$ ceph osd lspools
>> > > > >>          > > > > > > > >>>> 0 data,1 metadata,2 rbd,3 sc1,
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> root@hv-kvm-01:/home/indra# libvirtd
>> > > > >> --version
>> > > > >>          > > > > > > > >>>> libvirtd (libvirt) 0.9.13
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> I tried to add Primary Storage into
>> the
>> > > > >>         Cloudstack zone
>> > > > >>          > > which I
>> > > > >>          > > > > have
>> > > > >>          > > > > > > > >>>> created:
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> Add Primary Storage:
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>> Zone: my zone name
>> > > > >>          > > > > > > > >>>> Pod: my pod name
>> > > > >>          > > > > > > > >>>> Cluster: my cluster name
>> > > > >>          > > > > > > > >>>> Name: ceph-rbd-pri-storage
>> > > > >>          > > > > > > > >>>> Protocol: RBD
>> > > > >>          > > > > > > > >>>> RADOS Monitor: my first Ceph monitor
>> IP
>> > > > >> address
>> > > > >>          > > > > > > > >>>> RADOS Pool: sc1 (the pool name on
>> Ceph
>> > > > >> cluster)
>> > > > >>          > > > > > > > >>>> RADOS User: client.admin
>> > > > >>          > > > > > > > >>>> RADOS Secret:
>> > > > >>         /etc/ceph/ceph.client.admin.****keyring (keyring
>> > > > >>          > > > > file
>> > > > >>          > > > > > > > >>>> location)
>> > > > >>          > > > > > > > >>>>
>> > > > >>          > > > > > > > >>>
>> > > > >>          > > > > > > > >> This is your problem. That shouldn't
>> be the
>> > > > >>         location of the
>> > > > >>          > > file,
>> > > > >>          > > > > but
>> > > > >>          > > > > > > it
>> > > > >>          > > > > > > > >> should be the secret, which is a base64
>> > > > >>         encoded string.
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> $ ceph auth list
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> That should tell you what the secret
>> is.
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >>  Storage Tags: rbd
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> This is the error message when I tried
>> to
>> > > add
>> > > > >>         the primary
>> > > > >>          > > storage
>> > > > >>          > > > > by
>> > > > >>          > > > > > > > >> clicking OK:
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> DB Exception on:
>> > > > >>          > > com.mysql.jdbc.****JDBC4PreparedStatement@4b2eb56
>> > > > >>          > > > > **:
>> > > > >>          > > > > > > > >> INSERT INTO
>> > > > >>          > > > > > > > >> storage_pool (storage_pool.id
>> > > > >>         <http://storage_pool.id>, storage_pool.name
>> > > > >>         <http://storage_pool.name>,
>> > > > >>
>> > > > >>          > > > > storage_pool.uuid,
>> > > > >>          > > > > > > > >> storage_pool.pool_type,
>> > > storage_pool.created,
>> > > > >>          > > > > > > storage_pool.update_time,
>> > > > >>          > > > > > > > >> storage_pool.data_center_id,
>> > > > >> storage_pool.pod_id,
>> > > > >>          > > > > > > > >> storage_pool.available_bytes,
>> > > > >>         storage_pool.capacity_bytes,
>> > > > >>          > > > > > > > >> storage_pool.status,
>> storage_pool.scope,
>> > > > >>          > > > > > > storage_pool.storage_provider_****
>> > > > >>          > > > > > > > >> id,
>> > > > >>          > > > > > > > >> storage_pool.host_address,
>> > > storage_pool.path,
>> > > > >>          > > storage_pool.port,
>> > > > >>          > > > > > > > >> storage_pool.user_info,
>> > > > >>         storage_pool.cluster_id) VALUES (217,
>> > > > >>          > > > > > > > >> _binary'ceph-rbd-pri-storage',
>> > > > >>          > > > > > > > >>
>> > > > >>         _binary'a226c9a1-da78-3f3a-****b5ac-e18b925c9634', 'RBD',
>> > > > >>          > > > > '2013-07-10
>> > > > >>          > > > > > > > >> 09:08:28', null, 2, 2, 0, 0, 'Up',
>> null,
>> > > null,
>> > > > >>         null,
>> > > > >>          > > _binary'ceph/
>> > > > >>          > > > > > > > >> ceph.client.admin.keyring@10.
>> > > ****237.11.2/sc1<
>> > > > >>          > > > > > > http://ceph.client.admin.**
>> > > keyring@10.237.11.2/sc1<
>> http://ceph.client.admin.keyring@10.237.11.2/sc1>
>> > > > >> >',
>> > > > >>          > > > > > > > >> 6789, null, 2)
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> On the management-server.log file:
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >> 2013-07-10 17:08:28,845 DEBUG
>> > > > >>         [cloud.api.ApiServlet]
>> > > > >>          > > > > > > > >> (catalina-exec-2:null)
>> > > > >>          > > > > > > > >> ===START===  192.168.0.100 -- GET
>> > > > >>          > > > > > > > >>
>> > > > >>
>> > > command=createStoragePool&****zoneid=c116950e-e4ae-4f23-****a7e7-
>> > > > >>          > > > > > > > >>
>> > > > >>
>> 74a75c4ee638&podId=a748b063-****3a83-4175-a0e9-de39118fe5ce&**
>> > > > >>          > > > > > > > >>
>> > > > >>
>> > > clusterid=1f87eb09-324d-4d49-****83c2-88d84d7a15df&name=ceph-****
>> > > > >>          > > > > > > > >>
>> > > > >>
>> rbd-pri-storage&url=rbd%3A%2F%****2Fclient.admin%3A_etc%2Fc
>> > > > >>          > > > > > > > >>
>> > > > >>
>> eph%2Fceph.client.admin.****keyring%4010.237.11.2%2Fsc1&**
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > >
>> > > > >>         tags=rbd&response=json&****sessionkey=****
>> > > > >> rDRfWpqeKfQKbKZtHr398ULV%2F8k%****
>> > > > >>          > > > > > > > >> 3D&_=1373447307839
>> > > > >>          > > > > > > > >> 2013-07-10 17:08:28,862 DEBUG
>> > > > >>          > > [cloud.storage.****StorageManagerImpl]
>> > > > >>          > > > > > > > >> (catalina-exec-2:null) createPool
>> Params @
>> > > > >>         scheme - rbd
>> > > > >>          > > > > storageHost -
>> > > > >>          > > > > > > null
>> > > > >>          > > > > > > > >> hostPath - /ceph/ceph.client
>> > > > >>          > > > > > > > >> .admin.keyring@10.237.11.2/sc1
>> > > > >>         <http://admin.keyring@10.237.**11.2/sc1<
>> > > http://admin.keyring@10.237.11.2/sc1>>
>> > > > >> port - -1
>> > > > >>
>> > > > >>          > > > > > > > >> 2013-07-10 17:08:28,918 DEBUG
>> > > > >>          > > [cloud.storage.****StorageManagerImpl]
>> > > > >>          > > > > > > > >> (catalina-exec-2:null) In createPool
>> Setting
>> > > > >>         poolId - 217
>> > > > >>          > > uuid -
>> > > > >>          > > > > > > > >>
>> a226c9a1-da78-3f3a-b5ac-****e18b925c9634 z
>> > > > >>          > > > > > > > >> oneId - 2 podId - 2 poolName -
>> > > > >>         ceph-rbd-pri-storage
>> > > > >>          > > > > > > > >> 2013-07-10 17:08:28,921 DEBUG
>> > > > >>         [db.Transaction.Transaction]
>> > > > >>          > > > > > > > >> (catalina-exec-2:null) Rolling back the
>> > > > >>         transaction: Time = 3
>> > > > >>          > > > > Name =
>> > > > >>          > > > > > > > >> persist; called by -Transaction.rollbac
>> > > > >>          > > > > > > > >>
>> > > > >>
>> k:890-Transaction.removeUpTo:****833-Transaction.close:657-**
>> > > > >>          > > > > > > > >>
>> > > > >>         TransactionContextBuilder.****interceptException:63-**
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > >
>> > > > >>
>> > > ComponentInstantiationPostProc****essor$InterceptorDispatcher.**
>> > > > >> **interce
>> > > > >>          > > > > > > > >> pt:133-StorageManagerImpl.****
>> > > > >> createPool:1378-**
>> > > > >>          > > > > > > > >>
>> > > > >>
>> StorageManagerImpl.createPool:****147-CreateStoragePoolCmd.**
>> > > > >>          > > > > > > > >>
>> > > > >>         execute:123-ApiDispatcher.****dispatch:162-ApiServer.**
>> > > > >>          > > > > > > > >> queueCommand:505-ApiSe
>> > > > >>          > > > > > > > >>
>> > > > >>         rver.handleRequest:355-****ApiServlet.processRequest:302
>> > > > >>          > > > > > > > >> 2013-07-10 17:08:28,923 ERROR
>> > > > >>         [cloud.api.ApiServer]
>> > > > >>          > > > > > > (catalina-exec-2:null)
>> > > > >>          > > > > > > > >> unhandled exception executing api
>> command:
>> > > > >>         createStoragePool
>> > > > >>          > > > > > > > >>
>> > > > >>         com.cloud.utils.exception.****CloudRuntimeException: DB
>> > > > >>          > > Exception
>> > > > >>          > > > > on:
>> > > > >>          > > > > > > > >>
>> > > > >>         com.mysql.jdbc.****JDBC4PreparedStatement@4b2eb56****:
>> INSERT
>> > > > >>          > > INTO
>> > > > >>          > > > > > > > >> storage_pool (
>> > > > >>          > > > > > > > >> storage_pool.id <
>> http://storage_pool.id>,
>> > > > >>
>> > > > >>         storage_pool
>> > > > >>          > > > > > > > >> .name, storage_pool.uuid,
>> > > > >> storage_pool.pool_type,
>> > > > >>          > > > > > > storage_pool.created,
>> > > > >>          > > > > > > > >> storage_pool.update_time,
>> > > > >>         storage_pool.data_center_id,
>> > > > >>          > > > > > > > >> storage_pool.pod_id,
>> > > > >>          > > > > > > > >> storage_pool.availab
>> > > > >>          > > > > > > > >> le_bytes, storage_pool.capacity_bytes,
>> > > > >>         storage_pool.status,
>> > > > >>          > > > > > > > >> storage_pool.scope,
>> > > > >>         storage_pool.storage_provider_****id,
>> > > > >>          > > > > > > > >> storage_pool.host_address,
>> > > storage_pool.path,
>> > > > >>         storage_
>> > > > >>          > > > > > > > >> pool.port, storage_pool.user_info,
>> > > > >>         storage_pool.cluster_id)
>> > > > >>          > > VALUES
>> > > > >>          > > > > > > (217,
>> > > > >>          > > > > > > > >> _binary'ceph-rbd-pri-storage',
>> > > > >>          > > > > > > > >>
>> > > > >>         _binary'a226c9a1-da78-3f3a-****b5ac-e18b925c9634', 'RBD',
>> > > > >>          > > '2013-07-1
>> > > > >>          > > > > > > > >> 0 09:08:28', null, 2, 2, 0, 0, 'Up',
>> null,
>> > > > >>         null, null,
>> > > > >>          > > > > _binary'ceph/
>> > > > >>          > > > > > > > >> ceph.client.admin.keyring@10.
>> > > ****237.11.2/sc1<
>> > > > >>          > > > > > > http://ceph.client.admin.**
>> > > keyring@10.237.11.2/sc1<
>> http://ceph.client.admin.keyring@10.237.11.2/sc1>
>> > > > >> >',
>> > > > >>          > > > > > > > >> 6789, null, 2)
>> > > > >>          > > > > > > > >>          at
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > >
>> > > > >>          > >
>> > > > >>         com.cloud.utils.db.****GenericDaoBase.persist(****
>> > > > >> GenericDaoBase.java:1342)
>> > > > >>          > > > > > > > >>          at
>> > > > >>          > > > > > > > >>
>> > > > >>         com.cloud.storage.dao.****StoragePoolDaoImpl.persist(**
>> > > > >>          > > > > > > > >> StoragePoolDaoImpl.java:232)
>> > > > >>          > > > > > > > >>          at
>> > > > >>          > > > > > > > >>
>> > > > >>
>> > > com.cloud.utils.component.****ComponentInstantiationPostProc****
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > >
>> > > > >>          > > > >
>> > > > >>          > >
>> > > > >>         essor$InterceptorDispatcher.****intercept(****
>> > > > >> ComponentInstantiationPostProc***
>> > > > >>          > > > > > > > >> *es
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > > > >>
>> > > > >>          > > > > > >
>> > > > >>          > > > > > > --
>> > > > >>          > > > > > > Prasanna.,
>> > > > >>          > > > > > >
>> > > > >>          > > > > > > ------------------------
>> > > > >>          > > > > > > Powered by BigRock.com
>> > > > >>          > > > > > >
>> > > > >>          > > > > > >
>> > > > >>          > > > >
>> > > > >>          > > > > --
>> > > > >>          > > > > Prasanna.,
>> > > > >>          > > > >
>> > > > >>          > > > > ------------------------
>> > > > >>          > > > > Powered by BigRock.com
>> > > > >>          > > > >
>> > > > >>          > > > >
>> > > > >>          > >
>> > > > >>          > > --
>> > > > >>          > > Prasanna.,
>> > > > >>          > >
>> > > > >>          > > ------------------------
>> > > > >>          > > Powered by BigRock.com
>> > > > >>          > >
>> > > > >>          > >
>> > > > >>
>> > > > >>         --
>> > > > >>         Prasanna.,
>> > > > >>
>> > > > >>         ------------------------
>> > > > >>         Powered by BigRock.com
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >>
>> > > > >
>> > >
>> > > --
>> > > Prasanna.,
>> > >
>> > > ------------------------
>> > > Powered by BigRock.com
>> > >
>> > >
>>
>> --
>> Prasanna.,
>>
>> ------------------------
>> Powered by BigRock.com
>>
>>
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message