Hi Wido,

I'm in the project with Kimi and Nakajima-san.

[root@rx200s7-07m ~]# ceph -v
ceph version 0.61.4 (1669132fcfc27d0c0b5e5bb93ade59d147e23404)

Same version is installed into all the hosts (KVM host and Ceph nodes).

Here is KVM Agent log.
http://pastebin.com/5yG1uBuj
I had set the log level "DEBUG" and failed to create 2 instances
,"RBDVM-shimazaki-1" and "RBDVM-shimazaki-2".

Thanks,
Satoshi Shimazaki




2013/7/23 Kimihiko Kitase <kimihiko.kit...@citrix.co.jp>

> Hi Wido
>
> Thanks for you comment.
>
> If we create vm on the NFS primary storage and mount additional disk on
> the RBD storage, it works fine.
> If we check vm from virt manager, there is no virtual disk. So we believe
> the problem should be vm configuration...
>
> We will check ceph version tomorrow.
>
> Thanks
> Kimi
>
> -----Original Message-----
> From: Wido den Hollander [mailto:w...@widodh.nl]
> Sent: Monday, July 22, 2013 11:43 PM
> To: dev@cloudstack.apache.org
> Subject: Re: Problem in adding Ceph RBD storage to CloudStack
>
> Hi,
>
> On 07/22/2013 02:25 PM, Kimihiko Kitase wrote:
> > Wido, Thank you very much.
> >
> > CloudStack: 4.1.0
> > QEMU: 1.5.50
> > Libvirt: 0.10.2
>
> What version of Ceph on the nodes?
>
> $ ceph -v
>
> >
> > We will set "DEBUG" on the agent tomorrow. But the following is command
> CloudStack issue. We got this command at KVM host.
> >
> > [root@rx200s7-07m ~]# ps -ef|grep 1517
> > root     16099     1 27 19:36 ?        00:00:12 /usr/libexec/qemu-kvm
> -name i-2-1517-VM -S -M pc-i440fx-1.6 -enable-kvm -m 256 -smp
> 1,sockets=1,cores=1,threads=1 -uuid e67f1707-fe92-3426-978d-0441d5000d6a
> -no-user-config -nodefaults -chardev
> socket,id=charmonitor,path=/var/lib/libvirt/qemu/i-2-1517-VM.monitor,server,nowait
> -mon chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
> -boot dc -drive
> file=rbd:libvirt-pool/cd3688ab-e37b-4866-9ea7-4051b670a323:id=libvirt:key=AQC7OuZReMndFxAAY/qUwLbvfod6EMvgVWU21g==:auth_supported=cephx\;none:mon_host=192.168.10.20\:6789,if=none,id=drive-virtio-disk0,format=raw,cache=none
> -device
> virtio-blk-pci,bus=pci.0,addr=0x4,drive=drive-virtio-disk0,id=virtio-disk0
> -drive
> if=none,media=cdrom,id=drive-ide0-1-0,readonly=on,format=raw,cache=none
> -device ide-drive,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -netdev
> tap,fd=27,id=hostnet0,vhost=on,vhostfd=29 -device
> virtio-net-pci,netdev=hostnet0,id=net0,mac=02:00:0a:b9:00:16
>  ,bus=pci.
> 0,addr=0x3 -chardev pty,id=charserial0 -device
> isa-serial,chardev=charserial0,id=serial0 -usb -device usb-tablet,id=input0
> -vnc 0.0.0.0:3 -vga cirrus -device
> virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5
> >
>
> The argument to Qemu seems just fine, so I think the problem is not in
> CloudStack.
>
> Wido
>
> > Thanks
> > Kimi
> >
> > -----Original Message-----
> > From: Wido den Hollander [mailto:w...@widodh.nl]
> > Sent: Monday, July 22, 2013 7:47 PM
> > To: dev@cloudstack.apache.org
> > Subject: Re: Problem in adding Ceph RBD storage to CloudStack
> >
> > Hi,
> >
> > On 07/22/2013 12:43 PM, Kimihiko Kitase wrote:
> >> It seems secondary storage vm could copy template to primary storage
> successfully, but created VM doesn't point this vol..
> >> If we create vm manually and add this vol as boot vol, it works fine..
> >>
> >
> > Which version of CloudStack are you using?
> >
> > What is the Qemu version running on your hypervisor and what libvirt
> version?
> >
> > If you set the logging level on the Agent to "DEBUG", does it show
> deploying the VM with the correct XML parameters?
> >
> > I haven't seen the things you are reporting.
> >
> > Wido
> >
> >> So it seems cloudstack cannot configure VM correctly in ceph rbd
> environment.
> >>
> >> Any idea?
> >>
> >> Thanks
> >> Kimi
> >>
> >> -----Original Message-----
> >> From: Kimihiko Kitase [mailto:kimihiko.kit...@citrix.co.jp]
> >> Sent: Monday, July 22, 2013 7:11 PM
> >> To: dev@cloudstack.apache.org
> >> Subject: RE: Problem in adding Ceph RBD storage to CloudStack
> >>
> >> Hello
> >>
> >> I am in the project with Nakajima san.
> >>
> >> We succeeded to add RBD storage to primary storage.
> >> But when we try to boot centos as user instance, it fail during system
> logger process.
> >> It works fine when we boot centos using NFS storage.
> >> It works fine when we boot centos using NFS storage and add additional
> disk from RBD storage.
> >>
> >> Do you have any idea to resolve this issue?
> >>
> >> Thanks
> >> Kimi
> >>
> >> -----Original Message-----
> >> From: Takuma Nakajima [mailto:penguin.trance.2...@gmail.com]
> >> Sent: Saturday, July 20, 2013 12:23 PM
> >> To: dev@cloudstack.apache.org
> >> Subject: Re: Problem in adding Ceph RBD storage to CloudStack
> >>
> >> I'm sorry but I forgot to tell you that the environment does not have
> the internet connection.
> >> It is not allowed to make a direct connection to the internet because
> of the security policy.
> >>
> >> Wido,
> >>> No, it works for me like a charm :)
> >>>
> >>> Could you set the Agent logging to DEBUG as well and show the output
> >>> of
> >> that log? Maybe paste the log on pastebin.
> >>>
> >>> I'm interested in the XMLs the Agent is feeding to libvirt when
> >>> adding
> >> the RBD pool.
> >>
> >> I thought the new libvirt overwrites the old one, but actually both
> libvirt (with RBD and without RBD) were installed to the system. qemu was
> installed from the package and so it might have the dependency to the
> libvirt installed from the package. After deleting the both libvirt
> installed from source and package, then installed it from rpm package with
> RBD support, RBD storage was registered to the CloudStack successfully.
> >>
> >> David,
> >>> Why not 6.4?
> >>
> >> Because of no internet connection, packages in the local mirror
> repository may be old.
> >> I checked /etc/redhat-release and it showed the version is 6.3.
> >>
> >> In current state, although the RBD storage was installed, system VMs
> >> won't start with "Unable to get vms
> >> org.libvirt.LibvirtException: Domain not found: no domain with
> >> matching uuid 'xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx'" error like
> >> http://mail-archives.apache.org/mod_mbox/cloudstack-users/201303.mbox
> >> /
> >> %3CD2EE6B3265AD864EB3EA4F5C670D256F3546C1@EXMBX01L-CRP-03.webmdhealth.
> >> net%3E The uuid in the error message was not in the database of the
> >> management server nor in ceph storage node.
> >>
> >> I tried removing host from the CloudStack and cleaning up the computing
> node, but it cannot be added again to the CloudStack.
> >> agent log says it attempted to connect to localhost:8250 though the
> management server address is set to 10.40.1.190 in global settings.
> >> management server log is here: http://pastebin.com/muGz73c0
> >> (10.40.1.24 is the address of the computing node)
> >>
> >> Now the computing node is under rebuilding.
> >>
> >> Takuma Nakajima
> >>
> >> 2013/7/19 David Nalley <da...@gnsa.us>:
> >>> On Thu, Jul 18, 2013 at 12:09 PM, Takuma Nakajima
> >>> <penguin.trance.2...@gmail.com> wrote:
> >>>> Hi,
> >>>>
> >>>> I'm building a CloudStack 4.1 with Ceph RBD storage using RHEL 6.3
> >> recently
> >>>> but it fails when adding RBD storage to primary storage.
> >>>> Does anybody know about the problem?
> >>>
> >>>
> >>> Why not 6.4?
>

Reply via email to