Hi, If cinder-volume fails to connect and putting the admin keyring works it means that cinder is not configured properly. Please also try to add the following:
[client.cinder]
keyring = <path-to-keyring>
Same for Glance.
Btw: ceph.conf doesn’t need to be own by Cinder, just let mod +r and keep root
as owner.
––––
Sébastien Han
Cloud Engineer
"Always give 100%. Unless you're giving blood.”
Phone: +33 (0)1 49 70 99 72
Mail: [email protected]
Address : 10, rue de la Victoire - 75009 Paris
Web : www.enovance.com - Twitter : @enovance
On 17 Feb 2014, at 14:48, Ashish Chandra <[email protected]> wrote:
> Hi Sebastian, Jean;
>
> This is my ceph.conf looks like. It was auto generated using ceph-deploy.
>
> [global]
> fsid = afa13fcd-f662-4778-8389-85047645d034
> mon_initial_members = ceph-node1
> mon_host = 10.0.1.11
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> filestore_xattr_use_omap = true
>
> If I provide admin.keyring file to openstack node (in /etc/ceph) it works
> fine and issue is gone .
>
> Thanks
>
> Ashish
>
>
> On Mon, Feb 17, 2014 at 2:03 PM, Sebastien Han <[email protected]>
> wrote:
> Hi,
>
> Can I see your ceph.conf?
> I suspect that [client.cinder] and [client.glance] sections are missing.
>
> Cheers.
> ––––
> Sébastien Han
> Cloud Engineer
>
> "Always give 100%. Unless you're giving blood.”
>
> Phone: +33 (0)1 49 70 99 72
> Mail: [email protected]
> Address : 10, rue de la Victoire - 75009 Paris
> Web : www.enovance.com - Twitter : @enovance
>
> On 16 Feb 2014, at 06:55, Ashish Chandra <[email protected]> wrote:
>
> > Hi Jean,
> >
> > Here is the output for ceph auth list for client.cinder
> >
> > client.cinder
> > key: AQCKaP9ScNgiMBAAwWjFnyL69rBfMzQRSHOfoQ==
> > caps: [mon] allow r
> > caps: [osd] allow class-read object_prefix rbd_children, allow rwx
> > pool=volumes, allow rx pool=images
> >
> >
> > Here is the output of ceph -s:
> >
> > ashish@ceph-client:~$ ceph -s
> > cluster afa13fcd-f662-4778-8389-85047645d034
> > health HEALTH_OK
> > monmap e1: 1 mons at {ceph-node1=10.0.1.11:6789/0}, election epoch 1,
> > quorum 0 ceph-node1
> > osdmap e37: 3 osds: 3 up, 3 in
> > pgmap v84: 576 pgs, 6 pools, 0 bytes data, 0 objects
> > 106 MB used, 9076 MB / 9182 MB avail
> > 576 active+clean
> >
> > I created all the keyrings and copied as suggested by the guide.
> >
> >
> >
> >
> >
> >
> > On Sun, Feb 16, 2014 at 3:08 AM, Jean-Charles LOPEZ <[email protected]>
> > wrote:
> > Hi,
> >
> > what do you get when you run a 'ceph auth list' command for the user name
> > (client.cinder) you created for cinder? Are the caps and the key for this
> > user correct? No typo in the hostname in the cinder.conf file (host=) ? Did
> > you copy the keyring to the cinder running cinder (can’t really say from
> > your output and there is no ceph-s command to check the monitor names)?
> >
> > It could just be a typo in the ceph auth get-or-create command that’s
> > causing it.
> >
> > Rgds
> > JC
> >
> >
> >
> > On Feb 15, 2014, at 10:35, Ashish Chandra <[email protected]>
> > wrote:
> >
> >> Hi Cephers,
> >>
> >> I am trying to configure ceph rbd as backend for cinder and glance by
> >> following the steps mentioned in:
> >>
> >> http://ceph.com/docs/master/rbd/rbd-openstack/
> >>
> >> Before I start all openstack services are running normally and ceph
> >> cluster health shows "HEALTH_OK"
> >>
> >> But once I am done with all steps and restart openstack services,
> >> cinder-volume fails to start and throws an error.
> >>
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd Traceback (most
> >> recent call last):
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd File
> >> "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 262, in
> >> check_for_setup_error
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd with
> >> RADOSClient(self):
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd File
> >> "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 234, in __init__
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd self.cluster,
> >> self.ioctx = driver._connect_to_rados(pool)
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd File
> >> "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 282, in
> >> _connect_to_rados
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd
> >> client.connect()
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd File
> >> "/usr/lib/python2.7/dist-packages/rados.py", line 185, in connect
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd raise
> >> make_ex(ret, "error calling connect")
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd Error: error
> >> calling connect: error code 95
> >> 2014-02-16 00:01:42.582 TRACE cinder.volume.drivers.rbd
> >> 2014-02-16 00:01:42.591 ERROR cinder.volume.manager
> >> [req-8134a4d7-53f8-4ada-b4b5-4d96d7cad4bc None None] Error encountered
> >> during initialization of driver: RBDDriver
> >> 2014-02-16 00:01:42.592 ERROR cinder.volume.manager
> >> [req-8134a4d7-53f8-4ada-b4b5-4d96d7cad4bc None None] Bad or unexpected
> >> response from the storage volume backend API: error connecting to ceph
> >> cluster
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager Traceback (most recent
> >> call last):
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager File
> >> "/opt/stack/cinder/cinder/volume/manager.py", line 190, in init_host
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager
> >> self.driver.check_for_setup_error()
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager File
> >> "/opt/stack/cinder/cinder/volume/drivers/rbd.py", line 267, in
> >> check_for_setup_error
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager raise
> >> exception.VolumeBackendAPIException(data=msg)
> >> 2014-02-16 00:01:42.592 TRACE cinder.volume.manager
> >> VolumeBackendAPIException: Bad or unexpected response from the storage
> >> volume backend API: error connecting to ceph cluster
> >>
> >>
> >> Here is the content of my /etc/ceph in openstack node:
> >>
> >> ashish@ubuntu:/etc/ceph$ ls -lrt
> >> total 16
> >> -rw-r--r-- 1 cinder cinder 229 Feb 15 23:45 ceph.conf
> >> -rw-r--r-- 1 glance glance 65 Feb 15 23:46 ceph.client.glance.keyring
> >> -rw-r--r-- 1 cinder cinder 65 Feb 15 23:47 ceph.client.cinder.keyring
> >> -rw-r--r-- 1 cinder cinder 72 Feb 15 23:47
> >> ceph.client.cinder-backup.keyring
> >>
> >> I am really stuck and tried a lot. What Could possibly I be doing wrong.
> >>
> >>
> >> HELP.
> >>
> >>
> >> Thanks and Regards
> >> Ashish Chandra
> >>
> >> _______________________________________________
> >> ceph-users mailing list
> >> [email protected]
> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> >
> > _______________________________________________
> > ceph-users mailing list
> > [email protected]
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
