Hi, One more step debugging this issue (hypervisor/nova-compute node is XEN 4.4.2):
I think the problem is that libvirt is not getting the correct user or credentials tu access pool, on instance qemu log i see: xen be: qdisk-51760: error: Could not open 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or directory xen be: qdisk-51760: initialise() failed xen be: qdisk-51760: error: Could not open 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or directory xen be: qdisk-51760: initialise() failed xen be: qdisk-51760: error: Could not open 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or directory But using the user cinder on pool volumes : rbd ls -p volumes --id cinder test *volume-4d26bb31-91e8-4646-8010-82127b775c8e* volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a volume-7da08f12-fb0f-4269-931a-d528c1507fee Using: qemu-img info -f rbd rbd:volumes/test Does not work, but using directly the user cinder and the ceph.conf file works fine: *qemu-img info -f rbd rbd:volumes/test:id=cinder:conf=/etc/ceph/ceph.conf* I think nova.conf is set correctly (section libvirt): images_rbd_pool = volumes images_rbd_ceph_conf = /etc/ceph/ceph.conf hw_disk_discard=unmap rbd_user = cinder rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX And looking at libvirt: # virsh secret-list setlocale: No such file or directory UUID Usage -------------------------------------------------------------------------------- 67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX ceph client.cinder secret virsh secret-get-value 67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX setlocale: No such file or directory *AQAonAdWS3iMJxxxxxxj9iErv001a0k+vyFdUg==* cat /etc/ceph/ceph.client.cinder.keyring [client.cinder] key = *AQAonAdWS3iMJxxxxxxj9iErv001a0k+vyFdUg==* Any idea will be welcomed. regards, I 2015-11-04 10:51 GMT+01:00 Iban Cabrillo <cabri...@ifca.unican.es>: > Dear Cephers, > > I still can attach volume to my cloud machines, ceph version is 0.94.5 > (9764da52395923e0b32908d83a9f7304401fee43) and Openstack Juno > > Nova+cinder are able to create volumes on Ceph > cephvolume:~ # rados ls --pool volumes > rbd_header.1f7784a9e1c2e > rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a > rbd_directory > rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee > rbd_header.23d5e33b4c15c > rbd_id.volume-4d26bb31-91e8-4646-8010-82127b775c8e > rbd_header.20407190ce77f > > cloud:~ # cinder list > > +--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+ > | ID | > Status | Display Name | Size | Volume Type | Bootable | > Attached to | > > +--------------------------------------+--------+--------------+------+-------------+----------+--------------------------------|---------------------------------------------------------+ > | 4d26bb31-91e8-4646-8010-82127b775c8e | in-use | None | 2 > | rbd | false | > 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb | > > +--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+ > > > nova:~ # nova volume-attach 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb > 4d26bb31-91e8-4646-8010-82127b775c8e auto > +----------+------------------------------------------------------------+ > | Property | Value > | > +----------+------------------------------------------------------------+ > | device | /dev/xvdd > | > | id | 4d26bb31-91e8-4646-8010-82127b775c8e | > | serverId | 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb | > | volumeId | 4d26bb31-91e8-4646-8010-82127b775c8e | > +----------+--------------------------------------+ > > From nova-compute (Ubuntu 14.04 LTS \n \l) node I see the > attaching/detaching: > cloud01:~ # dpkg -l | grep ceph > ii ceph-common 0.94.5-1trusty > amd64 common utilities to mount and interact with a ceph storage > cluster > ii libcephfs1 0.94.5-1trusty > amd64 Ceph distributed file system client library > ii python-cephfs 0.94.5-1trusty > amd64 Python libraries for the Ceph libcephfs library > ii librbd1 0.94.5-1trusty > amd64 RADOS block device client library > ii python-rbd 0.94.5-1trusty > amd64 Python libraries for the Ceph librbd library > > *at cinder.conf* > > *rbd_user = cinder* > *rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx* > > *[rbd-cephvolume]* > *volume_backend_name=rbd* > *volume_driver = cinder.volume.drivers.rbd.RBDDriver* > *rbd_pool = volumes* > *rbd_ceph_conf = /etc/ceph/ceph.conf* > *rbd_flatten_volume_from_snapshot = false* > *rbd_max_clone_depth = 5* > *rbd_store_chunk_size = 4* > *rados_connect_timeout = -1* > *glance_api_version = 2* > > *in nova.conf* > *rbd_user=cinder* > > *# The libvirt UUID of the secret for the rbd_uservolumes* > *# (string value)* > *rbd_secret_uuid=67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx* > > > > *images_rbd_pool=volumes# Path to the ceph configuration file to use > (string value)images_rbd_ceph_conf=/etc/ceph/ceph.conf* > *ls -la /etc/libvirt/secrets* > *total 16* > *drwx------ 2 root root 4096 Nov 4 10:28 .* > *drwxr-xr-x 7 root root 4096 Oct 22 13:15 ..* > *-rw------- 1 root root 40 Nov 4 10:28 > 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxx.base64* > *-rw------- 1 root root 170 Nov 4 10:25 > 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxx.xml* > > > > 2015-11-04 10:39:42.573 11653 INFO nova.compute.manager > [req-8b2a9793-4b39-4cb0-b291-e492c350387e b7aadbb4a85745feb498b74e437129cc > ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance: > 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Detach volume > 4d26bb31-91e8-4646-8010-82127b775c8e from mountpoint /dev/xvdd > 2015-11-04 10:40:43.266 11653 INFO nova.compute.manager > [req-35218de0-3f26-496b-aad9-5c839143da17 b7aadbb4a85745feb498b74e437129cc > ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance: > 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Attaching volume > 4d26bb31-91e8-4646-8010-82127b775c8e to /dev/xvdd > > but one on cloud machine (SL6) the volume y never showed (xvdd). > [root@cloud5 ~]# cat /proc/partitions > major minor #blocks name > > 202 0 20971520 xvda > 202 16 209715200 xvdb > 202 32 10485760 xvdc > > Thanks in advance, I > > 2015-11-03 11:18 GMT+01:00 Iban Cabrillo <cabri...@ifca.unican.es>: > >> Hi all, >> During last week I been trying to deploy the pre-existing ceph >> cluster with out openstack intance. >> The ceph-cinder integration was easy (or at least I think so!!) >> There is only one volume to attach block storage to out cloud >> machines. >> >> The client.cinder has permission on this volume (following the guides) >> ............... >> client.cinder >> key: AQAonXXXXXXXRAAPIAj9iErv001a0k+vyFdUg== >> caps: [mon] allow r >> *caps: [osd] allow class-read object_prefix rbd_children, allow rwx >> pool=volumes* >> >> ceph.conf file seems to be OK: >> >> *[global]* >> *fsid = 6f5a65a7-316c-4825-afcb-428608941dd1* >> *mon_initial_members = cephadm, cephmon02, cephmon03* >> *mon_host = 10.10.3.1,10.10.3.2,10.10.3.3* >> *auth_cluster_required = cephx* >> *auth_service_required = cephx* >> *auth_client_required = cephx* >> *filestore_xattr_use_omap = true* >> *osd_pool_default_size = 2* >> *public_network = 10.10.0.0/16 <http://10.10.0.0/16>* >> *cluster_network = 192.168.254.0/27 <http://192.168.254.0/27>* >> >> *[osd]* >> *osd_journal_size = 20000* >> >> *[client.cinder]* >> *keyring = /etc/ceph/ceph.client.cinder.keyring* >> >> *[client]* >> *rbd cache = true* >> *rbd cache writethrough until flush = true* >> *admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok* >> >> >> The trouble seems that blocks are created using the client.admin instead >> of client.cinder >> >> From cinder machine: >> >> cinder:~ # rados ls --pool volumes >> rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a >> rbd_directory >> rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee >> rbd_header.23d5e33b4c15c >> rbd_header.20407190ce77f >> >> But if I try to look for using cinder client: >> >> >> cinder:~ #rados ls --pool volumes --secret client.cinder >> "empty answer" >> >> cinder:~ # ls -la /etc/ceph >> total 24 >> drwxr-xr-x 2 root root 4096 nov 3 10:17 . >> drwxr-xr-x 108 root root 4096 oct 29 09:52 .. >> -rw------- 1 root root 63 nov 3 10:17 ceph.client.admin.keyring >> -rw-r--r-- 1 cinder cinder 67 oct 28 13:44 ceph.client.cinder.keyring >> -rw-r--r-- 1 root root 454 oct 1 13:56 ceph.conf >> -rw-r--r-- 1 root root 73 sep 27 09:36 ceph.mon.keyring >> >> >> from a client (I have supposed that this machine only need the cinder >> key...) >> >> cloud28:~ # ls -la /etc/ceph/ >> total 28 >> drwx------ 2 root root 4096 nov 3 11:01 . >> drwxr-xr-x 116 root root 12288 oct 30 14:37 .. >> -rw-r--r-- 1 nova nova 67 oct 28 11:43 ceph.client.cinder.keyring >> -rw-r--r-- 1 root root 588 nov 3 10:59 ceph.conf >> -rw-r--r-- 1 root root 92 oct 26 16:59 rbdmap >> >> cloud28:~ # rbd -p volumes ls >> 2015-11-03 11:01:58.782795 7fc6c714b840 -1 monclient(hunting): ERROR: >> missing keyring, cannot use cephx for authentication >> 2015-11-03 11:01:58.782800 7fc6c714b840 0 librados: client.admin >> initialization error (2) No such file or directory >> rbd: couldn't connect to the cluster! >> >> Any help will be welcome. >> >> > > > -- > > ############################################################################ > Iban Cabrillo Bartolome > Instituto de Fisica de Cantabria (IFCA) > Santander, Spain > Tel: +34942200969 > PGP PUBLIC KEY: > http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC > > ############################################################################ > Bertrand Russell: > *"El problema con el mundo es que los estúpidos están seguros de todo y > los inteligentes están llenos de dudas*" > -- ############################################################################ Iban Cabrillo Bartolome Instituto de Fisica de Cantabria (IFCA) Santander, Spain Tel: +34942200969 PGP PUBLIC KEY: http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC ############################################################################ Bertrand Russell: *"El problema con el mundo es que los estúpidos están seguros de todo y los inteligentes están llenos de dudas*"
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com