Hi Vasilly,
  Thanks, but I still see the same error:

cinder.conf (of course I just restart the cinder-volume service)

# default volume type to use (string value)

[rbd-cephvolume]
rbd_user = cinder
rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx
volume_backend_name=rbd
volume_driver = cinder.volume.drivers.rbd.RBDDriver
rbd_pool = volumes
rbd_ceph_conf = /etc/ceph/ceph.conf
rbd_flatten_volume_from_snapshot = false
rbd_max_clone_depth = 5
rbd_store_chunk_size = 4
rados_connect_timeout = -1
glance_api_version = 2


  xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed
xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed
xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed
xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed
xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed
xen be: qdisk-51760: error: Could not open
'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
directory
xen be: qdisk-51760: initialise() failed

Regards, I

2015-11-06 13:00 GMT+01:00 Vasiliy Angapov <anga...@gmail.com>:

> At cinder.conf you should place this options:
>
> rbd_user = cinder
> rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx
>
> to [rbd-cephvolume] section instead of DEFAULT.
>
> 2015-11-06 19:45 GMT+08:00 Iban Cabrillo <cabri...@ifca.unican.es>:
> > Hi,
> >   One more step debugging this issue (hypervisor/nova-compute node is XEN
> > 4.4.2):
> >
> >   I think the problem is that libvirt is not getting the correct user or
> > credentials tu access pool, on instance qemu log i see:
> >
> > xen be: qdisk-51760: error: Could not open
> > 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
> > directory
> > xen be: qdisk-51760: initialise() failed
> > xen be: qdisk-51760: error: Could not open
> > 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
> > directory
> > xen be: qdisk-51760: initialise() failed
> > xen be: qdisk-51760: error: Could not open
> > 'volumes/volume-4d26bb31-91e8-4646-8010-82127b775c8e': No such file or
> > directory
> >
> > But using the user cinder on pool volumes :
> >
> > rbd ls -p volumes --id cinder
> > test
> > volume-4d26bb31-91e8-4646-8010-82127b775c8e
> > volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a
> > volume-7da08f12-fb0f-4269-931a-d528c1507fee
> >
> > Using:
> > qemu-img info -f rbd rbd:volumes/test
> > Does not work, but using directly the user cinder and the ceph.conf file
> > works fine:
> >
> > qemu-img info -f rbd rbd:volumes/test:id=cinder:conf=/etc/ceph/ceph.conf
> >
> > I think nova.conf is set correctly (section libvirt):
> > images_rbd_pool = volumes
> > images_rbd_ceph_conf = /etc/ceph/ceph.conf
> > hw_disk_discard=unmap
> > rbd_user = cinder
> > rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX
> >
> > And looking at libvirt:
> >
> > # virsh secret-list
> > setlocale: No such file or directory
> >  UUID                                  Usage
> >
> --------------------------------------------------------------------------------
> >  67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX  ceph client.cinder secret
> >
> >
> > virsh secret-get-value 67a6d4a1-e53a-42c7-9bc9-XXXXXXXXXXXX
> > setlocale: No such file or directory
> > AQAonAdWS3iMJxxxxxxj9iErv001a0k+vyFdUg==
> > cat /etc/ceph/ceph.client.cinder.keyring
> > [client.cinder]
> > key = AQAonAdWS3iMJxxxxxxj9iErv001a0k+vyFdUg==
> >
> >
> > Any idea will be welcomed.
> > regards, I
> >
> > 2015-11-04 10:51 GMT+01:00 Iban Cabrillo <cabri...@ifca.unican.es>:
> >>
> >> Dear Cephers,
> >>
> >>    I still can attach volume to my cloud machines, ceph version is
> 0.94.5
> >> (9764da52395923e0b32908d83a9f7304401fee43) and Openstack Juno
> >>
> >>    Nova+cinder are able to create volumes on Ceph
> >> cephvolume:~ # rados ls --pool volumes
> >> rbd_header.1f7784a9e1c2e
> >> rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a
> >> rbd_directory
> >> rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee
> >> rbd_header.23d5e33b4c15c
> >> rbd_id.volume-4d26bb31-91e8-4646-8010-82127b775c8e
> >> rbd_header.20407190ce77f
> >>
> >> cloud:~ # cinder list
> >>
> >>
> +--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+
> >> |                  ID                                               |
> >> Status  | Display Name | Size | Volume Type | Bootable |
> >> Attached to              |
> >>
> >>
> +--------------------------------------+--------+--------------+------+-------------+----------+--------------------------------|---------------------------------------------------------+
> >> | 4d26bb31-91e8-4646-8010-82127b775c8e | in-use |     None     |
>  2
> >> |             rbd     |  false       |
> 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb
> >> |
> >>
> >>
> +--------------------------------------+--------+--------------+------+-------------+----------+------------------------------------------------------------------------------------------+
> >>
> >>
> >>    nova:~ # nova volume-attach 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb
> >> 4d26bb31-91e8-4646-8010-82127b775c8e auto
> >>
> +----------+------------------------------------------------------------+
> >> | Property |                              Value
> >> |
> >>
> +----------+------------------------------------------------------------+
> >> | device      | /dev/xvdd
> >> |
> >> | id             | 4d26bb31-91e8-4646-8010-82127b775c8e |
> >> | serverId   | 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb  |
> >> | volumeId | 4d26bb31-91e8-4646-8010-82127b775c8e |
> >> +----------+--------------------------------------+
> >>
> >> From nova-compute (Ubuntu 14.04 LTS \n \l) node I see the
> >> attaching/detaching:
> >> cloud01:~ # dpkg -l | grep ceph
> >> ii  ceph-common                         0.94.5-1trusty
> >> amd64        common utilities to mount and interact with a ceph storage
> >> cluster
> >> ii  libcephfs1                               0.94.5-1trusty
> >> amd64        Ceph distributed file system client library
> >> ii  python-cephfs                         0.94.5-1trusty
> >> amd64        Python libraries for the Ceph libcephfs library
> >> ii  librbd1                                    0.94.5-1trusty
> >> amd64        RADOS block device client library
> >> ii  python-rbd                              0.94.5-1trusty
> >> amd64        Python libraries for the Ceph librbd library
> >>
> >> at cinder.conf
> >>
> >>  rbd_user = cinder
> >> rbd_secret_uuid = 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx
> >>
> >> [rbd-cephvolume]
> >> volume_backend_name=rbd
> >> volume_driver = cinder.volume.drivers.rbd.RBDDriver
> >> rbd_pool = volumes
> >> rbd_ceph_conf = /etc/ceph/ceph.conf
> >> rbd_flatten_volume_from_snapshot = false
> >> rbd_max_clone_depth = 5
> >> rbd_store_chunk_size = 4
> >> rados_connect_timeout = -1
> >> glance_api_version = 2
> >>
> >> in nova.conf
> >> rbd_user=cinder
> >>
> >> # The libvirt UUID of the secret for the rbd_uservolumes
> >> # (string value)
> >> rbd_secret_uuid=67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxxx
> >>
> >> images_rbd_pool=volumes
> >>
> >> # Path to the ceph configuration file to use (string value)
> >> images_rbd_ceph_conf=/etc/ceph/ceph.conf
> >>
> >> ls -la /etc/libvirt/secrets
> >> total 16
> >> drwx------ 2 root root 4096 Nov  4 10:28 .
> >> drwxr-xr-x 7 root root 4096 Oct 22 13:15 ..
> >> -rw------- 1 root root   40 Nov  4 10:28
> >> 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxx.base64
> >> -rw------- 1 root root  170 Nov  4 10:25
> >> 67a6d4a1-e53a-42c7-9bc9-xxxxxxxxxx.xml
> >>
> >>
> >>
> >> 2015-11-04 10:39:42.573 11653 INFO nova.compute.manager
> >> [req-8b2a9793-4b39-4cb0-b291-e492c350387e
> b7aadbb4a85745feb498b74e437129cc
> >> ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance:
> >> 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Detach volume
> >> 4d26bb31-91e8-4646-8010-82127b775c8e from mountpoint /dev/xvdd
> >> 2015-11-04 10:40:43.266 11653 INFO nova.compute.manager
> >> [req-35218de0-3f26-496b-aad9-5c839143da17
> b7aadbb4a85745feb498b74e437129cc
> >> ce2dd2951bd24c1ea3b43c3b3716f604 - - -] [instance:
> >> 59aa021e-bb4c-4154-9b18-9d09f5fd3aeb] Attaching volume
> >> 4d26bb31-91e8-4646-8010-82127b775c8e to /dev/xvdd
> >>
> >> but one on cloud machine (SL6) the volume y never showed (xvdd).
> >> [root@cloud5 ~]# cat /proc/partitions
> >> major minor  #blocks  name
> >>
> >>  202        0   20971520 xvda
> >>  202       16 209715200 xvdb
> >>  202       32   10485760 xvdc
> >>
> >> Thanks in advance, I
> >>
> >> 2015-11-03 11:18 GMT+01:00 Iban Cabrillo <cabri...@ifca.unican.es>:
> >>>
> >>> Hi all,
> >>>     During last week I been trying to deploy the pre-existing ceph
> >>> cluster with out openstack intance.
> >>>     The ceph-cinder integration was easy (or at least I think so!!)
> >>>     There is only one volume to attach block storage to out cloud
> >>> machines.
> >>>
> >>>     The client.cinder has permission on this volume (following the
> >>> guides)
> >>>     ...............
> >>>     client.cinder
> >>> key: AQAonXXXXXXXRAAPIAj9iErv001a0k+vyFdUg==
> >>> caps: [mon] allow r
> >>> caps: [osd] allow class-read object_prefix rbd_children, allow rwx
> >>> pool=volumes
> >>>
> >>>    ceph.conf file seems to be OK:
> >>>
> >>> [global]
> >>> fsid = 6f5a65a7-316c-4825-afcb-428608941dd1
> >>> mon_initial_members = cephadm, cephmon02, cephmon03
> >>> mon_host = 10.10.3.1,10.10.3.2,10.10.3.3
> >>> auth_cluster_required = cephx
> >>> auth_service_required = cephx
> >>> auth_client_required = cephx
> >>> filestore_xattr_use_omap = true
> >>> osd_pool_default_size = 2
> >>> public_network = 10.10.0.0/16
> >>> cluster_network = 192.168.254.0/27
> >>>
> >>> [osd]
> >>> osd_journal_size = 20000
> >>>
> >>> [client.cinder]
> >>> keyring = /etc/ceph/ceph.client.cinder.keyring
> >>>
> >>> [client]
> >>> rbd cache = true
> >>> rbd cache writethrough until flush = true
> >>> admin socket = /var/run/ceph/$cluster-$type.$id.$pid.$cctid.asok
> >>>
> >>>
> >>> The trouble seems that blocks are created using the client.admin
> instead
> >>> of client.cinder
> >>>
> >>> From cinder machine:
> >>>
> >>> cinder:~ # rados ls --pool volumes
> >>> rbd_id.volume-5e2ab5c2-4710-4c28-9755-b5bc4ff6a52a
> >>> rbd_directory
> >>> rbd_id.volume-7da08f12-fb0f-4269-931a-d528c1507fee
> >>> rbd_header.23d5e33b4c15c
> >>> rbd_header.20407190ce77f
> >>>
> >>> But if I try to look for using cinder client:
> >>>
> >>>
> >>>   cinder:~ #rados ls --pool volumes --secret client.cinder
> >>>   "empty answer"
> >>>
> >>> cinder:~ # ls -la /etc/ceph
> >>> total 24
> >>> drwxr-xr-x   2 root   root   4096 nov  3 10:17 .
> >>> drwxr-xr-x 108 root   root   4096 oct 29 09:52 ..
> >>> -rw-------   1 root   root     63 nov  3 10:17
> ceph.client.admin.keyring
> >>> -rw-r--r--   1 cinder cinder   67 oct 28 13:44
> ceph.client.cinder.keyring
> >>> -rw-r--r--   1 root   root    454 oct  1 13:56 ceph.conf
> >>> -rw-r--r--   1 root   root     73 sep 27 09:36 ceph.mon.keyring
> >>>
> >>>
> >>> from a client (I have supposed that this machine only need the cinder
> >>> key...)
> >>>
> >>> cloud28:~ # ls -la /etc/ceph/
> >>> total 28
> >>> drwx------   2 root root  4096 nov  3 11:01 .
> >>> drwxr-xr-x 116 root root 12288 oct 30 14:37 ..
> >>> -rw-r--r--   1 nova nova    67 oct 28 11:43 ceph.client.cinder.keyring
> >>> -rw-r--r--   1 root root   588 nov  3 10:59 ceph.conf
> >>> -rw-r--r--   1 root root    92 oct 26 16:59 rbdmap
> >>>
> >>> cloud28:~ # rbd -p volumes ls
> >>> 2015-11-03 11:01:58.782795 7fc6c714b840 -1 monclient(hunting): ERROR:
> >>> missing keyring, cannot use cephx for authentication
> >>> 2015-11-03 11:01:58.782800 7fc6c714b840  0 librados: client.admin
> >>> initialization error (2) No such file or directory
> >>> rbd: couldn't connect to the cluster!
> >>>
> >>> Any help will be welcome.
> >>>
> >>
> >>
> >>
> >> --
> >>
> >>
> ############################################################################
> >> Iban Cabrillo Bartolome
> >> Instituto de Fisica de Cantabria (IFCA)
> >> Santander, Spain
> >> Tel: +34942200969
> >> PGP PUBLIC KEY:
> >> http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
> >>
> >>
> ############################################################################
> >> Bertrand Russell:
> >> "El problema con el mundo es que los estúpidos están seguros de todo y
> los
> >> inteligentes están llenos de dudas"
> >
> >
> >
> >
> > --
> >
> ############################################################################
> > Iban Cabrillo Bartolome
> > Instituto de Fisica de Cantabria (IFCA)
> > Santander, Spain
> > Tel: +34942200969
> > PGP PUBLIC KEY:
> > http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
> >
> ############################################################################
> > Bertrand Russell:
> > "El problema con el mundo es que los estúpidos están seguros de todo y
> los
> > inteligentes están llenos de dudas"
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>



-- 
############################################################################
Iban Cabrillo Bartolome
Instituto de Fisica de Cantabria (IFCA)
Santander, Spain
Tel: +34942200969
PGP PUBLIC KEY:
http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
############################################################################
Bertrand Russell:
*"El problema con el mundo es que los estúpidos están seguros de todo y los
inteligentes están llenos de dudas*"
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to