Hello Sebastien / Community

I tried the commands mentioned in below email.


[root@rdo ~]#
[root@rdo ~]# cinder create 1
+---------------------+--------------------------------------+
|       Property      |                Value                 |
+---------------------+--------------------------------------+
|     attachments     |                  []                  |
|  availability_zone  |                 nova                 |
|       bootable      |                false                 |
|      created_at     |      2013-11-27T07:40:54.161478      |
| display_description |                 None                 |
|     display_name    |                 None                 |
|          id         | ae8cd686-5f1d-4c05-8c42-cb7622122a3e |
|       metadata      |                  {}                  |
|         size        |                  1                   |
|     snapshot_id     |                 None                 |
|     source_volid    |                 None                 |
|        status       |               creating               |
|     volume_type     |                 None                 |
+---------------------+--------------------------------------+
[root@rdo ~]#
[root@rdo ~]# cinder list
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
|                  ID                  | Status | Display Name | Size | Volume 
Type | Bootable | Attached to |
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
| ae8cd686-5f1d-4c05-8c42-cb7622122a3e | error  |     None     |  1   |     
None    |  false   |             |
+--------------------------------------+--------+--------------+------+-------------+----------+-------------+
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]# rbd -p ceph-volumes ls
rbd: pool ceph-volumes doesn't contain rbd images
[root@rdo ~]#
[root@rdo ~]#
[root@rdo ~]# rados lspools
data
metadata
rbd
ceph-images
ceph-volumes
[root@rdo ~]# rbd -p rbd ls
[root@rdo ~]# rbd -p data ls
foo
foo1
[root@rdo ~]#




I checked in cinder.log and got the below errors.


2013-11-27 09:44:14.830 3273 INFO cinder.volume.manager [-] Updating volume 
status
2013-11-27 09:44:14.830 3273 WARNING cinder.volume.manager [-] Unable to update 
stats, driver is uninitialized
2013-11-27 09:44:42.407 12007 INFO cinder.volume.manager [-] Updating volume 
status
2013-11-27 09:44:42.408 12007 WARNING cinder.volume.manager [-] Unable to 
update stats, driver is uninitialized
2013-11-27 09:44:51.799 4943 INFO cinder.volume.manager [-] Updating volume 
status
2013-11-27 09:44:51.799 4943 WARNING cinder.volume.manager [-] Unable to update 
stats, driver is uninitialized
2013-11-27 09:45:14.834 3273 INFO cinder.volume.manager [-] Updating volume 
status
2013-11-27 09:45:14.834 3273 WARNING cinder.volume.manager [-] Unable to update 
stats, driver is uninitialized
[root@rdo cinder]#




Output from my cinder.conf file



# Options defined in cinder.volume.utils
#

# The default block size used when copying/clearing volumes
# (string value)
#volume_dd_blocksize=1M


# Total option count: 382
volume_driver=cinder.volume.drivers.rbd.RBDDriver
rbd_pool=ceph-volumes
glance_api_version=2
rbd_user=volumes
rbd_secret_uuid=801a42ec-aec1-3ea8-d869-823c2de56b83

rootwrap_config=/etc/cinder/rootwrap.conf
sql_connection = mysql://cinder:root@localhost/cinder
api_paste_config = /etc/cinder/api-paste.ini

iscsi_helper=tgtadm
volume_name_template = volume-%s
volume_group = cinder-volumes
verbose = True
auth_strategy = keystone
#osapi_volume_listen_port=5900
[keystone_authtoken]
admin_tenant_name = admin
admin_user = admin
admin_password = root@123
auth_host = 192.168.1.49
auth_port = 35357
auth_protocol = http
signing_dirname = /tmp/keystone-signing-cinder
[root@rdo cinder]#



Can you suggest what is wrong with this configuration , i followed everything 
mentioned on http://ceph.com/docs/master/rbd/rbd-openstack/   and there is not 
much information available on internet WRT this error.

Please help.



Many Thanks
Karan Singh


----- Original Message -----
From: "Sebastien Han" <[email protected]>
To: "Karan Singh" <[email protected]>
Cc: [email protected]
Sent: Wednesday, 27 November, 2013 1:08:35 AM
Subject: Re: [ceph-users] how to Testing cinder and glance with CEPH

Hi,

Well after restarting the services run:

$ cinder create 1

Then you can check both status in Cinder and Ceph:

For Cinder run:
$ cinder list

For Ceph run:
$ rbd -p <cinder-pool> ls

If the image is there, you’re good.

Cheers.

–––– 
Sébastien Han 
Cloud Engineer 

"Always give 100%. Unless you're giving blood.” 

Phone: +33 (0)1 49 70 99 72 
Mail: [email protected] 
Address : 10, rue de la Victoire - 75009 Paris 
Web : www.enovance.com - Twitter : @enovance 

On 27 Nov 2013, at 00:04, Karan Singh <[email protected]> wrote:

> Hello Cephers
> 
> I was following http://ceph.com/docs/master/rbd/rbd-openstack/  for ceph and 
> openstack Integration , using this document ih ave done all the changes 
> required for this integration.
> 
> I am not sure how should i test my configuration , how should i make sure 
> integration is successful. Can you suggest some test that i can perform to 
> check my ceph and openstack integration .
> 
> FYI , in the document http://ceph.com/docs/master/rbd/rbd-openstack/  , 
> Nothing is mentioned after " Restart Openstack Services "  heading , but 
> there should be steps to test this inttegration , please suggest me here , i 
> am new to openstack great if you can give me some commanes used for testing.
> 
> 
> 
> Karan Singh
> CSC - IT Center for Science Ltd.
> P.O. Box 405, FI-02101 Espoo, FINLAND
> http://www.csc.fi/ | +358 (0) 503 812758
> 
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to