Hello Mr. Kees,
Thanks for your response!
My setup is
Openstack Node 1 -> controller + network + compute1 (Liberty Version)
Openstack node 2 --> Compute2
Ceph version Hammer
I am using dell storage with following status
DELL SAN storage is attached to both hosts as
[root@OSKVM1 ~]# iscsiadm -m node
10.35.0.3:3260,1
iqn.2001-05.com.equallogic:0-1cb196-07a83c107-47700000018575af-vol1
10.35.0.8:3260,1
iqn.2001-05.com.equallogic:0-1cb196-07a83c107-47700000018575af-vol1
10.35.0.*:3260,-1
iqn.2001-05.com.equallogic:0-1cb196-20d83c107-7290000002157606-vol2
10.35.0.8:3260,1
iqn.2001-05.com.equallogic:0-1cb196-20d83c107-7290000002157606-vol2
10.35.0.*:3260,-1
iqn.2001-05.com.equallogic:0-1cb196-f0783c107-70a000000245761a-vol3
10.35.0.8:3260,1
iqn.2001-05.com.equallogic:0-1cb196-f0783c107-70a000000245761a-vol3
10.35.0.*:3260,-1
iqn.2001-05.com.equallogic:0-1cb196-fda83c107-927000000275761a-vol4
10.35.0.8:3260,1
iqn.2001-05.com.equallogic:0-1cb196-fda83c107-927000000275761a-vol4
Since in my setup same LUNs are MAPPED to both hosts
i choose 2 LUNS on Openstack Node 1 and 2 on Openstack Node 2
*Node1 has *
/dev/sdc1 2.0T 3.1G 2.0T 1% /var/lib/ceph/osd/ceph-0
/dev/sdd1 2.0T 3.8G 2.0T 1% /var/lib/ceph/osd/ceph-1
*Node 2 has *
/dev/sdd1 2.0T 3.4G 2.0T 1% /var/lib/ceph/osd/ceph-2
/dev/sde1 2.0T 3.5G 2.0T 1% /var/lib/ceph/osd/ceph-3
[root@OSKVM1 ~]# ceph status
cluster 9f923089-a6c0-4169-ace8-ad8cc4cca116
health HEALTH_WARN
mon.OSKVM1 low disk space
monmap e1: 1 mons at {OSKVM1=10.24.0.4:6789/0}
election epoch 1, quorum 0 OSKVM1
osdmap e40: 4 osds: 4 up, 4 in
pgmap v1154: 576 pgs, 5 pools, 6849 MB data, 860 objects
13857 MB used, 8154 GB / 8168 GB avail
576 active+clean
*Can you please help me to know if it is correct configuration as per my
setup?*
After this setup, i am trying to configure Cinder and Glance to use RBD for
a backend.
Glance image is already stored in RBD.
Following this link http://docs.ceph.com/docs/master/rbd/rbd-openstack/
I have managed to install glance image in rbd. But i am finding some issue
in cinder configuration. Can you please help me on this?
As per link, i need to configure these parameters under [ceph] but i do not
have different section for [ceph]. infact i could find all these parameters
under [DEFAULT]. Is it ok to configure them under [DEFAULT].
CONFIGURING CINDER
<http://docs.ceph.com/docs/master/rbd/rbd-openstack/#configuring-cinder>
OpenStack requires a driver to interact with Ceph block devices. You must
also specify the pool name for the block device. On your OpenStack node,
edit/etc/cinder/cinder.conf by adding:
[DEFAULT]
...
enabled_backends = ceph
...
[ceph]
volume_driver = cinder.volume.drivers.rbd.RBDDriver
rbd_pool = volumes
rbd_ceph_conf = /etc/ceph/ceph.conf
rbd_flatten_volume_from_snapshot = false
rbd_max_clone_depth = 5
rbd_store_chunk_size = 4
rados_connect_timeout = -1
glance_api_version = 2
I find following error in cinder service status
systemctl status openstack-cinder-volume.service
Jul 07 09:37:01 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:01.058
136259 ERROR cinder.service [-] Manager for service cinder-volume
OSKVM1@ceph is reporting problems, not sending heartbeat. Service will
appear "down".
Jul 07 09:37:02 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:02.040
136259 WARNING cinder.volume.manager
[req-561ddd3c-9560-4374-a958-7a2c103af7ee - - - - -] Update driver status
failed: (config name ceph) is uninitialized.
Jul 07 09:37:11 OSKVM1 cinder-volume[136247]: 2016-07-07 09:37:11.059
136259 ERROR cinder.service [-] Manager for service cinder-volume
OSKVM1@ceph is reporting problems, not sending heartbeat. Service will
appear "down".
[root@OSKVM2 ~]# rbd -p images ls
a8b45c8a-a5c8-49d8-a529-1e4088bdbf3f
[root@OSKVM2 ~]# rados df
pool name KB objects clones degraded
unfound rd rd KB wr wr KB
backups 0 0 0 0
0 0 0 0 0
images 7013377 860 0 0
0 9486 7758 2580 7013377
rbd 0 0 0 0
0 0 0 0 0
vms 0 0 0 0
0 0 0 0 0
volumes 0 0 0 0
0 0 0 0 0
total used 14190236 860
total avail 8550637828
total space 8564828064
[root@OSKVM2 ~]# ceph auth list
installed auth entries:
mds.OSKVM1
key: AQCK6XtXNBFdDBAAXmX73gBqK3lyakSxxP+XjA==
caps: [mds] allow
caps: [mon] allow profile mds
caps: [osd] allow rwx
osd.0
key: AQAB4HtX7q27KBAAEqcuJXwXAJyD6a1Qu/MXqA==
caps: [mon] allow profile osd
caps: [osd] allow *
osd.1
key: AQC/4ntXFJGdFBAAADYH03iQTF4jWI1LnBZeJg==
caps: [mon] allow profile osd
caps: [osd] allow *
osd.2
key: AQCa43tXr12fDhAAzbq6FO2+8m9qg1B12/99Og==
caps: [mon] allow profile osd
caps: [osd] allow *
osd.3
key: AQA/5HtXDNfcLxAAJWawgxc1nd8CB+4uH/8fdQ==
caps: [mon] allow profile osd
caps: [osd] allow *
client.admin
key: AQBNknJXE/I2FRAA+caW02eje7GZ/uv1O6aUgA==
caps: [mds] allow
caps: [mon] allow *
caps: [osd] allow *
client.bootstrap-mds
key: AQBOknJXjLloExAAGjMRfjp5okI1honz9Nx4wg==
caps: [mon] allow profile bootstrap-mds
client.bootstrap-osd
key: AQBNknJXDUMFKBAAZ8/TfDkS0N7Q6CbaOG3DyQ==
caps: [mon] allow profile bootstrap-osd
client.bootstrap-rgw
key: AQBOknJXQAUiABAA6IB4p4RyUmrsxXk+pv4u7g==
caps: [mon] allow profile bootstrap-rgw
client.cinder
key: AQCIAHxX9ga8LxAAU+S3Vybdu+Cm2bP3lplGnA==
caps: [mon] allow r
caps: [osd] allow class-read object_prefix rbd_children, allow rwx
pool=volumes, allow rwx pool=vms, allow rx pool=images
client.cinder-backup
key: AQCXAHxXAVSNKhAAV1d/ZRMsrriDOt+7pYgJIg==
caps: [mon] allow r
caps: [osd] allow class-read object_prefix rbd_children, allow rwx
pool=backups
client.glance
key: AQCVAHxXupPdLBAA7hh1TJZnvSmFSDWbQiaiEQ==
caps: [mon] allow r
caps: [osd] allow class-read object_prefix rbd_children, allow rwx
pool=images
Regards
Gaurav Goyal
On Thu, Jul 7, 2016 at 2:54 AM, Kees Meijs <[email protected]> wrote:
> Hi Gaurav,
>
> Unfortunately I'm not completely sure about your setup, but I guess it
> makes sense to configure Cinder and Glance to use RBD for a backend. It
> seems to me, you're trying to store VM images directly on an OSD
> filesystem.
>
> Please refer to http://docs.ceph.com/docs/master/rbd/rbd-openstack/ for
> details.
>
> Regards,
> Kees
>
> On 06-07-16 23:03, Gaurav Goyal wrote:
> >
> > I am installing ceph hammer and integrating it with openstack Liberty
> > for the first time.
> >
> > My local disk has only 500 GB but i need to create 600 GB VM. SO i
> > have created a soft link to ceph filesystem as
> >
> > lrwxrwxrwx 1 root root 34 Jul 6 13:02 instances ->
> > /var/lib/ceph/osd/ceph-0/instances [root@OSKVM1 nova]# pwd
> > /var/lib/nova [root@OSKVM1 nova]#
> >
> > now when i am trying to create an instance it is giving the following
> > error as checked from nova-compute.log
> > I need your help to fix this issue.
> >
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com