Hello everybody,
I want to build a new architecture with Ceph for storage backend.
For the moment I’ve got only one server with this specs :
1 RAID-1 SSD : OS + OSD journals
12x 4To : OSD daemons.
I never reached the « clean state » on my cluster and I’m always in HEALTH_WARN
mode like this :
health HEALTH_WARN 25 pgs degraded; 24 pgs incomplete; 24 pgs stuck
inactive; 64 pgs stuck unclean; 25 pgs undersized
I tried to add 3 —> 12 OSD but it’s always the same problem.
What is the right configuration to have a valid cluster please ?
# cat ceph.conf
[global]
fsid = 588595a0-3570-44bb-af77-3c0eaa28fbdb
mon_initial_members = drt-marco
mon_host = 172.16.21.4
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true
public network = 172.16.21.0/24
[osd]
osd journal size = 10000
osd crush chooseleaf type = 0
osd pool default size = 1
NB : I use ceph-deploy for debian wheezy to deploy the services.
Thank you so much for your help !
k.
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com