Am Montag, 6. Mai 2013, 16:41:43 schrieb Wido den Hollander:
> On 05/06/2013 04:15 PM, Guido Winkelmann wrote:
> > Am Montag, 6. Mai 2013, 16:05:31 schrieb Wido den Hollander:
> >> On 05/06/2013 04:00 PM, Guido Winkelmann wrote:
> >>> Hi,
> >>>
> >>> How do I run a 1-node cluster with no replication?
> >>>
> >>> I'm trying to run a small 1-node cluster on my local workstation and
> >>> another on my notebook for experimentation/development purposes, but
> >>> since I only have on OSD, I'm always getting HEALTH_WARN as the cluster
> >>> status from ceph -s. Can I somehow tell ceph to just not bother with
> >>> replication for this cluster?
> >>
> >> Have you set min_size to 1 for all the pools?
> >
> > You mean in the crushmap?
>
> No, it's pool setting.
>
> See: http://ceph.com/docs/master/rados/operations/pools/#set-pool-values
Hm, I set that to 1 now, and nothing changed:
$ ceph osd lspools
0 data,1 metadata,2 rbd,
$ ceph osd pool set data min_size 1
set pool 0 min_size to 1
$ ceph osd pool set rbd min_size 1
set pool 2 min_size to 1
$ ceph osd pool set metadata min_size 1
set pool 1 min_size to 1
$ ceph -s
health HEALTH_WARN 384 pgs degraded; 384 pgs stuck unclean; recovery
1552/3104 degraded (50.000%)
monmap e1: 1 mons at {alpha=x.x.x.x:6789/0}, election epoch 1, quorum 0
alpha
osdmap e16: 1 osds: 1 up, 1 in
pgmap v104: 384 pgs: 384 active+degraded; 6112 MB data, 7144 MB used, 456
GB / 465 GB avail; 1552/3104 degraded (50.000%)
mdsmap e14: 1/1/1 up {0=a=up:active}
Are you sure this is a different setting than what you can see and change in
the crushmap? Because that would be quite confusing...
Guido
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com