Hi: -----------------------------ceph osd tree----------------------------------- # id weight type name up/down reweight -1 1.82 root default -2 1.82 host storage1 0 0.91 osd.0 up 1 1 0.91 osd.1 up 1
Best wishes, Mika 2014-10-29 17:05 GMT+08:00 Irek Fasikhov <[email protected]>: > ceph osd tree please :) > > 2014-10-29 12:03 GMT+03:00 Vickie CH <[email protected]>: > >> Dear all, >> Thanks for the reply. >> Pool replicated size is 2. Because the replicated size parameter already >> write into ceph.conf before deploy. >> Because not familiar crush map. I will according Mark's information to >> do a test that change the crush map to see the result. >> >> -----------ceph.conf------------------ >> [global] >> fsid = c404ded6-4086-4f0b-b479- >> 89bc018af954 >> mon_initial_members = storage0 >> mon_host = 192.168.1.10 >> auth_cluster_required = cephx >> auth_service_required = cephx >> auth_client_required = cephx >> filestore_xattr_use_omap = true >> >> *osd_pool_default_size = 2osd_pool_default_min_size = 1* >> osd_pool_default_pg_num = 128 >> osd_journal_size = 2048 >> osd_pool_default_pgp_num = 128 >> osd_mkfs_type = xfs >> ------------------------------------------- >> >> ----------------------ceph osd dump result ----------------------------- >> pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash >> rjenkins pg_num 64 pgp_num 64 last_change 14 flags hashpspool >> crash_replay_interval 45 stripe_width 0 >> pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 0 >> object_hash rjenkins pg_num 64 pgp_num 64 last_change 15 flags hashpspool >> stripe_width 0 >> pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash >> rjenkins pg_num 64 pgp_num 64 last_change 16 flags hashpspool stripe_width 0 >> max_osd 2 >> >> ------------------------------------------------------------------------------ >> >> Best wishes, >> Mika >> >> Best wishes, >> Mika >> >> 2014-10-29 16:56 GMT+08:00 Mark Kirkwood <[email protected]>: >> >>> That is not my experience: >>> >>> $ ceph -v >>> ceph version 0.86-579-g06a73c3 (06a73c39169f2f332dec760f56d3ec >>> 20455b1646) >>> >>> $ cat /etc/ceph/ceph.conf >>> [global] >>> ... >>> osd pool default size = 2 >>> >>> $ ceph osd dump|grep size >>> pool 2 'hot' replicated size 2 min_size 1 crush_ruleset 0 object_hash >>> rjenkins pg_num 128 pgp_num 128 last_change 47 flags >>> hashpspool,incomplete_clones tier_of 1 cache_mode writeback target_bytes >>> 2000000000 hit_set bloom{false_positive_probability: 0.05, target_size: >>> 0, seed: 0} 3600s x1 stripe_width 0 >>> pool 10 '.rgw.root' replicated size 2 min_size 1 crush_ruleset 0 >>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 102 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 11 '.rgw.control' replicated size 2 min_size 1 crush_ruleset 0 >>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 104 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 12 '.rgw' replicated size 2 min_size 1 crush_ruleset 0 object_hash >>> rjenkins pg_num 8 pgp_num 8 last_change 106 owner 18446744073709551615 >>> flags hashpspool stripe_width 0 >>> pool 13 '.rgw.gc' replicated size 2 min_size 1 crush_ruleset 0 >>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 107 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 14 '.users.uid' replicated size 2 min_size 1 crush_ruleset 0 >>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 108 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 15 '.rgw.buckets.index' replicated size 2 min_size 1 crush_ruleset >>> 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 110 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 16 '.rgw.buckets' replicated size 2 min_size 1 crush_ruleset 0 >>> object_hash rjenkins pg_num 8 pgp_num 8 last_change 112 owner >>> 18446744073709551615 flags hashpspool stripe_width 0 >>> pool 17 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash >>> rjenkins pg_num 1024 pgp_num 1024 last_change 186 flags hashpspool >>> stripe_width 0 >>> >>> >>> >>> >>> >>> >>> On 29/10/14 21:46, Irek Fasikhov wrote: >>> >>>> Hi. >>>> This parameter does not apply to pools by default. >>>> ceph osd dump | grep pool. see size=? >>>> >>>> >>>> 2014-10-29 11:40 GMT+03:00 Vickie CH <[email protected] >>>> <mailto:[email protected]>>: >>>> >>>> Der Irek: >>>> >>>> Thanks for your reply. >>>> Even already set "osd_pool_default_size = 2" the cluster still need >>>> 3 different hosts right? >>>> Is this default number can be changed by user and write into >>>> ceph.conf before deploy? >>>> >>>> >>>> Best wishes, >>>> Mika >>>> >>>> 2014-10-29 16:29 GMT+08:00 Irek Fasikhov <[email protected] >>>> <mailto:[email protected]>>: >>>> >>>> Hi. >>>> >>>> Because the disc requires three different hosts, the default >>>> number of replications 3. >>>> >>>> 2014-10-29 10:56 GMT+03:00 Vickie CH <[email protected] >>>> <mailto:[email protected]>>: >>>> >>>> >>>> Hi all, >>>> Try to use two OSDs to create a cluster. After the >>>> deply finished, I found the health status is "88 >>>> active+degraded" "104 active+remapped". Before use 2 osds to >>>> create cluster the result is ok. I'm confuse why this >>>> situation happened. Do I need to set crush map to fix this >>>> problem? >>>> >>>> >>>> ----------ceph.conf--------------------------------- >>>> [global] >>>> fsid = c404ded6-4086-4f0b-b479-89bc018af954 >>>> mon_initial_members = storage0 >>>> mon_host = 192.168.1.10 >>>> auth_cluster_required = cephx >>>> auth_service_required = cephx >>>> auth_client_required = cephx >>>> filestore_xattr_use_omap = true >>>> osd_pool_default_size = 2 >>>> osd_pool_default_min_size = 1 >>>> osd_pool_default_pg_num = 128 >>>> osd_journal_size = 2048 >>>> osd_pool_default_pgp_num = 128 >>>> osd_mkfs_type = xfs >>>> --------------------------------------------------------- >>>> >>>> -----------ceph -s----------------------------------- >>>> cluster c404ded6-4086-4f0b-b479-89bc018af954 >>>> health HEALTH_WARN 88 pgs degraded; 192 pgs stuck >>>> unclean >>>> monmap e1: 1 mons at {storage0=192.168.10.10:6789/0 >>>> <http://192.168.10.10:6789/0>}, election epoch 2, quorum 0 >>>> storage0 >>>> osdmap e20: 2 osds: 2 up, 2 in >>>> pgmap v45: 192 pgs, 3 pools, 0 bytes data, 0 objects >>>> 79752 kB used, 1858 GB / 1858 GB avail >>>> 88 active+degraded >>>> 104 active+remapped >>>> -------------------------------------------------------- >>>> >>>> >>>> Best wishes, >>>> Mika >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> [email protected] <mailto:[email protected] >>>> > >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>>> >>>> >>>> -- >>>> С уважением, Фасихов Ирек Нургаязович >>>> Моб.: +79229045757 >>>> >>>> >>>> >>>> >>>> >>>> -- >>>> С уважением, Фасихов Ирек Нургаязович >>>> Моб.: +79229045757 >>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> [email protected] >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >>> >> > > > -- > С уважением, Фасихов Ирек Нургаязович > Моб.: +79229045757 >
_______________________________________________ ceph-users mailing list [email protected] http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
