Dear all,
Thanks for the reply.
Pool replicated size is 2. Because the replicated size parameter already
write into ceph.conf before deploy.
Because not familiar crush map.  I will according Mark's information to do
a test that change the crush map to see the result.

-----------ceph.conf------------------
[global]
fsid = c404ded6-4086-4f0b-b479-
89bc018af954
mon_initial_members = storage0
mon_host = 192.168.1.10
auth_cluster_required = cephx
auth_service_required = cephx
auth_client_required = cephx
filestore_xattr_use_omap = true

*osd_pool_default_size = 2osd_pool_default_min_size = 1*
osd_pool_default_pg_num = 128
osd_journal_size = 2048
osd_pool_default_pgp_num = 128
osd_mkfs_type = xfs
-------------------------------------------

----------------------ceph osd dump result -----------------------------
pool 0 'data' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 14 flags hashpspool
crash_replay_interval 45 stripe_width 0
pool 1 'metadata' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 15 flags hashpspool stripe_width 0
pool 2 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
rjenkins pg_num 64 pgp_num 64 last_change 16 flags hashpspool stripe_width 0
max_osd 2
------------------------------------------------------------------------------

Best wishes,
Mika

Best wishes,
Mika

2014-10-29 16:56 GMT+08:00 Mark Kirkwood <[email protected]>:

> That is not my experience:
>
> $ ceph -v
> ceph version 0.86-579-g06a73c3 (06a73c39169f2f332dec760f56d3ec20455b1646)
>
> $ cat /etc/ceph/ceph.conf
> [global]
> ...
> osd pool default size = 2
>
> $ ceph osd dump|grep size
> pool 2 'hot' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 128 pgp_num 128 last_change 47 flags
> hashpspool,incomplete_clones tier_of 1 cache_mode writeback target_bytes
> 2000000000 hit_set bloom{false_positive_probability: 0.05, target_size:
> 0, seed: 0} 3600s x1 stripe_width 0
> pool 10 '.rgw.root' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 102 owner
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 11 '.rgw.control' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 104 owner
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 12 '.rgw' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 106 owner 18446744073709551615
> flags hashpspool stripe_width 0
> pool 13 '.rgw.gc' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 8 pgp_num 8 last_change 107 owner 18446744073709551615
> flags hashpspool stripe_width 0
> pool 14 '.users.uid' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 108 owner
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 15 '.rgw.buckets.index' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 110 owner
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 16 '.rgw.buckets' replicated size 2 min_size 1 crush_ruleset 0
> object_hash rjenkins pg_num 8 pgp_num 8 last_change 112 owner
> 18446744073709551615 flags hashpspool stripe_width 0
> pool 17 'rbd' replicated size 2 min_size 1 crush_ruleset 0 object_hash
> rjenkins pg_num 1024 pgp_num 1024 last_change 186 flags hashpspool
> stripe_width 0
>
>
>
>
>
>
> On 29/10/14 21:46, Irek Fasikhov wrote:
>
>> Hi.
>> This parameter does not apply to pools by default.
>> ceph osd dump | grep pool. see size=?
>>
>>
>> 2014-10-29 11:40 GMT+03:00 Vickie CH <[email protected]
>> <mailto:[email protected]>>:
>>
>>     Der Irek:
>>
>>     Thanks for your reply.
>>     Even already set "osd_pool_default_size = 2" the cluster still need
>>     3 different hosts right?
>>     Is this default number can be changed by user and write into
>>     ceph.conf before deploy?
>>
>>
>>     Best wishes,
>>     Mika
>>
>>     2014-10-29 16:29 GMT+08:00 Irek Fasikhov <[email protected]
>>     <mailto:[email protected]>>:
>>
>>         Hi.
>>
>>         Because the disc requires three different hosts, the default
>>         number of replications 3.
>>
>>         2014-10-29 10:56 GMT+03:00 Vickie CH <[email protected]
>>         <mailto:[email protected]>>:
>>
>>
>>             Hi all,
>>                    Try to use two OSDs to create a cluster. After the
>>             deply finished, I found the health status is "88
>>             active+degraded" "104 active+remapped". Before use 2 osds to
>>             create cluster the result is ok. I'm confuse why this
>>             situation happened. Do I need to set crush map to fix this
>>             problem?
>>
>>
>>             ----------ceph.conf---------------------------------
>>             [global]
>>             fsid = c404ded6-4086-4f0b-b479-89bc018af954
>>             mon_initial_members = storage0
>>             mon_host = 192.168.1.10
>>             auth_cluster_required = cephx
>>             auth_service_required = cephx
>>             auth_client_required = cephx
>>             filestore_xattr_use_omap = true
>>             osd_pool_default_size = 2
>>             osd_pool_default_min_size = 1
>>             osd_pool_default_pg_num = 128
>>             osd_journal_size = 2048
>>             osd_pool_default_pgp_num = 128
>>             osd_mkfs_type = xfs
>>             ---------------------------------------------------------
>>
>>             -----------ceph -s-----------------------------------
>>             cluster c404ded6-4086-4f0b-b479-89bc018af954
>>                   health HEALTH_WARN 88 pgs degraded; 192 pgs stuck
>> unclean
>>                   monmap e1: 1 mons at {storage0=192.168.10.10:6789/0
>>             <http://192.168.10.10:6789/0>}, election epoch 2, quorum 0
>>             storage0
>>                   osdmap e20: 2 osds: 2 up, 2 in
>>                    pgmap v45: 192 pgs, 3 pools, 0 bytes data, 0 objects
>>                          79752 kB used, 1858 GB / 1858 GB avail
>>                                88 active+degraded
>>                               104 active+remapped
>>             --------------------------------------------------------
>>
>>
>>             Best wishes,
>>             Mika
>>
>>             _______________________________________________
>>             ceph-users mailing list
>>             [email protected] <mailto:[email protected]>
>>             http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>>
>>         --
>>         С уважением, Фасихов Ирек Нургаязович
>>         Моб.: +79229045757
>>
>>
>>
>>
>>
>> --
>> С уважением, Фасихов Ирек Нургаязович
>> Моб.: +79229045757
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> [email protected]
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to