Re: [ceph-users] PG stuck in active+clean+remapped

2019-04-17 Thread Vladimir Prokofev
-users-ceph.com/2017-January/015550.html пн, 1 апр. 2019 г. в 12:23, Vladimir Prokofev : > As we fixed failed node next day, cluster rebalanced to it's original > state without any issues, so crush dump would be irrelevant at this point I > guess. Will have to wait for next occurence.

Re: [ceph-users] PG stuck in active+clean+remapped

2019-04-01 Thread Vladimir Prokofev
"require_feature_tunables2": 1, "has_v2_rules": 0, "require_feature_tunables3": 1, "has_v3_rules": 0, "has_v4_buckets": 0, "require_feature_tunables5": 0, "has_v5_rules": 0 }, вс

[ceph-users] PG stuck in active+clean+remapped

2019-03-26 Thread Vladimir Prokofev
CEPH 12.2.11, pool size 3, min_size 2. One node went down today(private network interface started flapping, and after a while OSD processes crashed), no big deal, cluster recovered, but not completely. 1 PG stuck in active+clean+remapped state. PG_STAT OBJECTS MISSING_ON_PRIMARY DEGRADED

Re: [ceph-users] Problem replacing osd with ceph-deploy

2019-02-01 Thread Vladimir Prokofev
Your output looks a bit weird, but still, this is normal for bluestore. It creates small separate data partition that is presented as XFS mounted in /var/lib/ceph/osd, while real data partition is hidden as raw(bluestore) block device. It's no longer possible to check disk utilisation with df

[ceph-users] osd deployment: DB/WAL links

2019-01-21 Thread Vladimir Prokofev
Hello list. Today while redeploying an OSD I've noticed that links to DB/WAL devices are pointing to partitions themselves, not to the partition UUID how it was before. I think that changed with latest ceph-deploy. I'm using 12.2.2 on my mon/osd nodes. ceph-deploy is 2.0.1 on admin node. All

Re: [ceph-users] yet another deep-scrub performance topic

2018-12-12 Thread Vladimir Prokofev
Thank you all for your input. My best guess at the moment is that deep-scrub performs as it should, and the issue is that it just has no limits on its performance, so it uses all the OSD time it can. Even if it has lower priority than client IO, it still can fill disk queue, and effectively

[ceph-users] yet another deep-scrub performance topic

2018-12-10 Thread Vladimir Prokofev
Hello list. Deep scrub totally kills cluster performance. First of all, it takes several minutes to complete: 2018-12-09 01:39:53.857994 7f2d32fde700 0 log_channel(cluster) log [DBG] : 4.75 deep-scrub starts 2018-12-09 01:46:30.703473 7f2d32fde700 0 log_channel(cluster) log [DBG] : 4.75

Re: [ceph-users] limited disk slots - should I ran OS on SD card ?

2018-08-15 Thread Vladimir Prokofev
I'm running small CEPH cluster of 9 OSD nodes, with systems hosted on USB sticks exactly for the same reason - not enough disk slots. Works fine for almost 2 years now. 2018-08-15 1:13 GMT+03:00 Paul Emmerich : > I've seen the OS running on SATA DOMs and cheap USB sticks. > It works well for

Re: [ceph-users] rgw make container private again

2018-03-30 Thread Vladimir Prokofev
k if this issue is reproducible? If so - this seems to be a ticket-worthy. 2018-03-30 17:40 GMT+03:00 Vladimir Prokofev <v...@prokofev.me>: > CEPH 12.2.2, RGW. > I'm using it as an object storage endpoint for Openstack. > > Recently while browsing an object storage from Horizon

[ceph-users] rgw make container private again

2018-03-30 Thread Vladimir Prokofev
CEPH 12.2.2, RGW. I'm using it as an object storage endpoint for Openstack. Recently while browsing an object storage from Horizon, I accidently marked container as public. The issue is - I can't make it private again! Docs state that to do it I should simply delete X-Container-Read metadata, but

Re: [ceph-users] Stuck in creating+activating

2018-03-17 Thread Vladimir Prokofev
My first guess would be PG overdose protection kicked in [1][2] You can try fixing it by increasing allowed number of PG per OSD with ceph tell mon.* injectargs '--mon_max_pg_per_osd 500' ceph tell osd.* injectargs '--mon_max_pg_per_osd 500' and then triggering CRUSH algorithm update by restarting

Re: [ceph-users] Random health OSD_SCRUB_ERRORS on various OSDs, after pg repair back to HEALTH_OK

2018-03-05 Thread Vladimir Prokofev
194 Temperature_Celsius 0x0022 031 017 000Old_age Always > - 31 (0 17 0 0 0) > 195 Hardware_ECC_Recovered 0x001a 005 001 000Old_age Always > - 224621855 > 197 Current_Pending_Sector 0x0012 100 100 000Old_age Always &

Re: [ceph-users] Random health OSD_SCRUB_ERRORS on various OSDs, after pg repair back to HEALTH_OK

2018-03-05 Thread Vladimir Prokofev
> candidate had a read error speaks for itself - while scrubbing it coudn't read data. I had similar issue, and it was just OSD dying - errors and relocated sectors in SMART, just replaced the disk. But in your case it seems that errors are on different OSDs? Are your OSDs all healthy? You can use

Re: [ceph-users] All pools full after one OSD got OSD_FULL state

2018-03-05 Thread Vladimir Prokofev
I'll pitch in my personal expirience. When single OSD in a pool becomes full(95% used), then all client IO writes to this pool must stop, even if other OSDs are almost free. This is done for the purpose of data intergity. [1] To avoid this you need to balance your failure domains. For example,

Re: [ceph-users] PG overdose protection causing PG unavailability

2018-02-24 Thread Vladimir Prokofev
In my case we don't even use "default" ruleset in CRUSH - no pool has this ruleset associated with it. Adding OSDs doesn't lead to any PG recalculation or data movement. It's triggered only after modifying CRUSH map and placing OSDs in appropriate failure domain. This way you can add any number of

[ceph-users] flatten clones are not sparse?

2018-02-15 Thread Vladimir Prokofev
CEPH 12.2.2 So I have a snapshot of an image: $ rbd du ssd-volumes/volume-4fed3f31-0802-4850-91bc-17e6da05697d NAME PROVISIONED USED volume-4fed3f31-0802-4850-91bc-17e6da05697d@snapshot-d2145e21-99a7-4e2e-9138-ab3e975f8113 20480M 8088M

[ceph-users] consequence of losing WAL/DB device with bluestore?

2018-01-29 Thread Vladimir Prokofev
Hello. In short: what are the consequence of loosing external WAL/DB device(assuming it’s SSD) in bluestore? In comparison with filestore - we used to have an external SSD for journaling multiple HDD OSDs. Hardware failure of such a device would not be that big of a deal, as we can quickly use

Re: [ceph-users] Increasing PG number

2018-01-02 Thread Vladimir Prokofev
Increased number of PGs in multiple pools in a production cluster on 12.2.2 recently - zero issues. CEPH claims that increasing pg_num and pgp_num are safe operations, which are essential for it's ability to scale, and this sounds pretty reasonable to me. [1] [1]

Re: [ceph-users] luminous OSD_ORPHAN

2017-12-20 Thread Vladimir Prokofev
impact. The fix was also quite easy to figure out. 2017-12-20 4:03 GMT+03:00 Brad Hubbard <bhubb...@redhat.com>: > Version? > > See http://tracker.ceph.com/issues/22346 for a (limited) explanation. > > On Tue, Dec 19, 2017 at 6:35 PM, Vladimir Prokofev <v...@prokofev.me&g

Re: [ceph-users] luminous OSD_ORPHAN

2017-12-19 Thread Vladimir Prokofev
OSDs. Case closed. 2017-12-19 10:39 GMT+03:00 Vladimir Prokofev <v...@prokofev.me>: > Hello. > > After some furious "ceph-deploy osd prepare/osd zap" cycles to figure out > a correct command for ceph-deploy to create a bluestore HDD with wal/db > SSD, I now have

[ceph-users] luminous OSD_ORPHAN

2017-12-18 Thread Vladimir Prokofev
Hello. After some furious "ceph-deploy osd prepare/osd zap" cycles to figure out a correct command for ceph-deploy to create a bluestore HDD with wal/db SSD, I now have orphant OSDs, which are nowhere to be found in CRUSH map! $ ceph health detail HEALTH_WARN 4 osds exist in the crush map but

Re: [ceph-users] Fwd: FileStore vs BlueStore

2017-09-20 Thread Vladimir Prokofev
> So, filestore: > - write data to the journal > - write metadata > - move data from journal to definitive storage > > Bluestore: > - write data to definitive storage (free space, not overwriting anything) > - write metadata Wait, doesn't that mean that when I move to Bluestore from Filestore

Re: [ceph-users] Restart ceph cluster

2017-05-12 Thread Vladimir Prokofev
ore than ceph question, but still. Surely "ceph tell" is a goto solution for any configuration updates. 2017-05-12 17:42 GMT+03:00 mj <li...@merit.unu.edu>: > Hi, > > On 05/12/2017 03:35 PM, Vladimir Prokofev wrote: > >> My best guess is that using systemd you can

Re: [ceph-users] Restart ceph cluster

2017-05-12 Thread Vladimir Prokofev
Intersting question. AFAIK, there's no built-in solution. Also, if you think about it, restarting whole cluster at once can lead to service interruption, as you can easily bring all PG copies in stale state for a short time, and even longer if some OSDs won't go up for some reason. You should

[ceph-users] snapshot removal slows cluster

2017-04-26 Thread Vladimir Prokofev
Hello ceph-users. Short description: during snapshot removal osd usilisation goes up to 100%, which leads to slow requests and VM failures due to IOPS stall. We're using Openstack Cinder with CEPH cluster as a volume backend. CEPH version is 10.2.6. We also using cinder-backup to create backups