Hi Brad,
Thanks, that helped to get the query info on the inconsistent PG 18.2:
{
"state": "active+clean+inconsistent",
"snap_trimq": "[]",
"snap_trimq_len": 0,
"epoch": 121293,
"up": [
21,
28
],
"acting": [
21,
28
],
"actingbackfill": [
"21",
"28"
],
"info": {
"pgid": "18.2",
"last_update": "121290'698339",
"last_complete": "121290'698339",
"log_tail": "121272'696825",
"last_user_version": 698319,
"last_backfill": "MAX",
"last_backfill_bitwise": 0,
"purged_snaps": [],
"history": {
"epoch_created": 24431,
"epoch_pool_created": 24431,
"last_epoch_started": 121152,
"last_interval_started": 121151,
"last_epoch_clean": 121152,
"last_interval_clean": 121151,
"last_epoch_split": 0,
"last_epoch_marked_full": 106367,
"same_up_since": 121148,
"same_interval_since": 121151,
"same_primary_since": 121020,
"last_scrub": "121290'698339",
"last_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_deep_scrub": "121290'698339",
"last_deep_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_clean_scrub_stamp": "2018-06-11 15:28:20.335739"
},
"stats": {
"version": "121290'698339",
"reported_seq": "1055277",
"reported_epoch": "121293",
"state": "active+clean+inconsistent",
"last_fresh": "2018-06-27 08:33:20.764603",
"last_change": "2018-06-27 03:55:44.291146",
"last_active": "2018-06-27 08:33:20.764603",
"last_peered": "2018-06-27 08:33:20.764603",
"last_clean": "2018-06-27 08:33:20.764603",
"last_became_active": "2018-06-21 16:35:46.487783",
"last_became_peered": "2018-06-21 16:35:46.487783",
"last_unstale": "2018-06-27 08:33:20.764603",
"last_undegraded": "2018-06-27 08:33:20.764603",
"last_fullsized": "2018-06-27 08:33:20.764603",
"mapping_epoch": 121151,
"log_start": "121272'696825",
"ondisk_log_start": "121272'696825",
"created": 24431,
"last_epoch_clean": 121152,
"parent": "0.0",
"parent_split_bits": 0,
"last_scrub": "121290'698339",
"last_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_deep_scrub": "121290'698339",
"last_deep_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_clean_scrub_stamp": "2018-06-11 15:28:20.335739",
"log_size": 1514,
"ondisk_log_size": 1514,
"stats_invalid": false,
"dirty_stats_invalid": false,
"omap_stats_invalid": false,
"hitset_stats_invalid": false,
"hitset_bytes_stats_invalid": false,
"pin_stats_invalid": true,
"snaptrimq_len": 0,
"stat_sum": {
"num_bytes": 0,
"num_objects": 116,
"num_object_clones": 0,
"num_object_copies": 232,
"num_objects_missing_on_primary": 0,
"num_objects_missing": 0,
"num_objects_degraded": 0,
"num_objects_misplaced": 0,
"num_objects_unfound": 0,
"num_objects_dirty": 111,
"num_whiteouts": 0,
"num_read": 168436,
"num_read_kb": 25417188,
"num_write": 3370202,
"num_write_kb": 0,
"num_scrub_errors": 2,
"num_shallow_scrub_errors": 0,
"num_deep_scrub_errors": 2,
"num_objects_recovered": 207,
"num_bytes_recovered": 0,
"num_keys_recovered": 9482826,
"num_objects_omap": 107,
"num_objects_hit_set_archive": 0,
"num_bytes_hit_set_archive": 0,
"num_flush": 0,
"num_flush_kb": 0,
"num_evict": 0,
"num_evict_kb": 0,
"num_promote": 0,
"num_flush_mode_high": 0,
"num_flush_mode_low": 0,
"num_evict_mode_some": 0,
"num_evict_mode_full": 0,
"num_objects_pinned": 0,
"num_legacy_snapsets": 0
},
"up": [
21,
28
],
"acting": [
21
28
],
"blocked_by": [],
"up_primary": 21,
"acting_primary": 21
},
"empty": 0,
"dne": 0,
"incomplete": 0,
"last_epoch_started": 121152,
"hit_set_history": {
"current_last_update": "0'0",
"history": []
}
},
"peer_info": [
{
"peer": "28",
"pgid": "18.2",
"last_update": "121290'698339",
"last_complete": "121172'661331",
"log_tail": "121127'652751",
"last_user_version": 0,
"last_backfill": "MAX",
"last_backfill_bitwise": 1,
"purged_snaps": [],
"history": {
"epoch_created": 24431,
"epoch_pool_created": 24431,
"last_epoch_started": 121152,
"last_interval_started": 121151,
"last_epoch_clean": 121152,
"last_interval_clean": 121151,
"last_epoch_split": 0,
"last_epoch_marked_full": 106367,
"same_up_since": 121148,
"same_interval_since": 121151,
"same_primary_since": 121020,
"last_scrub": "121290'698339",
"last_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_deep_scrub": "121290'698339",
"last_deep_scrub_stamp": "2018-06-27 03:55:44.291060",
"last_clean_scrub_stamp": "2018-06-11 15:28:20.335739"
},
"stats": {
"version": "121131'654251",
"reported_seq": "959540",
"reported_epoch": "121150",
"state":
"active+undersized+degraded+remapped+inconsistent+backfilling",
"last_fresh": "2018-06-21 16:35:44.468284",
"last_change": "2018-06-21 16:34:12.447803",
"last_active": "2018-06-21 16:35:44.468284",
"last_peered": "2018-06-21 16:35:44.468284",
"last_clean": "2018-06-21 16:27:07.835328",
"last_became_active": "2018-06-21 16:33:24.246631",
"last_became_peered": "2018-06-21 16:33:24.246631",
"last_unstale": "2018-06-21 16:35:44.468284",
"last_undegraded": "2018-06-21 16:33:23.997020",
"last_fullsized": "2018-06-21 16:33:23.994195",
"mapping_epoch": 121151,
"log_start": "121127'652725",
"created": 24431,
"last_epoch_clean": 121145,
"parent": "0.0",
"parent_split_bits": 0,
"last_scrub": "121131'654251",
"last_scrub_stamp": "2018-06-21 16:27:07.835266",
"last_deep_scrub": "121131'654251",
"last_deep_scrub_stamp": "2018-06-21 16:27:07.835266",
"last_clean_scrub_stamp": "2018-06-11 15:28:20.335739",
"log_size": 1526,
"ondisk_log_size": 1526,
"stats_invalid": false,
"dirty_stats_invalid": false,
"omap_stats_invalid": false,
"hitset_stats_invalid": false,
"hitset_bytes_stats_invalid": false,
"pin_stats_invalid": true,
"snaptrimq_len": 0,
"stat_sum": {
"num_bytes": 0,
"num_objects": 69,
"num_object_clones": 0,
"num_object_copies": 138,
"num_objects_missing_on_primary": 0,
"num_objects_missing": 0,
"num_objects_degraded": 1,
"num_objects_misplaced": 0,
"num_objects_unfound": 0,
"num_objects_dirty": 64,
"num_whiteouts": 0,
"num_read": 14057,
"num_read_kb": 454200,
"num_write": 797911,
"num_write_kb": 0,
"num_scrub_errors": 0,
"num_shallow_scrub_errors": 0,
"num_deep_scrub_errors": 0,
"num_objects_recovered": 207,
"num_bytes_recovered": 0,
"num_keys_recovered": 9482826,
"num_objects_omap": 60,
"num_objects_hit_set_archive": 0,
"num_bytes_hit_set_archive": 0,
"num_flush": 0,
"num_flush_kb": 0,
"num_evict": 0,
"num_evict_kb": 0,
"num_promote": 0,
"num_flush_mode_high": 0,
"num_flush_mode_low": 0,
"num_evict_mode_some": 0,
"num_evict_mode_full": 0,
"num_objects_pinned": 0,
"num_legacy_snapsets": 0
},
"up": [
21,
28
],
"acting": [
21,
28
],
"blocked_by": [],
"up_primary": 21,
"acting_primary": 21
},
"empty": 0,
"dne": 0,
"incomplete": 0,
"last_epoch_started": 121152,
"hit_set_history": {
"current_last_update": "0'0",
"history": []
}
}
],
"recovery_state": [
{
"name": "Started/Primary/Active",
"enter_time": "2018-06-21 16:35:46.478007",
"might_have_unfound": [],
"recovery_progress": {
"backfill_targets": [],
"waiting_on_backfill": [],
"last_backfill_started": "MIN",
"backfill_info": {
"begin": "MIN",
"end": "MIN",
"objects": []
},
"peer_backfill_info": [],
"backfills_in_flight": [],
"recovering": [],
"pg_backend": {
"pull_from_peer": [],
"pushing": []
}
},
"scrub": {
"scrubber.epoch_start": "121151",
"scrubber.active": false,
"scrubber.state": "INACTIVE",
"scrubber.start": "MIN",
"scrubber.end": "MIN",
"scrubber.subset_last_update": "0'0",
"scrubber.deep": false,
"scrubber.seed": 0,
"scrubber.waiting_on": 0,
"scrubber.waiting_on_whom": []
}
},
{
"name": "Started",
"enter_time": "2018-06-21 16:35:45.052939"
}
],
"agent_state": {}
}
Thanks for trying to help out.
Cheers
----- Original Message -----
> From: "Brad Hubbard" <[email protected]>
> To: "Andrei Mikhailovsky" <[email protected]>
> Cc: "ceph-users" <[email protected]>
> Sent: Wednesday, 27 June, 2018 00:18:19
> Subject: Re: [ceph-users] fixing unrepairable inconsistent PG
> Try setting the osd caps to 'allow *' for client.admin or running the
> command using an id that has that access such as
> mgr.arh-ibstorage1-ib.
>
> On Wed, Jun 27, 2018 at 1:32 AM, Andrei Mikhailovsky <[email protected]>
> wrote:
>> Hi Brad,
>>
>> Here is the output of the "ceph auth list" command (I have removed the key:
>> line
>> which was present in every single entry, including the osd.21):
>>
>> # ceph auth list
>> installed auth entries:
>>
>> mds.arh-ibstorage1-ib
>> caps: [mds] allow
>> caps: [mgr] allow profile mds
>> caps: [mon] allow profile mds
>> caps: [osd] allow *
>> mds.arh-ibstorage2-ib
>> caps: [mds] allow
>> caps: [mgr] allow profile mds
>> caps: [mon] allow profile mds
>> caps: [osd] allow *
>> osd.0
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.1
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.10
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.11
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.12
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.13
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.14
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.15
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.16
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.17
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.18
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.19
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.2
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.20
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.21
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.22
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.23
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.24
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.25
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.27
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.28
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.29
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.3
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.30
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.4
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.5
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.6
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.7
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.8
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> osd.9
>> caps: [mgr] allow profile osd
>> caps: [mon] allow profile osd
>> caps: [osd] allow *
>> client.admin
>> caps: [mds] allow rwx
>> caps: [mgr] allow *
>> caps: [mon] allow rwx
>> caps: [osd] allow rwx
>> client.arh-ibstorage1-ib.csprdc.arhont.com
>> caps: [mgr] allow r
>> caps: [mon] allow rw
>> caps: [osd] allow rwx
>> client.bootstrap-mds
>> caps: [mgr] allow r
>> caps: [mon] allow profile bootstrap-mds
>> client.bootstrap-mgr
>> caps: [mon] allow profile bootstrap-mgr
>> client.bootstrap-osd
>> caps: [mgr] allow r
>> caps: [mon] allow profile bootstrap-osd
>> client.bootstrap-rgw
>> caps: [mgr] allow r
>> caps: [mon] allow profile bootstrap-rgw
>> client.ceph-monitors
>> caps: [mgr] allow r
>> caps: [mon] allow r
>> client.libvirt
>> caps: [mgr] allow r
>> caps: [mon] allow r
>> caps: [osd] allow class-read object_prefix rbd_children, allow rwx
>> pool=libvirt-pool
>> client.primary-ubuntu-1
>> caps: [mgr] allow r
>> caps: [mon] allow r
>> caps: [osd] allow rwx pool=Primary-ubuntu-1
>> client.radosgw1.gateway
>> caps: [mgr] allow r
>> caps: [mon] allow rwx
>> caps: [osd] allow rwx
>> client.radosgw2.gateway
>> caps: [mgr] allow r
>> caps: [mon] allow rw
>> caps: [osd] allow rwx
>> client.ssdcs
>> caps: [mgr] allow r
>> caps: [mon] allow r
>> caps: [osd] allow class-read object_prefix rbd_children, allow rwx
>> pool=ssdcs
>>
>> mgr.arh-ibstorage1-ib
>> caps: [mds] allow *
>> caps: [mon] allow profile mgr
>> caps: [osd] allow *
>> mgr.arh-ibstorage2-ib
>> caps: [mds] allow *
>> caps: [mon] allow profile mgr
>> caps: [osd] allow *
>>
>>
>>
>>
>>
>>
>>
>> I have ran this command on all pgs in the cluster and it shows the same error
>> message for all of them. For example:
>>
>> Error EPERM: problem getting command descriptions from pg.5.1c9
>>
>> Andrei
>>
>>
>> ----- Original Message -----
>>> From: "Brad Hubbard" <[email protected]>
>>> To: "Andrei Mikhailovsky" <[email protected]>
>>> Cc: "ceph-users" <[email protected]>
>>> Sent: Tuesday, 26 June, 2018 01:10:34
>>> Subject: Re: [ceph-users] fixing unrepairable inconsistent PG
>>
>>> Interesing...
>>>
>>> Can I see the output of "ceph auth list" and can you test whether you
>>> can query any other pg that has osd.21 as its primary?
>>>
>>> On Mon, Jun 25, 2018 at 8:04 PM, Andrei Mikhailovsky <[email protected]>
>>> wrote:
>>>> Hi Brad,
>>>>
>>>> here is the output:
>>>>
>>>> --------------
>>>>
>>>> root@arh-ibstorage1-ib:/home/andrei# ceph --debug_ms 5 --debug_auth 20 pg
>>>> 18.2
>>>> query
>>>> 2018-06-25 10:59:12.100302 7fe23eaa1700 2 Event(0x7fe2400e0140 nevent=5000
>>>> time_id=1).set_owner idx=0 owner=140609690670848
>>>> 2018-06-25 10:59:12.100398 7fe23e2a0700 2 Event(0x7fe24010d030 nevent=5000
>>>> time_id=1).set_owner idx=1 owner=140609682278144
>>>> 2018-06-25 10:59:12.100445 7fe23da9f700 2 Event(0x7fe240139ec0 nevent=5000
>>>> time_id=1).set_owner idx=2 owner=140609673885440
>>>> 2018-06-25 10:59:12.100793 7fe244b28700 1 Processor -- start
>>>> 2018-06-25 10:59:12.100869 7fe244b28700 1 -- - start start
>>>> 2018-06-25 10:59:12.100882 7fe244b28700 5 adding auth protocol: cephx
>>>> 2018-06-25 10:59:12.101046 7fe244b28700 2 auth: KeyRing::load: loaded key
>>>> file
>>>> /etc/ceph/ceph.client.admin.keyring
>>>> 2018-06-25 10:59:12.101244 7fe244b28700 1 -- - --> 192.168.168.201:6789/0
>>>> --
>>>> auth(proto 0 30 bytes epoch 0) v1 -- 0x7fe240174b80 con 0
>>>> 2018-06-25 10:59:12.101264 7fe244b28700 1 -- - --> 192.168.168.202:6789/0
>>>> --
>>>> auth(proto 0 30 bytes epoch 0) v1 -- 0x7fe240175010 con 0
>>>> 2018-06-25 10:59:12.101690 7fe23e2a0700 1 -- 192.168.168.201:0/3046734987
>>>> learned_addr learned my addr 192.168.168.201:0/3046734987
>>>> 2018-06-25 10:59:12.101890 7fe23e2a0700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_CONNECTING_WAIT_ACK_SEQ
>>>> pgs=0 cs=0 l=1)._process_connection got newly_acked_seq 0 vs out_seq 0
>>>> 2018-06-25 10:59:12.102030 7fe23da9f700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1
>>>> s=STATE_CONNECTING_WAIT_ACK_SEQ
>>>> pgs=0 cs=0 l=1)._process_connection got newly_acked_seq 0 vs out_seq 0
>>>> 2018-06-25 10:59:12.102450 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 1 0x7fe234002670 mon_map magic: 0 v1
>>>> 2018-06-25 10:59:12.102494 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 2 0x7fe234002b70 auth_reply(proto 2 0 (0) Success) v1
>>>> 2018-06-25 10:59:12.102542 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 1 ==== mon_map magic: 0 v1 ==== 505+0+0
>>>> (2386987630 0 0) 0x7fe234002670 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.102629 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1
>>>> ====
>>>> 33+0+0 (1469975654 0 0) 0x7fe234002b70 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.102655 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.102657 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.102658 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.102661 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service auth
>>>> 2018-06-25 10:59:12.102662 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 0 need 53
>>>> 2018-06-25 10:59:12.102666 7fe23ca9d700 10 cephx client: handle_response
>>>> ret = 0
>>>> 2018-06-25 10:59:12.102671 7fe23ca9d700 10 cephx client: got initial
>>>> server
>>>> challenge 6522ec95fb2eb487
>>>> 2018-06-25 10:59:12.102673 7fe23ca9d700 10 cephx client: validate_tickets:
>>>> want=53 need=53 have=0
>>>> 2018-06-25 10:59:12.102674 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.102675 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.102676 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.102676 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service auth
>>>> 2018-06-25 10:59:12.102677 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 0 need 53
>>>> 2018-06-25 10:59:12.102678 7fe23ca9d700 10 cephx client: want=53 need=53
>>>> have=0
>>>> 2018-06-25 10:59:12.102680 7fe23ca9d700 10 cephx client: build_request
>>>> 2018-06-25 10:59:12.102702 7fe23da9f700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=333625 cs=1 l=1). rx
>>>> mon.0
>>>> seq 1 0x7fe228001490 mon_map magic: 0 v1
>>>> 2018-06-25 10:59:12.102739 7fe23ca9d700 10 cephx client: get auth session
>>>> key:
>>>> client_challenge 80f2a24093f783c5
>>>> 2018-06-25 10:59:12.102743 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.202:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 --
>>>> 0x7fe224002080
>>>> con 0
>>>> 2018-06-25 10:59:12.102737 7fe23da9f700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=333625 cs=1 l=1). rx
>>>> mon.0
>>>> seq 2 0x7fe2280019c0 auth_reply(proto 2 0 (0) Success) v1
>>>> 2018-06-25 10:59:12.102776 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.0 192.168.168.201:6789/0 1 ==== mon_map magic: 0 v1 ==== 505+0+0
>>>> (2386987630 0 0) 0x7fe228001490 con 0x7fe24017a420
>>>> 2018-06-25 10:59:12.102821 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.0 192.168.168.201:6789/0 2 ==== auth_reply(proto 2 0 (0) Success) v1
>>>> ====
>>>> 33+0+0 (3800394028 0 0) 0x7fe2280019c0 con 0x7fe24017a420
>>>> 2018-06-25 10:59:12.102833 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.102834 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.102835 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.102836 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service auth
>>>> 2018-06-25 10:59:12.102837 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 0 need 53
>>>> 2018-06-25 10:59:12.102839 7fe23ca9d700 10 cephx client: handle_response
>>>> ret = 0
>>>> 2018-06-25 10:59:12.102841 7fe23ca9d700 10 cephx client: got initial
>>>> server
>>>> challenge ccd69ce967642f7
>>>> 2018-06-25 10:59:12.102842 7fe23ca9d700 10 cephx client: validate_tickets:
>>>> want=53 need=53 have=0
>>>> 2018-06-25 10:59:12.102843 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.102843 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.102844 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.102845 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service auth
>>>> 2018-06-25 10:59:12.102845 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 0 need 53
>>>> 2018-06-25 10:59:12.102846 7fe23ca9d700 10 cephx client: want=53 need=53
>>>> have=0
>>>> 2018-06-25 10:59:12.102848 7fe23ca9d700 10 cephx client: build_request
>>>> 2018-06-25 10:59:12.102881 7fe23ca9d700 10 cephx client: get auth session
>>>> key:
>>>> client_challenge 6ddb6fdc4176ea6a
>>>> 2018-06-25 10:59:12.102884 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.201:6789/0 -- auth(proto 2 32 bytes epoch 0) v1 --
>>>> 0x7fe2240032d0
>>>> con 0
>>>> 2018-06-25 10:59:12.103402 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 3 0x7fe234002200 auth_reply(proto 2 0 (0) Success) v1
>>>> 2018-06-25 10:59:12.103449 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1
>>>> ====
>>>> 206+0+0 (195487815 0 0) 0x7fe234002200 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.103468 7fe23ca9d700 10 cephx client: handle_response
>>>> ret = 0
>>>> 2018-06-25 10:59:12.103469 7fe23ca9d700 10 cephx client:
>>>> get_auth_session_key
>>>> 2018-06-25 10:59:12.103471 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> got 1 keys
>>>> 2018-06-25 10:59:12.103472 7fe23ca9d700 10 cephx: got key for service_id
>>>> auth
>>>> 2018-06-25 10:59:12.103508 7fe23ca9d700 10 cephx: ticket.secret_id=3687
>>>> 2018-06-25 10:59:12.103510 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> service auth secret_id 3687 session_key [KEY] validity=43200.000000
>>>> 2018-06-25 10:59:12.103527 7fe23ca9d700 10 cephx: ticket expires=2018-06-25
>>>> 22:59:12.103526 renew_after=2018-06-25 19:59:12.103526
>>>> 2018-06-25 10:59:12.103533 7fe23ca9d700 10 cephx client: want=53 need=53
>>>> have=0
>>>> 2018-06-25 10:59:12.103534 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.103535 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.103536 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.103537 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 32 need 21
>>>> 2018-06-25 10:59:12.103539 7fe23ca9d700 10 cephx client: validate_tickets:
>>>> want=53 need=21 have=32
>>>> 2018-06-25 10:59:12.103540 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.103541 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.103542 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.103542 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 32 need 21
>>>> 2018-06-25 10:59:12.103543 7fe23ca9d700 10 cephx client: want=53 need=21
>>>> have=32
>>>> 2018-06-25 10:59:12.103544 7fe23ca9d700 10 cephx client: build_request
>>>> 2018-06-25 10:59:12.103545 7fe23ca9d700 10 cephx client: get service keys:
>>>> want=53 need=21 have=32
>>>> 2018-06-25 10:59:12.103570 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.202:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 --
>>>> 0x7fe224007010
>>>> con 0
>>>> 2018-06-25 10:59:12.103657 7fe23da9f700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=333625 cs=1 l=1). rx
>>>> mon.0
>>>> seq 3 0x7fe228001020 auth_reply(proto 2 0 (0) Success) v1
>>>> 2018-06-25 10:59:12.103709 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.0 192.168.168.201:6789/0 3 ==== auth_reply(proto 2 0 (0) Success) v1
>>>> ====
>>>> 206+0+0 (2366624548 0 0) 0x7fe228001020 con 0x7fe24017a420
>>>> 2018-06-25 10:59:12.103729 7fe23ca9d700 10 cephx client: handle_response
>>>> ret = 0
>>>> 2018-06-25 10:59:12.103731 7fe23ca9d700 10 cephx client:
>>>> get_auth_session_key
>>>> 2018-06-25 10:59:12.103733 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> got 1 keys
>>>> 2018-06-25 10:59:12.103734 7fe23ca9d700 10 cephx: got key for service_id
>>>> auth
>>>> 2018-06-25 10:59:12.103774 7fe23ca9d700 10 cephx: ticket.secret_id=3687
>>>> 2018-06-25 10:59:12.103776 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> service auth secret_id 3687 session_key [KEY] validity=43200.000000
>>>> 2018-06-25 10:59:12.103792 7fe23ca9d700 10 cephx: ticket expires=2018-06-25
>>>> 22:59:12.103791 renew_after=2018-06-25 19:59:12.103791
>>>> 2018-06-25 10:59:12.103798 7fe23ca9d700 10 cephx client: want=53 need=53
>>>> have=0
>>>> 2018-06-25 10:59:12.103799 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.103800 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.103801 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.103802 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 32 need 21
>>>> 2018-06-25 10:59:12.103804 7fe23ca9d700 10 cephx client: validate_tickets:
>>>> want=53 need=21 have=32
>>>> 2018-06-25 10:59:12.103806 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mon
>>>> 2018-06-25 10:59:12.103806 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service osd
>>>> 2018-06-25 10:59:12.103807 7fe23ca9d700 10 cephx: set_have_need_key no
>>>> handler
>>>> for service mgr
>>>> 2018-06-25 10:59:12.103808 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 32 need 21
>>>> 2018-06-25 10:59:12.103808 7fe23ca9d700 10 cephx client: want=53 need=21
>>>> have=32
>>>> 2018-06-25 10:59:12.103812 7fe23ca9d700 10 cephx client: build_request
>>>> 2018-06-25 10:59:12.103813 7fe23ca9d700 10 cephx client: get service keys:
>>>> want=53 need=21 have=32
>>>> 2018-06-25 10:59:12.103834 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.201:6789/0 -- auth(proto 2 165 bytes epoch 0) v1 --
>>>> 0x7fe224009dd0
>>>> con 0
>>>> 2018-06-25 10:59:12.104168 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 4 0x7fe234002200 auth_reply(proto 2 0 (0) Success) v1
>>>> 2018-06-25 10:59:12.104201 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 4 ==== auth_reply(proto 2 0 (0) Success) v1
>>>> ====
>>>> 580+0+0 (56981162 0 0) 0x7fe234002200 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.104223 7fe23ca9d700 10 cephx client: handle_response
>>>> ret = 0
>>>> 2018-06-25 10:59:12.104226 7fe23ca9d700 10 cephx client:
>>>> get_principal_session_key session_key [KEY]
>>>> 2018-06-25 10:59:12.104238 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> got 3 keys
>>>> 2018-06-25 10:59:12.104240 7fe23ca9d700 10 cephx: got key for service_id
>>>> mon
>>>> 2018-06-25 10:59:12.104276 7fe23ca9d700 10 cephx: ticket.secret_id=44205
>>>> 2018-06-25 10:59:12.104277 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> service mon secret_id 44205 session_key [KEY] validity=3600.000000
>>>> 2018-06-25 10:59:12.104285 7fe23ca9d700 10 cephx: ticket expires=2018-06-25
>>>> 11:59:12.104284 renew_after=2018-06-25 11:44:12.104284
>>>> 2018-06-25 10:59:12.104290 7fe23ca9d700 10 cephx: got key for service_id
>>>> osd
>>>> 2018-06-25 10:59:12.104313 7fe23ca9d700 10 cephx: ticket.secret_id=44205
>>>> 2018-06-25 10:59:12.104314 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> service osd secret_id 44205 session_key [KEY] validity=3600.000000
>>>> 2018-06-25 10:59:12.104329 7fe23ca9d700 10 cephx: ticket expires=2018-06-25
>>>> 11:59:12.104329 renew_after=2018-06-25 11:44:12.104329
>>>> 2018-06-25 10:59:12.104333 7fe23ca9d700 10 cephx: got key for service_id
>>>> mgr
>>>> 2018-06-25 10:59:12.104355 7fe23ca9d700 10 cephx: ticket.secret_id=204
>>>> 2018-06-25 10:59:12.104356 7fe23ca9d700 10 cephx:
>>>> verify_service_ticket_reply
>>>> service mgr secret_id 204 session_key [KEY] validity=3600.000000
>>>> 2018-06-25 10:59:12.104368 7fe23ca9d700 10 cephx: ticket expires=2018-06-25
>>>> 11:59:12.104368 renew_after=2018-06-25 11:44:12.104368
>>>> 2018-06-25 10:59:12.104373 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 53 need 0
>>>> 2018-06-25 10:59:12.104376 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1 s=STATE_OPEN pgs=333625 cs=1
>>>> l=1).mark_down
>>>> 2018-06-25 10:59:12.104384 7fe23ca9d700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6789/0 conn(0x7fe24017a420 :-1 s=STATE_OPEN pgs=333625 cs=1
>>>> l=1)._stop
>>>> 2018-06-25 10:59:12.104426 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.202:6789/0 -- mon_subscribe({monmap=0+}) v2 -- 0x7fe240180bb0
>>>> con 0
>>>> 2018-06-25 10:59:12.104442 7fe23ca9d700 10 cephx: validate_tickets want 53
>>>> have
>>>> 53 need 0
>>>> 2018-06-25 10:59:12.104444 7fe23ca9d700 20 cephx client: need_tickets:
>>>> want=53
>>>> have=53 need=0
>>>> 2018-06-25 10:59:12.104481 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.202:6789/0 -- mon_subscribe({mgrmap=0+}) v2 -- 0x7fe240175010
>>>> con 0
>>>> 2018-06-25 10:59:12.104573 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.202:6789/0 -- mon_subscribe({osdmap=0}) v2 -- 0x7fe24017ea90
>>>> con 0
>>>> 2018-06-25 10:59:12.104979 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 5 0x7fe234002b90 mon_map magic: 0 v1
>>>> 2018-06-25 10:59:12.105008 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 5 ==== mon_map magic: 0 v1 ==== 505+0+0
>>>> (2386987630 0 0) 0x7fe234002b90 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.105022 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 6 0x7fe234001a60 mgrmap(e 139) v1
>>>> 2018-06-25 10:59:12.105058 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 6 ==== mgrmap(e 139) v1 ==== 381+0+0
>>>> (56579516 0
>>>> 0) 0x7fe234001a60 con 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.105066 7fe23e2a0700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=472363 cs=1 l=1). rx
>>>> mon.1
>>>> seq 7 0x7fe234002610 osd_map(121251..121251 src has 120729..121251) v3
>>>> 2018-06-25 10:59:12.105110 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> mon.1 192.168.168.202:6789/0 7 ==== osd_map(121251..121251 src has
>>>> 120729..121251) v3 ==== 18118+0+0 (421862548 0 0) 0x7fe234002610 con
>>>> 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.105405 7fe23da9f700 10 cephx client: build_authorizer
>>>> for
>>>> service mgr
>>>> 2018-06-25 10:59:12.105685 7fe23da9f700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6840/32624 conn(0x7fe2240127b0 :-1
>>>> s=STATE_CONNECTING_WAIT_ACK_SEQ pgs=0 cs=0 l=1)._process_connection got
>>>> newly_acked_seq 0 vs out_seq 0
>>>> 2018-06-25 10:59:12.105720 7fe23da9f700 10 In get_auth_session_handler for
>>>> protocol 2
>>>> 2018-06-25 10:59:12.108653 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> -->
>>>> 192.168.168.203:6828/43673 -- command(tid 1: {"prefix":
>>>> "get_command_descriptions", "pgid": "18.2"}) v1 -- 0x7fe240184580 con 0
>>>> 2018-06-25 10:59:12.109327 7fe23eaa1700 10 cephx client: build_authorizer
>>>> for
>>>> service osd
>>>> 2018-06-25 10:59:12.109828 7fe23eaa1700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.203:6828/43673 conn(0x7fe240180f20 :-1
>>>> s=STATE_CONNECTING_WAIT_ACK_SEQ pgs=0 cs=0 l=1)._process_connection got
>>>> newly_acked_seq 0 vs out_seq 0
>>>> 2018-06-25 10:59:12.109875 7fe23eaa1700 10 In get_auth_session_handler for
>>>> protocol 2
>>>> 2018-06-25 10:59:12.109921 7fe23eaa1700 10 _calc_signature seq 1
>>>> front_crc_ =
>>>> 2696387361 middle_crc = 0 data_crc = 0 sig = 10077981589201542762
>>>> 2018-06-25 10:59:12.109930 7fe23eaa1700 20 Putting signature in client
>>>> message(seq # 1): sig = 10077981589201542762
>>>> 2018-06-25 10:59:12.110382 7fe23eaa1700 10 _calc_signature seq 1
>>>> front_crc_ =
>>>> 1943489909 middle_crc = 0 data_crc = 0 sig = 6955259887491975287
>>>> 2018-06-25 10:59:12.110394 7fe23eaa1700 5 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.203:6828/43673 conn(0x7fe240180f20 :-1
>>>> s=STATE_OPEN_MESSAGE_READ_FOOTER_AND_DISPATCH pgs=26353 cs=1 l=1). rx
>>>> osd.21
>>>> seq 1 0x7fe238000f20 command_reply(tid 1: -1 ) v1
>>>> 2018-06-25 10:59:12.110436 7fe23ca9d700 1 -- 192.168.168.201:0/3046734987
>>>> <==
>>>> osd.21 192.168.168.203:6828/43673 1 ==== command_reply(tid 1: -1 ) v1 ====
>>>> 8+0+0 (1943489909 0 0) 0x7fe238000f20 con 0x7fe240180f20
>>>> Error EPERM: problem getting command descriptions from pg.18.2
>>>> 2018-06-25 10:59:12.112168 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.203:6828/43673 conn(0x7fe240180f20 :-1 s=STATE_OPEN pgs=26353
>>>> cs=1
>>>> l=1).mark_down
>>>> 2018-06-25 10:59:12.112190 7fe244b28700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.203:6828/43673 conn(0x7fe240180f20 :-1 s=STATE_OPEN pgs=26353
>>>> cs=1
>>>> l=1)._stop
>>>> 2018-06-25 10:59:12.112337 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6840/32624 conn(0x7fe2240127b0 :-1 s=STATE_OPEN pgs=575947
>>>> cs=1
>>>> l=1).mark_down
>>>> 2018-06-25 10:59:12.112348 7fe244b28700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:6840/32624 conn(0x7fe2240127b0 :-1 s=STATE_OPEN pgs=575947
>>>> cs=1
>>>> l=1)._stop
>>>> 2018-06-25 10:59:12.112367 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1 s=STATE_OPEN pgs=472363 cs=1
>>>> l=1).mark_down
>>>> 2018-06-25 10:59:12.112372 7fe244b28700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.202:6789/0 conn(0x7fe240176dc0 :-1 s=STATE_OPEN pgs=472363 cs=1
>>>> l=1)._stop
>>>> 2018-06-25 10:59:12.112519 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections
>>>> 2018-06-25 10:59:12.112530 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections mark down 192.168.168.201:6840/32624 0x7fe2240127b0
>>>> 2018-06-25 10:59:12.112538 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections mark down 192.168.168.201:6789/0 0x7fe24017a420
>>>> 2018-06-25 10:59:12.112543 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections mark down 192.168.168.203:6828/43673 0x7fe240180f20
>>>> 2018-06-25 10:59:12.112549 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections mark down 192.168.168.202:6789/0 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.112554 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections delete 0x7fe2240127b0
>>>> 2018-06-25 10:59:12.112570 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections delete 0x7fe240176dc0
>>>> 2018-06-25 10:59:12.112577 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections delete 0x7fe24017a420
>>>> 2018-06-25 10:59:12.112582 7fe244b28700 5 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections delete 0x7fe240180f20
>>>> 2018-06-25 10:59:12.112701 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> shutdown_connections
>>>> 2018-06-25 10:59:12.112752 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> wait
>>>> complete.
>>>> 2018-06-25 10:59:12.112764 7fe244b28700 1 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:0/3046734987 conn(0x7fe240167220 :-1 s=STATE_NONE pgs=0
>>>> cs=0
>>>> l=0).mark_down
>>>> 2018-06-25 10:59:12.112770 7fe244b28700 2 -- 192.168.168.201:0/3046734987
>>>> >>
>>>> 192.168.168.201:0/3046734987 conn(0x7fe240167220 :-1 s=STATE_NONE pgs=0
>>>> cs=0
>>>> l=0)._stop
>>>>
>>>>
>>>> ----------
>>>>
>>>>
>>>> Thanks
>>>>
>>>> ----- Original Message -----
>>>>> From: "Brad Hubbard" <[email protected]>
>>>>> To: "Andrei Mikhailovsky" <[email protected]>
>>>>> Cc: "ceph-users" <[email protected]>
>>>>> Sent: Monday, 25 June, 2018 02:28:55
>>>>> Subject: Re: [ceph-users] fixing unrepairable inconsistent PG
>>>>
>>>>> Can you try the following?
>>>>>
>>>>> $ ceph --debug_ms 5 --debug_auth 20 pg 18.2 query
>>>>>
>>>>> On Fri, Jun 22, 2018 at 7:54 PM, Andrei Mikhailovsky <[email protected]>
>>>>> wrote:
>>>>>> Hi Brad,
>>>>>>
>>>>>> here is the output of the command (replaced the real auth key with
>>>>>> [KEY]):
>>>>>>
>>>>>>
>>>>>> ----------------
>>>>>>
>>>>>> 2018-06-22 10:47:27.659895 7f70ef9e6700 10 monclient:
>>>>>> build_initial_monmap
>>>>>> 2018-06-22 10:47:27.661995 7f70ef9e6700 10 monclient: init
>>>>>> 2018-06-22 10:47:27.662002 7f70ef9e6700 5 adding auth protocol: cephx
>>>>>> 2018-06-22 10:47:27.662004 7f70ef9e6700 10 monclient: auth_supported 2
>>>>>> method
>>>>>> cephx
>>>>>> 2018-06-22 10:47:27.662221 7f70ef9e6700 2 auth: KeyRing::load: loaded
>>>>>> key file
>>>>>> /etc/ceph/ceph.client.admin.keyring
>>>>>> 2018-06-22 10:47:27.662338 7f70ef9e6700 10 monclient: _reopen_session
>>>>>> rank -1
>>>>>> 2018-06-22 10:47:27.662425 7f70ef9e6700 10 monclient(hunting): picked
>>>>>> mon.noname-b con 0x7f70e8176c80 addr 192.168.168.202:6789/0
>>>>>> 2018-06-22 10:47:27.662484 7f70ef9e6700 10 monclient(hunting): picked
>>>>>> mon.noname-a con 0x7f70e817a2e0 addr 192.168.168.201:6789/0
>>>>>> 2018-06-22 10:47:27.662534 7f70ef9e6700 10 monclient(hunting):
>>>>>> _renew_subs
>>>>>> 2018-06-22 10:47:27.662544 7f70ef9e6700 10 monclient(hunting):
>>>>>> authenticate will
>>>>>> time out at 2018-06-22 10:52:27.662543
>>>>>> 2018-06-22 10:47:27.663831 7f70d77fe700 10 monclient(hunting):
>>>>>> handle_monmap
>>>>>> mon_map magic: 0 v1
>>>>>> 2018-06-22 10:47:27.663885 7f70d77fe700 10 monclient(hunting): got
>>>>>> monmap 20,
>>>>>> mon.noname-b is now rank -1
>>>>>> 2018-06-22 10:47:27.663889 7f70d77fe700 10 monclient(hunting): dump:
>>>>>> epoch 20
>>>>>> fsid 51e9f641-372e-44ec-92a4-b9fe55cbf9fe
>>>>>> last_changed 2018-06-16 23:14:48.936175
>>>>>> created 0.000000
>>>>>> 0: 192.168.168.201:6789/0 mon.arh-ibstorage1-ib
>>>>>> 1: 192.168.168.202:6789/0 mon.arh-ibstorage2-ib
>>>>>> 2: 192.168.168.203:6789/0 mon.arh-ibstorage3-ib
>>>>>>
>>>>>> 2018-06-22 10:47:27.664005 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664020 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664021 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664025 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service auth
>>>>>> 2018-06-22 10:47:27.664026 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 0 need 53
>>>>>> 2018-06-22 10:47:27.664032 7f70d77fe700 10 monclient(hunting): my
>>>>>> global_id is
>>>>>> 411322261
>>>>>> 2018-06-22 10:47:27.664035 7f70d77fe700 10 cephx client: handle_response
>>>>>> ret = 0
>>>>>> 2018-06-22 10:47:27.664046 7f70d77fe700 10 cephx client: got initial
>>>>>> server
>>>>>> challenge d66f2dffc2113d43
>>>>>> 2018-06-22 10:47:27.664049 7f70d77fe700 10 cephx client:
>>>>>> validate_tickets:
>>>>>> want=53 need=53 have=0
>>>>>>
>>>>>> 2018-06-22 10:47:27.664052 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664053 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664054 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664055 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service auth
>>>>>> 2018-06-22 10:47:27.664056 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 0 need 53
>>>>>> 2018-06-22 10:47:27.664057 7f70d77fe700 10 cephx client: want=53 need=53
>>>>>> have=0
>>>>>> 2018-06-22 10:47:27.664061 7f70d77fe700 10 cephx client: build_request
>>>>>> 2018-06-22 10:47:27.664145 7f70d77fe700 10 cephx client: get auth
>>>>>> session key:
>>>>>> client_challenge d4c95f637e641b55
>>>>>> 2018-06-22 10:47:27.664175 7f70d77fe700 10 monclient(hunting):
>>>>>> handle_monmap
>>>>>> mon_map magic: 0 v1
>>>>>> 2018-06-22 10:47:27.664208 7f70d77fe700 10 monclient(hunting): got
>>>>>> monmap 20,
>>>>>> mon.arh-ibstorage1-ib is now rank 0
>>>>>> 2018-06-22 10:47:27.664211 7f70d77fe700 10 monclient(hunting): dump:
>>>>>> epoch 20
>>>>>> fsid 51e9f641-372e-44ec-92a4-b9fe55cbf9fe
>>>>>> last_changed 2018-06-16 23:14:48.936175
>>>>>> created 0.000000
>>>>>> 0: 192.168.168.201:6789/0 mon.arh-ibstorage1-ib
>>>>>> 1: 192.168.168.202:6789/0 mon.arh-ibstorage2-ib
>>>>>> 2: 192.168.168.203:6789/0 mon.arh-ibstorage3-ib
>>>>>>
>>>>>> 2018-06-22 10:47:27.664241 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664244 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664245 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664246 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service auth
>>>>>> 2018-06-22 10:47:27.664247 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 0 need 53
>>>>>> 2018-06-22 10:47:27.664251 7f70d77fe700 10 monclient(hunting): my
>>>>>> global_id is
>>>>>> 411323061
>>>>>> 2018-06-22 10:47:27.664253 7f70d77fe700 10 cephx client: handle_response
>>>>>> ret = 0
>>>>>> 2018-06-22 10:47:27.664256 7f70d77fe700 10 cephx client: got initial
>>>>>> server
>>>>>> challenge d5d3c1e5bcf3c0b8
>>>>>> 2018-06-22 10:47:27.664258 7f70d77fe700 10 cephx client:
>>>>>> validate_tickets:
>>>>>> want=53 need=53 have=0
>>>>>> 2018-06-22 10:47:27.664260 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664261 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664262 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664263 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service auth
>>>>>> 2018-06-22 10:47:27.664264 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 0 need 53
>>>>>> 2018-06-22 10:47:27.664265 7f70d77fe700 10 cephx client: want=53 need=53
>>>>>> have=0
>>>>>> 2018-06-22 10:47:27.664268 7f70d77fe700 10 cephx client: build_request
>>>>>> 2018-06-22 10:47:27.664328 7f70d77fe700 10 cephx client: get auth
>>>>>> session key:
>>>>>> client_challenge d31821a6437d4974
>>>>>> 2018-06-22 10:47:27.664651 7f70d77fe700 10 cephx client: handle_response
>>>>>> ret = 0
>>>>>> 2018-06-22 10:47:27.664667 7f70d77fe700 10 cephx client:
>>>>>> get_auth_session_key
>>>>>> 2018-06-22 10:47:27.664673 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> got 1 keys
>>>>>> 2018-06-22 10:47:27.664676 7f70d77fe700 10 cephx: got key for service_id
>>>>>> auth
>>>>>> 2018-06-22 10:47:27.664766 7f70d77fe700 10 cephx: ticket.secret_id=3681
>>>>>> 2018-06-22 10:47:27.664774 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> service auth secret_id 3681 session_key [KEY] validity=43200.000000
>>>>>> 2018-06-22 10:47:27.664806 7f70d77fe700 10 cephx: ticket
>>>>>> expires=2018-06-22
>>>>>> 22:47:27.664805 renew_after=2018-06-22 19:47:27.664805
>>>>>> 2018-06-22 10:47:27.664825 7f70d77fe700 10 cephx client: want=53
>>>>>> need=53 have=0
>>>>>> 2018-06-22 10:47:27.664827 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664829 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664830 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664832 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 32 need 21
>>>>>> 2018-06-22 10:47:27.664836 7f70d77fe700 10 cephx client:
>>>>>> validate_tickets:
>>>>>> want=53 need=21 have=32
>>>>>> 2018-06-22 10:47:27.664837 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.664839 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.664840 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.664841 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 32 need 21
>>>>>> 2018-06-22 10:47:27.664842 7f70d77fe700 10 cephx client: want=53 need=21
>>>>>> have=32
>>>>>> 2018-06-22 10:47:27.664844 7f70d77fe700 10 cephx client: build_request
>>>>>> 2018-06-22 10:47:27.664846 7f70d77fe700 10 cephx client: get service
>>>>>> keys:
>>>>>> want=53 need=21 have=32
>>>>>> 2018-06-22 10:47:27.664928 7f70d77fe700 10 cephx client: handle_response
>>>>>> ret = 0
>>>>>> 2018-06-22 10:47:27.664933 7f70d77fe700 10 cephx client:
>>>>>> get_auth_session_key
>>>>>> 2018-06-22 10:47:27.664935 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> got 1 keys
>>>>>> 2018-06-22 10:47:27.664937 7f70d77fe700 10 cephx: got key for service_id
>>>>>> auth
>>>>>> 2018-06-22 10:47:27.664985 7f70d77fe700 10 cephx: ticket.secret_id=3681
>>>>>> 2018-06-22 10:47:27.664987 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> service auth secret_id 3681 session_key [KEY] validity=43200.000000
>>>>>> 2018-06-22 10:47:27.665009 7f70d77fe700 10 cephx: ticket
>>>>>> expires=2018-06-22
>>>>>> 22:47:27.665008 renew_after=2018-06-22 19:47:27.665008
>>>>>> 2018-06-22 10:47:27.665017 7f70d77fe700 10 cephx client: want=53
>>>>>> need=53 have=0
>>>>>> 2018-06-22 10:47:27.665019 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.665020 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.665024 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.665026 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 32 need 21
>>>>>> 2018-06-22 10:47:27.665029 7f70d77fe700 10 cephx client:
>>>>>> validate_tickets:
>>>>>> want=53 need=21 have=32
>>>>>> 2018-06-22 10:47:27.665031 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mon
>>>>>> 2018-06-22 10:47:27.665032 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service osd
>>>>>> 2018-06-22 10:47:27.665033 7f70d77fe700 10 cephx: set_have_need_key no
>>>>>> handler
>>>>>> for service mgr
>>>>>> 2018-06-22 10:47:27.665034 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 32 need 21
>>>>>> 2018-06-22 10:47:27.665035 7f70d77fe700 10 cephx client: want=53 need=21
>>>>>> have=32
>>>>>> 2018-06-22 10:47:27.665037 7f70d77fe700 10 cephx client: build_request
>>>>>> 2018-06-22 10:47:27.665039 7f70d77fe700 10 cephx client: get service
>>>>>> keys:
>>>>>> want=53 need=21 have=32
>>>>>> 2018-06-22 10:47:27.665354 7f70d77fe700 10 cephx client: handle_response
>>>>>> ret = 0
>>>>>> 2018-06-22 10:47:27.665365 7f70d77fe700 10 cephx client:
>>>>>> get_principal_session_key session_key [KEY]
>>>>>> 2018-06-22 10:47:27.665377 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> got 3 keys
>>>>>> 2018-06-22 10:47:27.665379 7f70d77fe700 10 cephx: got key for service_id
>>>>>> mon
>>>>>> 2018-06-22 10:47:27.665419 7f70d77fe700 10 cephx: ticket.secret_id=44133
>>>>>> 2018-06-22 10:47:27.665425 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> service mon secret_id 44133 session_key [KEY] validity=3600.000000
>>>>>> 2018-06-22 10:47:27.665437 7f70d77fe700 10 cephx: ticket
>>>>>> expires=2018-06-22
>>>>>> 11:47:27.665436 renew_after=2018-06-22 11:32:27.665436
>>>>>> 2018-06-22 10:47:27.665443 7f70d77fe700 10 cephx: got key for service_id
>>>>>> osd
>>>>>> 2018-06-22 10:47:27.665476 7f70d77fe700 10 cephx: ticket.secret_id=44133
>>>>>> 2018-06-22 10:47:27.665478 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> service osd secret_id 44133 session_key [KEY] validity=3600.000000
>>>>>> 2018-06-22 10:47:27.665497 7f70d77fe700 10 cephx: ticket
>>>>>> expires=2018-06-22
>>>>>> 11:47:27.665496 renew_after=2018-06-22 11:32:27.665496
>>>>>> 2018-06-22 10:47:27.665506 7f70d77fe700 10 cephx: got key for service_id
>>>>>> mgr
>>>>>> 2018-06-22 10:47:27.665539 7f70d77fe700 10 cephx: ticket.secret_id=132
>>>>>> 2018-06-22 10:47:27.665546 7f70d77fe700 10 cephx:
>>>>>> verify_service_ticket_reply
>>>>>> service mgr secret_id 132 session_key [KEY] validity=3600.000000
>>>>>> 2018-06-22 10:47:27.665564 7f70d77fe700 10 cephx: ticket
>>>>>> expires=2018-06-22
>>>>>> 11:47:27.665564 renew_after=2018-06-22 11:32:27.665564
>>>>>> 2018-06-22 10:47:27.665573 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 53 need 0
>>>>>> 2018-06-22 10:47:27.665602 7f70d77fe700 1 monclient: found
>>>>>> mon.arh-ibstorage2-ib
>>>>>> 2018-06-22 10:47:27.665617 7f70d77fe700 20 monclient: _un_backoff
>>>>>> reopen_interval_multipler now 1
>>>>>> 2018-06-22 10:47:27.665636 7f70d77fe700 10 monclient: _send_mon_message
>>>>>> to
>>>>>> mon.arh-ibstorage2-ib at 192.168.168.202:6789/0
>>>>>> 2018-06-22 10:47:27.665656 7f70d77fe700 10 cephx: validate_tickets want
>>>>>> 53 have
>>>>>> 53 need 0
>>>>>> 2018-06-22 10:47:27.665658 7f70d77fe700 20 cephx client: need_tickets:
>>>>>> want=53
>>>>>> have=53 need=0
>>>>>> 2018-06-22 10:47:27.665661 7f70d77fe700 20 monclient:
>>>>>> _check_auth_rotating not
>>>>>> needed by client.admin
>>>>>> 2018-06-22 10:47:27.665678 7f70ef9e6700 5 monclient: authenticate
>>>>>> success,
>>>>>> global_id 411322261
>>>>>> 2018-06-22 10:47:27.665694 7f70ef9e6700 10 monclient: _renew_subs
>>>>>> 2018-06-22 10:47:27.665698 7f70ef9e6700 10 monclient: _send_mon_message
>>>>>> to
>>>>>> mon.arh-ibstorage2-ib at 192.168.168.202:6789/0
>>>>>> 2018-06-22 10:47:27.665817 7f70ef9e6700 10 monclient: _renew_subs
>>>>>> 2018-06-22 10:47:27.665828 7f70ef9e6700 10 monclient: _send_mon_message
>>>>>> to
>>>>>> mon.arh-ibstorage2-ib at 192.168.168.202:6789/0
>>>>>> 2018-06-22 10:47:27.666069 7f70d77fe700 10 monclient: handle_monmap
>>>>>> mon_map
>>>>>> magic: 0 v1
>>>>>> 2018-06-22 10:47:27.666102 7f70d77fe700 10 monclient: got monmap 20,
>>>>>> mon.arh-ibstorage2-ib is now rank 1
>>>>>> 2018-06-22 10:47:27.666110 7f70d77fe700 10 monclient: dump:
>>>>>>
>>>>>> epoch 20
>>>>>> fsid 51e9f641-372e-44ec-92a4-b9fe55cbf9fe
>>>>>> last_changed 2018-06-16 23:14:48.936175
>>>>>> created 0.000000
>>>>>> 0: 192.168.168.201:6789/0 mon.arh-ibstorage1-ib
>>>>>> 1: 192.168.168.202:6789/0 mon.arh-ibstorage2-ib
>>>>>> 2: 192.168.168.203:6789/0 mon.arh-ibstorage3-ib
>>>>>>
>>>>>> 2018-06-22 10:47:27.666617 7f70eca43700 10 cephx client:
>>>>>> build_authorizer for
>>>>>> service mgr
>>>>>> 2018-06-22 10:47:27.667043 7f70eca43700 10 In get_auth_session_handler
>>>>>> for
>>>>>> protocol 2
>>>>>> 2018-06-22 10:47:27.678417 7f70eda45700 10 cephx client:
>>>>>> build_authorizer for
>>>>>> service osd
>>>>>> 2018-06-22 10:47:27.678914 7f70eda45700 10 In get_auth_session_handler
>>>>>> for
>>>>>> protocol 2
>>>>>> 2018-06-22 10:47:27.679003 7f70eda45700 10 _calc_signature seq 1
>>>>>> front_crc_ =
>>>>>> 2696387361 middle_crc = 0 data_crc = 0 sig = 929021353460216573
>>>>>> 2018-06-22 10:47:27.679026 7f70eda45700 20 Putting signature in client
>>>>>> message(seq # 1): sig = 929021353460216573
>>>>>> 2018-06-22 10:47:27.679520 7f70eda45700 10 _calc_signature seq 1
>>>>>> front_crc_ =
>>>>>> 1943489909 middle_crc = 0 data_crc = 0 sig = 10026640535487722288
>>>>>> Error EPERM: problem getting command descriptions from pg.18.2
>>>>>> 2018-06-22 10:47:27.681798 7f70ef9e6700 10 monclient: shutdown
>>>>>>
>>>>>>
>>>>>> -----------------
>>>>>>
>>>>>>
>>>>>> From what I can see the auth works:
>>>>>>
>>>>>> 2018-06-22 10:47:27.665678 7f70ef9e6700 5 monclient: authenticate
>>>>>> success,
>>>>>> global_id 411322261
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> ----- Original Message -----
>>>>>>> From: "Brad Hubbard" <[email protected]>
>>>>>>> To: "Andrei" <[email protected]>
>>>>>>> Cc: "ceph-users" <[email protected]>
>>>>>>> Sent: Friday, 22 June, 2018 02:05:51
>>>>>>> Subject: Re: [ceph-users] fixing unrepairable inconsistent PG
>>>>>>
>>>>>>> That seems like an authentication issue?
>>>>>>>
>>>>>>> Try running it like so...
>>>>>>>
>>>>>>> $ ceph --debug_monc 20 --debug_auth 20 pg 18.2 query
>>>>>>>
>>>>>>> On Thu, Jun 21, 2018 at 12:18 AM, Andrei Mikhailovsky
>>>>>>> <[email protected]> wrote:
>>>>>>>> Hi Brad,
>>>>>>>>
>>>>>>>> Yes, but it doesn't show much:
>>>>>>>>
>>>>>>>> ceph pg 18.2 query
>>>>>>>> Error EPERM: problem getting command descriptions from pg.18.2
>>>>>>>>
>>>>>>>> Cheers
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> ----- Original Message -----
>>>>>>>>> From: "Brad Hubbard" <[email protected]>
>>>>>>>>> To: "andrei" <[email protected]>
>>>>>>>>> Cc: "ceph-users" <[email protected]>
>>>>>>>>> Sent: Wednesday, 20 June, 2018 00:02:07
>>>>>>>>> Subject: Re: [ceph-users] fixing unrepairable inconsistent PG
>>>>>>>>
>>>>>>>>> Can you post the output of a pg query?
>>>>>>>>>
>>>>>>>>> On Tue, Jun 19, 2018 at 11:44 PM, Andrei Mikhailovsky
>>>>>>>>> <[email protected]> wrote:
>>>>>>>>>> A quick update on my issue. I have noticed that while I was trying
>>>>>>>>>> to move
>>>>>>>>>> the problem object on osds, the file attributes got lost on one of
>>>>>>>>>> the osds,
>>>>>>>>>> which is I guess why the error messages showed the no attribute bit.
>>>>>>>>>>
>>>>>>>>>> I then copied the attributes metadata to the problematic object and
>>>>>>>>>> restarted the osds in question. Following a pg repair I got a
>>>>>>>>>> different
>>>>>>>>>> error:
>>>>>>>>>>
>>>>>>>>>> 2018-06-19 13:51:05.846033 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 2 :
>>>>>>>>>> cluster [ERR] 18.2 shard 21: soid
>>>>>>>>>> 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> omap_digest 0x25e8a1da != omap_digest 0x21c7f871 from auth oi
>>>>>>>>>> 18:45f87722:::.dir.default.80018061.2:head(106137'603495
>>>>>>>>>> osd.21.0:41403910
>>>>>>>>>> dirty|omap|data_digest|omap_digest s 0 uv 603494 dd ffffffff od
>>>>>>>>>> 21c7f871
>>>>>>>>>> alloc_hint [0 0 0])
>>>>>>>>>> 2018-06-19 13:51:05.846042 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 3 :
>>>>>>>>>> cluster [ERR] 18.2 shard 28: soid
>>>>>>>>>> 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> omap_digest 0x25e8a1da != omap_digest 0x21c7f871 from auth oi
>>>>>>>>>> 18:45f87722:::.dir.default.80018061.2:head(106137'603495
>>>>>>>>>> osd.21.0:41403910
>>>>>>>>>> dirty|omap|data_digest|omap_digest s 0 uv 603494 dd ffffffff od
>>>>>>>>>> 21c7f871
>>>>>>>>>> alloc_hint [0 0 0])
>>>>>>>>>> 2018-06-19 13:51:05.846046 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 4 :
>>>>>>>>>> cluster [ERR] 18.2 soid 18:45f87722:::.dir.default.80018061.2:head:
>>>>>>>>>> failed
>>>>>>>>>> to pick suitable auth object
>>>>>>>>>> 2018-06-19 13:51:05.846118 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 5 :
>>>>>>>>>> cluster [ERR] repair 18.2 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> no '_'
>>>>>>>>>> attr
>>>>>>>>>> 2018-06-19 13:51:05.846129 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 6 :
>>>>>>>>>> cluster [ERR] repair 18.2 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> no
>>>>>>>>>> 'snapset' attr
>>>>>>>>>> 2018-06-19 13:51:09.810878 osd.21 osd.21 192.168.168.203:6828/24339
>>>>>>>>>> 7 :
>>>>>>>>>> cluster [ERR] 18.2 repair 4 errors, 0 fixed
>>>>>>>>>>
>>>>>>>>>> It mentions that there is an incorrect omap_digest . How do I go
>>>>>>>>>> about
>>>>>>>>>> fixing this?
>>>>>>>>>>
>>>>>>>>>> Cheers
>>>>>>>>>>
>>>>>>>>>> ________________________________
>>>>>>>>>>
>>>>>>>>>> From: "andrei" <[email protected]>
>>>>>>>>>> To: "ceph-users" <[email protected]>
>>>>>>>>>> Sent: Tuesday, 19 June, 2018 11:16:22
>>>>>>>>>> Subject: [ceph-users] fixing unrepairable inconsistent PG
>>>>>>>>>>
>>>>>>>>>> Hello everyone
>>>>>>>>>>
>>>>>>>>>> I am having trouble repairing one inconsistent and stubborn PG. I
>>>>>>>>>> get the
>>>>>>>>>> following error in ceph.log:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2018-06-19 11:00:00.000225 mon.arh-ibstorage1-ib mon.0
>>>>>>>>>> 192.168.168.201:6789/0 675 : cluster [ERR] overall HEALTH_ERR noout
>>>>>>>>>> flag(s)
>>>>>>>>>> set; 4 scrub errors; Possible data damage: 1 pg inconsistent;
>>>>>>>>>> application
>>>>>>>>>> not enabled on 4 pool(s)
>>>>>>>>>> 2018-06-19 11:09:24.586392 mon.arh-ibstorage1-ib mon.0
>>>>>>>>>> 192.168.168.201:6789/0 841 : cluster [ERR] Health check update:
>>>>>>>>>> Possible
>>>>>>>>>> data damage: 1 pg inconsistent, 1 pg repair (PG_DAMAGED)
>>>>>>>>>> 2018-06-19 11:09:27.139504 osd.21 osd.21 192.168.168.203:6828/4003 2
>>>>>>>>>> :
>>>>>>>>>> cluster [ERR] 18.2 soid 18:45f87722:::.dir.default.80018061.2:head:
>>>>>>>>>> failed
>>>>>>>>>> to pick suitable object info
>>>>>>>>>> 2018-06-19 11:09:27.139545 osd.21 osd.21 192.168.168.203:6828/4003 3
>>>>>>>>>> :
>>>>>>>>>> cluster [ERR] repair 18.2 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> no '_'
>>>>>>>>>> attr
>>>>>>>>>> 2018-06-19 11:09:27.139550 osd.21 osd.21 192.168.168.203:6828/4003 4
>>>>>>>>>> :
>>>>>>>>>> cluster [ERR] repair 18.2 18:45f87722:::.dir.default.80018061.2:head
>>>>>>>>>> no
>>>>>>>>>> 'snapset' attr
>>>>>>>>>>
>>>>>>>>>> 2018-06-19 11:09:35.484402 osd.21 osd.21 192.168.168.203:6828/4003 5
>>>>>>>>>> :
>>>>>>>>>> cluster [ERR] 18.2 repair 4 errors, 0 fixed
>>>>>>>>>> 2018-06-19 11:09:40.601657 mon.arh-ibstorage1-ib mon.0
>>>>>>>>>> 192.168.168.201:6789/0 844 : cluster [ERR] Health check update:
>>>>>>>>>> Possible
>>>>>>>>>> data damage: 1 pg inconsistent (PG_DAMAGED)
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> I have tried to follow a few instructions on the PG repair, including
>>>>>>>>>> removal of the 'broken' object .dir.default.80018061.2
>>>>>>>>>> from primary osd following by the pg repair. After that didn't
>>>>>>>>>> work, I've
>>>>>>>>>> done the same for the secondary osd. Still the same issue.
>>>>>>>>>>
>>>>>>>>>> Looking at the actual object on the file system, the file size is 0
>>>>>>>>>> for both
>>>>>>>>>> primary and secondary objects. The md5sum is the same too. The
>>>>>>>>>> broken PG
>>>>>>>>>> belongs to the radosgw bucket called .rgw.buckets.index
>>>>>>>>>>
>>>>>>>>>> What else can I try to get the thing fixed?
>>>>>>>>>>
>>>>>>>>>> Cheers
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> ceph-users mailing list
>>>>>>>>>> [email protected]
>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> _______________________________________________
>>>>>>>>>> ceph-users mailing list
>>>>>>>>>> [email protected]
>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Cheers,
>>>>>>>>> Brad
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Cheers,
>>>>>>> Brad
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Cheers,
>>>>> Brad
>>>>
>>>>
>>>
>>>
>>>
>>> --
>>> Cheers,
>>> Brad
>>
>>
>
>
>
> --
> Cheers,
> Brad
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com