Hello,
the client is using this version:
root@ld3955:~# ceph versions
{
"mon": {
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 3
},
"mgr": {
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 3
},
"osd": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 161,
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 277
},
"mds": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 2
},
"overall": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 163,
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 283
}
}
The server is using this version:
root@ld5505:~# ceph versions
{
"mon": {
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 3
},
"mgr": {
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 3
},
"osd": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 161,
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 277
},
"mds": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 2
},
"overall": {
"ceph version 14.2.4 (65249672c6e6d843510e7e01f8a4b976dcac3db1)
nautilus (stable)": 163,
"ceph version 14.2.4-1-gd592e56
(d592e56e74d94c6a05b9240fcb0031868acefbab) nautilus (stable)": 283
}
}
There's no issue with timesync though.
Am 15.11.2019 um 13:22 schrieb Wido den Hollander:
>
> On 11/15/19 11:38 AM, Thomas Schneider wrote:
>> Hi,
>>
>> when I execute this command
>> rbd ls -l <pool-name>
>> to list all RBDs I get spamming errors:
>>
> Those errors are weird. Can you share the Ceph cluster version and the
> clients?
>
> $ ceph versions
>
> And then also use rpm/dpkg to check which version of Ceph runs on the
> client.
>
> Are you also sure the time is in sync on the client?
>
> Wido
>
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Sender did not set
>> CEPH_MSG_FOOTER_SIGNED.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Message signature
>> does not match contents.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Signature on message:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 sig: 0
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Locally calculated
>> signature:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1
>> sig_check:6229148783016323662
>> 2019-11-15 11:29:19.428 7fd852678700 0 Signature failed.
>> 2019-11-15 11:29:19.428 7fd852678700 0 --1- 10.97.206.91:0/2841811017
>>>> v1:10.97.206.97:6884/265976 conn(0x7fd834090770 0x7fd83408c190 :-1
>> s=READ_FOOTER_AND_DISPATCH pgs=42068 cs=1 l=1).handle_message_footer
>> Signature check failed
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Sender did not set
>> CEPH_MSG_FOOTER_SIGNED.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Message signature
>> does not match contents.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Signature on message:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 sig: 0
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Locally calculated
>> signature:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1
>> sig_check:4639667068846516939
>> 2019-11-15 11:29:19.428 7fd852678700 0 Signature failed.
>> 2019-11-15 11:29:19.428 7fd852678700 0 --1- 10.97.206.91:0/2841811017
>>>> v1:10.97.206.97:6884/265976 conn(0x7fd83408c990 0x7fd83408b160 :-1
>> s=READ_FOOTER_AND_DISPATCH pgs=42069 cs=1 l=1).handle_message_footer
>> Signature check failed
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Sender did not set
>> CEPH_MSG_FOOTER_SIGNED.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 Message signature
>> does not match contents.
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Signature on message:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1 sig: 0
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1Locally calculated
>> signature:
>> 2019-11-15 11:29:19.428 7fd852678700 0 SIGN: MSG 1
>> sig_check:12754808375040063976
>> 2019-11-15 11:29:19.428 7fd852678700 0 Signature failed.
>> 2019-11-15 11:29:19.428 7fd852678700 0 --1- 10.97.206.91:0/2841811017
>>>> v1:10.97.206.97:6884/265976 conn(0x7fd834090770 0x7fd83408c190 :-1
>> s=READ_FOOTER_AND_DISPATCH pgs=42070 cs=1 l=1).handle_message_footer
>> Signature check failed
>>
>> I allready stopped and started all MON services w/o success.
>>
>> This is the output of config dump:
>> root@ld3955:~# ceph config-key dump | grep config/
>> "config/mgr/mgr/balancer/active": "true",
>> "config/mgr/mgr/balancer/mode": "upmap",
>> "config/mgr/mgr/balancer/pool_ids": "11,59,60,61",
>> "config/mgr/mgr/balancer/upmap_max_iterations": "2",
>> "config/mgr/mgr/dashboard/url_prefix": "dashboard",
>> "config/mgr/mgr/devicehealth/enable_monitoring": "false",
>> "config/mgr/target_max_misplaced_ratio": "0.010000",
>>
>> And this is the config of an OSD showing that cephx authentication is
>> activated:
>> root@ld3955:~# ceph config show osd.0
>> NAME VALUE
>> SOURCE OVERRIDES IGNORES
>> auth_client_required cephx file
>> auth_cluster_required cephx file
>> auth_service_required cephx file
>> bluestore_block_db_size 10737418240 file
>> cephx_cluster_require_signatures false file
>> cephx_require_signatures false file
>> cephx_sign_messages false file
>> cluster_network 192.168.1.0/27 file
>> daemonize false
>> override
>> debug_ms 0/0 file
>> keyring $osd_data/keyring
>> default
>> leveldb_log
>> default
>> mon_allow_pool_delete true file
>> mon_host 10.97.206.93 10.97.206.94 10.97.206.95 file
>> mon_osd_full_ratio 0.850000 file
>> mon_osd_nearfull_ratio 0.750000 file
>> osd_crush_update_on_start false file
>> osd_deep_scrub_interval 1209600.000000 file
>> osd_journal_size 1024 file
>> osd_max_backfills 2 file
>> osd_op_queue wpq file
>> osd_op_queue_cut_off high file
>> osd_pool_default_min_size 2 file
>> osd_pool_default_size 3 file
>> osd_scrub_begin_hour 21 file
>> osd_scrub_end_hour 8 file
>> osd_scrub_sleep 0.100000 file
>> public_network 10.97.206.0/24 file
>> rbd_default_features 61
>> default
>> setgroup ceph
>> cmdline
>> setuser ceph
>> cmdline
>>
>> How can I fix this error?
>>
>> THX
>>
>> _______________________________________________
>> ceph-users mailing list -- [email protected]
>> To unsubscribe send an email to [email protected]
>>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]