Thank you again for all the help. I changed the mon_initial_members in 
/etc/ceph/es-c1.conf to only include the current (bootstrapped) monitor and the 
command succeeds.

As for the custom cluster name. Is that something people would recommend or not?
________________________________
From: Tuffli, Chuck <[email protected]>
Sent: Thursday, May 13, 2021 9:50 AM
To: Eugen Block <[email protected]>
Cc: [email protected] <[email protected]>
Subject: [ceph-users] Re: monitor connection error

OK, that change makes the error message go away, but the ceph command then 
seemingly hangs:

[centos@cnode-01 ~]$ time sudo ceph --cluster es-c1 --status
^CCluster connection aborted

real    6m33.555s
user    0m0.140s
sys     0m0.041s
________________________________
From: Eugen Block <[email protected]>
Sent: Thursday, May 13, 2021 9:15 AM
To: Tuffli, Chuck <[email protected]>
Cc: [email protected] <[email protected]>
Subject: Re: [ceph-users] Re: monitor connection error

Sorry, hit the send the send button too early. Could you try and
rename the keyring file like this?

es-c1.client.admin.keyring:

Zitat von Eugen Block <[email protected]>:

> I think you need to rename the admin keyring file accordingly to the
> cluster name, too.
>
> Zitat von "Tuffli, Chuck" <[email protected]>:
>
>> Thank you for trying to reproduce my issue. I think I did that step:
>>
>> [centos@cnode-01 ~]$ ls -l /etc/ceph/ceph.client.admin.keyring
>> -rw-------. 1 root root 151 May 10 23:20 /etc/ceph/ceph.client.admin.keyring
>> [centos@cnode-01 ~]$ sudo cat /etc/ceph/ceph.client.admin.keyring
>> [client.admin]
>>        key = AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q==
>>        caps mds = "allow *"
>>        caps mgr = "allow *"
>>        caps mon = "allow *"
>>        caps osd = "allow *"
>> [centos@cnode-01 ~]$ sudo grep -r AQC6v5 /etc/ceph /tmp
>> /etc/ceph/ceph.client.admin.keyring:    key =
>> AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q==
>> /tmp/ceph.mon.keyring:  key = AQC6v5lgi0JBAhAAJ9Duj11SufIKidydIgO82Q==
>> [centos@cnode-01 ~]$
>>
>> ________________________________
>> From: Eugen Block <[email protected]>
>> Sent: Thursday, May 13, 2021 12:37 AM
>> To: Tuffli, Chuck <[email protected]>
>> Cc: [email protected] <[email protected]>
>> Subject: Re: [ceph-users] Re: monitor connection error
>>
>> I could reproduce your issue, it seems you're missing a keyring file
>> in /etc/ceph. Did you miss step 9 from the manual deployment guide?
>>
>> sudo ceph-authtool --create-keyring
>> /etc/ceph/ceph.client.admin.keyring --gen-key -n client.admin --cap
>> mon 'allow *' --cap osd 'allow *' --cap mds 'allow *' --cap mgr 'allow
>> *'
>>
>>
>>
>>
>> Zitat von "Tuffli, Chuck" <[email protected]>:
>>
>>>> -----Original Message-----
>>>> From: Eugen Block [mailto:[email protected]]
>>>> Sent: Tuesday, May 11, 2021 11:39 PM
>>>> To: [email protected]
>>>> Subject: [ceph-users] Re: monitor connection error
>>>>
>>>> Hi,
>>>>
>>>>> What is this error trying to tell me? TIA
>>>>
>>>> it tells you that the cluster is not reachable to the client, this
>>>> can have various
>>>> reasons.
>>>>
>>>> Can you show the output of your conf file?
>>>>
>>>> cat /etc/ceph/es-c1.conf
>>>
>>> [centos@cnode-01 ~]$ cat /etc/ceph/es-c1.conf
>>> [global]
>>> fsid = 3c5da069-2a03-4a5a-8396-53776286c858
>>> mon_initial_members = cnode-01,cnode-02,cnode-03
>>> mon_host = 192.168.122.39
>>> public_network = 192.168.122.0/24
>>> auth_cluster_required = cephx
>>> auth_service_required = cephx
>>> auth_client_required = cephx
>>> osd_journal_size = 1024
>>> osd_pool_default_size = 3
>>> osd_pool_default_min_size = 2
>>> osd_pool_default_pg_num = 333
>>> osd_pool_default_pgp_num = 333
>>> osd_crush_chooseleaf_type = 1
>>> [centos@cnode-01 ~]$
>>>
>>>> Is the monitor service up running? I take it you don't use cephadm
>>>> yet so it's not
>>>> a containerized environment?
>>>
>>> Correct, this is bare metal and not a containerized environment. And
>>> I believe it is running:
>>> [centos@cnode-01 ~]$ sudo systemctl --all | grep ceph
>>>  ceph-crash.service
>>>                      loaded    active   running   Ceph crash dump
>>> collector
>>>  [email protected]
>>>                      loaded    active   running   Ceph cluster
>>> monitor daemon
>>>  system-ceph\x2dmon.slice
>>>                      loaded    active   active
>>> system-ceph\x2dmon.slice
>>>  ceph-mon.target
>>>                      loaded    active   active    ceph target
>>> allowing to start/stop all [email protected] instances at once
>>>  ceph.target
>>>                      loaded    active   active    ceph target
>>> allowing to start/stop all ceph*@.service instances at once
>>> [centos@cnode-01 ~]$
>>>
>>>> Regards,
>>>> Eugen
>>>>
>>>>
>>>> Zitat von "Tuffli, Chuck" <[email protected]>:
>>>>
>>>>> Hi
>>>>>
>>>>> I'm new to ceph and have been following the Manual Deployment document
>>>>> [1]. The process seems to work correctly until step 18 ("Verify that
>>>>> the monitor is running"):
>>>>>
>>>>> [centos@cnode-01 ~]$ uname -a
>>>>> Linux cnode-01 3.10.0-693.5.2.el7.x86_64 #1 SMP Fri Oct 20 20:32:50
>>>>> UTC 2017 x86_64 x86_64 x86_64 GNU/Linux
>>>>> [centos@cnode-01 ~]$ ceph -v
>>>>> ceph version 15.2.11 (e3523634d9c2227df9af89a4eac33d16738c49cb)
>>>>> octopus (stable)
>>>>> [centos@cnode-01 ~]$ sudo ceph --cluster es-c1 -s [errno 2] RADOS
>>>>> object not found (error connecting to the cluster)
>>>>> [centos@cnode-01 ~]$
>>>>>
>>>>> What is this error trying to tell me? TIA
>>>>>
>>>>> [1]
>>>>> INVALID URI REMOVED
>>>>> nual-deployment/__;!!NpxR!1-v_Ql6E-l3P_E8DvIfk_YtknPrVFeZ5sFaPHLlsJVY8
>>>>> PmzP7kySRbr1rYqbFiZ1$
>>>> _______________________________________________
>>>>> ceph-users mailing list -- [email protected] To unsubscribe send an
>>>>> email to [email protected]
>>>>
>>>>
>>>> _______________________________________________
>>>> ceph-users mailing list -- [email protected] To unsubscribe send
>>>> an email to
>>>> [email protected]



_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to