I always see this:

[mon01][DEBUG ]     "mons": [
[mon01][DEBUG ]       {
[mon01][DEBUG ]         "addr": "[fd91:462b:4243:47e::1:1]:6789/0",
[mon01][DEBUG ]         "name": "mon01",
[mon01][DEBUG ]         "public_addr": "[fd91:462b:4243:47e::1:1]:6789/0",
[mon01][DEBUG ]         "rank": 0
[mon01][DEBUG ]       },
[mon01][DEBUG ]       {
[mon01][DEBUG ]         "addr": "0.0.0.0:0/1",
[mon01][DEBUG ]         "name": "mon02",
[mon01][DEBUG ]         "public_addr": "0.0.0.0:0/1",
[mon01][DEBUG ]         "rank": 1
[mon01][DEBUG ]       },
[mon01][DEBUG ]       {
[mon01][DEBUG ]         "addr": "0.0.0.0:0/2",
[mon01][DEBUG ]         "name": "mon03",
[mon01][DEBUG ]         "public_addr": "0.0.0.0:0/2",
[mon01][DEBUG ]         "rank": 2
[mon01][DEBUG ]       }
[mon01][DEBUG ]     ]


DNS is working fine and the hostnames are also listed in /etc/hosts.
I already purged the mon but still the same problem.

- Kevin


2018-02-23 10:26 GMT+01:00 Kevin Olbrich <k...@sv01.de>:

> Hi!
>
> On a new cluster, I get the following error. All 3x mons are connected to
> the same switch and ping between them works (firewalls disabled).
> Mon-nodes are Ubuntu 16.04 LTS on Cep Luminous.
>
>
> [ceph_deploy.mon][ERROR ] Some monitors have still not reached quorum:
> [ceph_deploy.mon][ERROR ] mon03
> [ceph_deploy.mon][ERROR ] mon02
> [ceph_deploy.mon][ERROR ] mon01
>
>
> root@adminnode:~# cat ceph.conf
> [global]
> fsid = 2689defb-8715-47bb-8d78-e862089adf7a
> ms_bind_ipv6 = true
> mon_initial_members = mon01, mon02, mon03
> mon_host = [fd91:462b:4243:47e::1:1],[fd91:462b:4243:47e::1:2],[
> fd91:462b:4243:47e::1:3]
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> public network = fdd1:ecbd:731f:ee8e::/64
> cluster network = fd91:462b:4243:47e::/64
>
>
> root@mon01:~# ip a
> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN group
> default qlen 1000
>     link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
>     inet 127.0.0.1/8 scope host lo
>        valid_lft forever preferred_lft forever
>     inet6 ::1/128 scope host
>        valid_lft forever preferred_lft forever
> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc pfifo_fast
> state UP group default qlen 1000
>     link/ether b8:ae:ed:e9:b6:61 brd ff:ff:ff:ff:ff:ff
>     inet 172.17.1.1/16 brd 172.17.255.255 scope global eth0
>        valid_lft forever preferred_lft forever
>     inet6 fd91:462b:4243:47e::1:1/64 scope global
>        valid_lft forever preferred_lft forever
>     inet6 fe80::baae:edff:fee9:b661/64 scope link
>        valid_lft forever preferred_lft forever
> 3: wlan0: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN group
> default qlen 1000
>     link/ether 00:db:df:64:34:d5 brd ff:ff:ff:ff:ff:ff
> 4: eth0.22@eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 9000 qdisc noqueue
> state UP group default qlen 1000
>     link/ether b8:ae:ed:e9:b6:61 brd ff:ff:ff:ff:ff:ff
>     inet6 fdd1:ecbd:731f:ee8e::1:1/64 scope global
>        valid_lft forever preferred_lft forever
>     inet6 fe80::baae:edff:fee9:b661/64 scope link
>        valid_lft forever preferred_lft forever
>
>
> Don't mind wlan0, thats because this node is built from an Intel NUC.
>
> Any idea?
>
> Kind regards
> Kevin
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to