the exact same 3 servers have been down and everything has worked? do you run ceph mons on every server?
On Fri, 28 Sep 2018, 21:19 Gilberto Nunes, <[email protected]> wrote: > Everything is working couple of hours ago! > Due a power failure, 3 hosts are down, but I suppose with 3 host the > cluster garantee the quorum and allow ceph continues to work. > This already happen before and ceph works perfectly with 3 servers > > --- > Gilberto Nunes Ferreira > > (47) 3025-5907 > (47) 99676-7530 - Whatsapp / Telegram > > Skype: gilberto.nunes36 > > > > > > Em sex, 28 de set de 2018 às 17:13, Woods, Ken A (DNR) < > [email protected]> > escreveu: > > > So the times are massively different. Fix that. > > > > And corosync needs multicast , and that’s tested using omping, not ping. > > > > Go back through the initial set up documentation and read what is > required > > for the basic network configuration. > > > > If corosync and ceph are both not working, start there. > > > > > > > On Sep 28, 2018, at 12:08, Gilberto Nunes <[email protected]> > > wrote: > > > > > > pve-ceph01:~# ssh pve-ceph01 date > > > Fri Sep 28 17:06:34 -03 2018 > > > pve-ceph01:~# ssh pve-ceph02 date > > > Fri Sep 28 17:06:37 -03 2018 > > > pve-ceph01:~# ssh pve-ceph05 date > > > Fri Sep 28 17:06:39 -03 2018 > > > > > > pve-ceph01:~# ping -c 1 pve-ceph01 > > > PING pve-ceph01.cepam.com.br (10.10.10.100) 56(84) bytes of data. > > > 64 bytes from pve-ceph01.cepam.com.br (10.10.10.100): icmp_seq=1 > ttl=64 > > > time=0.020 ms > > > > > > --- pve-ceph01.cepam.com.br ping statistics --- > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms > > > rtt min/avg/max/mdev = 0.020/0.020/0.020/0.000 ms > > > pve-ceph01:~# ping -c 1 pve-ceph02 > > > once. > > > PING pve-ceph02.cepam.com.br (10.10.10.110) 56(84) bytes of data. > > > 64 bytes from pve-ceph02.cepam.com.br (10.10.10.110): icmp_seq=1 > ttl=64 > > > time=0.120 ms > > > > > > --- pve-ceph02.cepam.com.br ping statistics --- > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms > > > rtt min/avg/max/mdev = 0.120/0.120/0.120/0.000 ms > > > pve-ceph01:~# ping -c 1 pve-ceph05 > > > PING pve-ceph05.cepam.com.br (10.10.10.140) 56(84) bytes of data. > > > 64 bytes from pve-ceph05.cepam.com.br (10.10.10.140): icmp_seq=1 > ttl=64 > > > time=0.078 ms > > > > > > --- pve-ceph05.cepam.com.br ping statistics --- > > > 1 packets transmitted, 1 received, 0% packet loss, time 0ms > > > rtt min/avg/max/mdev = 0.078/0.078/0.078/0.000 ms > > > > > > I can communicate with other... > > > ceph command stuck > > > all ceph service it's appears to be running... > > > > > > > > > --- > > > Gilberto Nunes Ferreira > > > > > > (47) 3025-5907 > > > (47) 99676-7530 - Whatsapp / Telegram > > > > > > Skype: gilberto.nunes36 > > > > > > > > > > > > > > > > > > Em sex, 28 de set de 2018 às 17:02, Woods, Ken A (DNR) < > > [email protected]> > > > escreveu: > > > > > >> Network issues? > > >> Time issues? > > >> > > >> > > >>> On Sep 28, 2018, at 11:50, Gilberto Nunes < > [email protected]> > > >> wrote: > > >>> > > >>> Hi there > > >>> I have a 6 server Ceph Cluster maded with proxmox 5.2 > > >>> Suddenly, after power failure, I have only 3 servers UP, but even > with > > 3 > > >>> server, Ceph Cluster doesn't work. > > >>> pveceph status give me a timeout > > >>> pveceph status got timeout > > >>> > > >>> Any advice? > > >>> > > >>> > > >>> --- > > >>> Gilberto Nunes Ferreira > > >>> > > >>> (47) 3025-5907 > > >>> (47) 99676-7530 - Whatsapp / Telegram > > >>> > > >>> Skype: gilberto.nunes36 > > >>> _______________________________________________ > > >>> pve-user mailing list > > >>> [email protected] > > >>> > > >> > > > https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=9BUtASlCjgOxfXJ55imNPY2xKhHLVcP7E8Klu3Ddc0E&s=XjLyjlncigLbXaMeQ-15anc5Wjnneqwz-n1VWZ_zj_E&e= > > >> _______________________________________________ > > >> pve-user mailing list > > >> [email protected] > > >> > > > https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e= > > >> > > > _______________________________________________ > > > pve-user mailing list > > > [email protected] > > > > > > https://urldefense.proofpoint.com/v2/url?u=https-3A__pve.proxmox.com_cgi-2Dbin_mailman_listinfo_pve-2Duser&d=DwIGaQ&c=teXCf5DW4bHgLDM-H5_GmQ&r=THf3d3FQjCY5FQHo3goSprNAh9vsOWPUM7J0jwvvVwM&m=hPQWd4X3X9m5HCF_f-fftckCIT3Ix1hGUcOW__0w32A&s=rpDVR25KfEMwiCsK3aj5RrfERa2_hxyOHkuuxfgUYho&e= > > _______________________________________________ > > pve-user mailing list > > [email protected] > > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > > > _______________________________________________ > pve-user mailing list > [email protected] > https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user > _______________________________________________ pve-user mailing list [email protected] https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
