Re: [PVE-User] Proxmox and non-proxmox hypervisors

2016-10-26 Thread Wolfgang Link
Proxmox VE use it own management stack and makes no use of libvirt, so
we can only manage other Proxmox VE hosts and there KVM and LXC containers.

If you like to use it you have to migrate all you kvm machines to
Proxmox VE, but this is no problem because only the config differs.

On 10/27/2016 12:25 AM, Leroy Tennison wrote:
> We have a number of kvm hypervisors currently in use, some on Ubuntu, a few 
> on openSUSE. We would like to provide the Proxmox web interface for users 
> rather than virt-manager (or the CLI). I understand we would need one Proxmox 
> hypervisor on Debian. If we did that, would the web interface be able to 
> manage the other hypervisors or is the web interface possible only because of 
> Proxmox software running on the hypervisor? If it is the latter, would it be 
> possible to install only the Proxmox software supporting the web interface on 
> the other hypervisors? Thanks. 
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> 

___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


[PVE-User] Proxmox and non-proxmox hypervisors

2016-10-26 Thread Leroy Tennison
We have a number of kvm hypervisors currently in use, some on Ubuntu, a few on 
openSUSE. We would like to provide the Proxmox web interface for users rather 
than virt-manager (or the CLI). I understand we would need one Proxmox 
hypervisor on Debian. If we did that, would the web interface be able to manage 
the other hypervisors or is the web interface possible only because of Proxmox 
software running on the hypervisor? If it is the latter, would it be possible 
to install only the Proxmox software supporting the web interface on the other 
hypervisors? Thanks. 
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread David Lawley
Watch nics, my hp units had broadcom nics and the drivers were buggy as 
heck.  Might be better now, but it was a few months of not knowing when 
they might crap out.. switched to intel and smooth sailing ever since




On 10/26/2016 7:47 AM, Info -- Asistar srl wrote:


Hello, I need some help

I want to buy a new HP server and mount proxmox 4.3 on it, because our 8
year old server is faulty.
I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the forum i
have seen that it can have issues with the RAID.
Can someone tell me if these issues are addressed or if it continues to be
an unsupported server? If so, can someone tell me an alternative machine
with similar specs that is compatible with proxmox?

Thank You in advance
Ivan Bolognani






___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user



___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread Silvestre Figueroa
in my case, in The Smart Array that I mentioned, the battery pack was
included.

Exact description in bare metal box:* Smart Array P440ar with 2GB FBWC
*




2016-10-26 13:11 GMT-03:00 Info -- Asistar srl :

> thanks for the reply.
> I buy the model DL380 Gen9 E5-2620V4 with 64Gb RAM
> I also need the battery for smart arrays?
>
>
>
> Ivan
>
>
> Il giorno 26/10/16, 14:37, "pve-user on behalf of Silvestre Figueroa"
>  silvestrefigue...@gmail.com> ha scritto:
>
> >Hi all!
> >
> >Im from Argentina, my english is not so good but I think I can explain me.
> >
> >I have 2 HP servers at PVE 3.2 Cluster, 1 HP DL380 GEN8 y 1 HP DL380 GEN9
> >(same description that you post).
> >
> >No problems with hardware RAID... dl380 gen9 has a Smart Array P440ar con
> >2GB FBWC RAID Controller.
> >
> >works out of the box like a punch!
> >
> >
> >
> >
> >
> >2016-10-26 9:13 GMT-03:00 Alex Bueno :
> >
> >> I have 3 hp servers, 2 dl360 gen9 and 1 dl160 gen 9 with p440 raid
> >> controller. 0 issues, work out of the box.
> >>
> >> El dc., 26 oct. 2016 , 13:47, Info -- Asistar srl  va
> >> escriure:
> >>
> >> >
> >> > Hello, I need some help
> >> >
> >> > I want to buy a new HP server and mount proxmox 4.3 on it, because
> >>our 8
> >> > year old server is faulty.
> >> > I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the
> >>forum i
> >> > have seen that it can have issues with the RAID.
> >> > Can someone tell me if these issues are addressed or if it continues
> >>to
> >> be
> >> > an unsupported server? If so, can someone tell me an alternative
> >>machine
> >> > with similar specs that is compatible with proxmox?
> >> >
> >> > Thank You in advance
> >> > Ivan Bolognani
> >> >
> >> >
> >> >
> >> >
> >> >
> >> >
> >> > ___
> >> > pve-user mailing list
> >> > pve-user@pve.proxmox.com
> >> > http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >> >
> >> ___
> >> pve-user mailing list
> >> pve-user@pve.proxmox.com
> >> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >>
> >
> >
> >
> >--
> >ING. SILVESTRE E. FIGUEROA
> >IT Consultant
> >Mobile: +542616462632
> >Skype: silvestrefigueroa
> >___
> >pve-user mailing list
> >pve-user@pve.proxmox.com
> >http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
>
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>



-- 
ING. SILVESTRE E. FIGUEROA
IT Consultant
Mobile: +542616462632
Skype: silvestrefigueroa
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread Michael Rasmussen
On Wed, 26 Oct 2016 18:11:03 +0200
Info  --  Asistar srl  wrote:

> thanks for the reply.
> I buy the model DL380 Gen9 E5-2620V4 with 64Gb RAM
Remember ECC RAM.

> I also need the battery for smart arrays?
> 
If you do not plan to use ZFS, which sort of nullifies the use of a
RAID controller, then yes. If you care for your data a BBU is
absolutely vital and will give you a noticeable performance boost.

-- 
Hilsen/Regards
Michael Rasmussen

Get my public GnuPG keys:
michael  rasmussen  cc
http://pgp.mit.edu:11371/pks/lookup?op=get=0xD3C9A00E
mir  datanom  net
http://pgp.mit.edu:11371/pks/lookup?op=get=0xE501F51C
mir  miras  org
http://pgp.mit.edu:11371/pks/lookup?op=get=0xE3E80917
--
/usr/games/fortune -es says:
Depends on how you define "always".  :-)
-- Larry Wall in <199710211647.jaa17...@wall.org>


pgpNirhAMyJcO.pgp
Description: OpenPGP digital signature
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread Info -- Asistar srl
thanks for the reply.
I buy the model DL380 Gen9 E5-2620V4 with 64Gb RAM
I also need the battery for smart arrays?



Ivan


Il giorno 26/10/16, 14:37, "pve-user on behalf of Silvestre Figueroa"
 ha scritto:

>Hi all!
>
>Im from Argentina, my english is not so good but I think I can explain me.
>
>I have 2 HP servers at PVE 3.2 Cluster, 1 HP DL380 GEN8 y 1 HP DL380 GEN9
>(same description that you post).
>
>No problems with hardware RAID... dl380 gen9 has a Smart Array P440ar con
>2GB FBWC RAID Controller.
>
>works out of the box like a punch!
>
>
>
>
>
>2016-10-26 9:13 GMT-03:00 Alex Bueno :
>
>> I have 3 hp servers, 2 dl360 gen9 and 1 dl160 gen 9 with p440 raid
>> controller. 0 issues, work out of the box.
>>
>> El dc., 26 oct. 2016 , 13:47, Info -- Asistar srl  va
>> escriure:
>>
>> >
>> > Hello, I need some help
>> >
>> > I want to buy a new HP server and mount proxmox 4.3 on it, because
>>our 8
>> > year old server is faulty.
>> > I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the
>>forum i
>> > have seen that it can have issues with the RAID.
>> > Can someone tell me if these issues are addressed or if it continues
>>to
>> be
>> > an unsupported server? If so, can someone tell me an alternative
>>machine
>> > with similar specs that is compatible with proxmox?
>> >
>> > Thank You in advance
>> > Ivan Bolognani
>> >
>> >
>> >
>> >
>> >
>> >
>> > ___
>> > pve-user mailing list
>> > pve-user@pve.proxmox.com
>> > http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>> >
>> ___
>> pve-user mailing list
>> pve-user@pve.proxmox.com
>> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>
>
>
>-- 
>ING. SILVESTRE E. FIGUEROA
>IT Consultant
>Mobile: +542616462632
>Skype: silvestrefigueroa
>___
>pve-user mailing list
>pve-user@pve.proxmox.com
>http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread Silvestre Figueroa
Hi all!

Im from Argentina, my english is not so good but I think I can explain me.

I have 2 HP servers at PVE 3.2 Cluster, 1 HP DL380 GEN8 y 1 HP DL380 GEN9
(same description that you post).

No problems with hardware RAID... dl380 gen9 has a Smart Array P440ar con
2GB FBWC RAID Controller.

works out of the box like a punch!





2016-10-26 9:13 GMT-03:00 Alex Bueno :

> I have 3 hp servers, 2 dl360 gen9 and 1 dl160 gen 9 with p440 raid
> controller. 0 issues, work out of the box.
>
> El dc., 26 oct. 2016 , 13:47, Info -- Asistar srl  va
> escriure:
>
> >
> > Hello, I need some help
> >
> > I want to buy a new HP server and mount proxmox 4.3 on it, because our 8
> > year old server is faulty.
> > I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the forum i
> > have seen that it can have issues with the RAID.
> > Can someone tell me if these issues are addressed or if it continues to
> be
> > an unsupported server? If so, can someone tell me an alternative machine
> > with similar specs that is compatible with proxmox?
> >
> > Thank You in advance
> > Ivan Bolognani
> >
> >
> >
> >
> >
> >
> > ___
> > pve-user mailing list
> > pve-user@pve.proxmox.com
> > http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
> >
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>



-- 
ING. SILVESTRE E. FIGUEROA
IT Consultant
Mobile: +542616462632
Skype: silvestrefigueroa
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Help for new server

2016-10-26 Thread Alex Bueno
I have 3 hp servers, 2 dl360 gen9 and 1 dl160 gen 9 with p440 raid
controller. 0 issues, work out of the box.

El dc., 26 oct. 2016 , 13:47, Info -- Asistar srl  va
escriure:

>
> Hello, I need some help
>
> I want to buy a new HP server and mount proxmox 4.3 on it, because our 8
> year old server is faulty.
> I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the forum i
> have seen that it can have issues with the RAID.
> Can someone tell me if these issues are addressed or if it continues to be
> an unsupported server? If so, can someone tell me an alternative machine
> with similar specs that is compatible with proxmox?
>
> Thank You in advance
> Ivan Bolognani
>
>
>
>
>
>
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


[PVE-User] Help for new server

2016-10-26 Thread Info -- Asistar srl

Hello, I need some help

I want to buy a new HP server and mount proxmox 4.3 on it, because our 8
year old server is faulty.
I thought of this: HP Proliant DL380 Gen9 E5-2620v4 but from the forum i
have seen that it can have issues with the RAID.
Can someone tell me if these issues are addressed or if it continues to be
an unsupported server? If so, can someone tell me an alternative machine
with similar specs that is compatible with proxmox?

Thank You in advance
Ivan Bolognani






___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Szabolcs F.
Hello again,

sorry for another followup. I just realised that 4 of the 12 cluster nodes
still have PVE Manager version 4.2 and the other 8 nodes have version 4.3.
Can this be the reason of all my troubles?

I'm in the process of updating these 4 nodes. These 4 nodes were installed
with the Proxmox install media, but the other 8 nodes were installed with
Debian 8 first. So the 4 outdated nodes didn't have the 'deb
http://download.proxmox.com/debian jessie pve-no-subscription' repo file.
Adding this repo made the 4.3 updates available.



On Wed, Oct 26, 2016 at 12:20 PM, Szabolcs F.  wrote:

> Hi Michael,
>
> I can change to LACP, sure. Would it be better than simple active-backup?
> I haven't got too much experience with LACP though.
>
> On Wed, Oct 26, 2016 at 11:55 AM, Michael Rasmussen  wrote:
>
>> Is it possible to switch to 802.3ad bond mode?
>>
>> On October 26, 2016 11:12:06 AM GMT+02:00, "Szabolcs F." <
>> subc...@gmail.com> wrote:
>>
>>> Hi Lutz,
>>>
>>> my bondXX files look like this: http://pastebin.com/GX8x3ZaN
>>> and my corosync.conf : http://pastebin.com/2ss0AAEr
>>>
>>> Mutlicast is enabled on my switches.
>>>
>>> The problem is I don't have a way to to replicate the problem, it seems to
>>> happen randomly, so I'm unsure how to do more tests. At the moment my
>>> cluster is working fine for about 16 hours. Any ideas forcing the issue?
>>>
>>> Thanks,
>>> Szabolcs
>>>
>>> On Wed, Oct 26, 2016 at 9:17 AM, Lutz Willek 
>>> wrote:
>>>
>>>  Am 24.10.2016 um 15:16 schrieb Szabolcs F.:

  Corosync has a lot of
> these in the /var/logs/daemon.log :
>  http://pastebin.com/ajhE8Rb9



  please carefully check your (node/switch/multicast) network configuration,
  and please paste your corosync configuration file and output of
  /proc/net/bonding/bondXX

  just a guess:

  * powerdown 1/3 - 1/2 of your nodes, adjust quorum (pvecm expect)
--> Problems still occours?

  * during "problem time"
--> omping is still ok?

  https://pve.proxmox.com/wiki/Troubleshooting_multicast,_quor
  um_and_cluster_issues


  Freundliche Grüße / Best Regards

   Lutz Willek

  --
 --
 creating IT solutions
  Lutz Willek science + computing ag
  Senior Systems Engineer Geschäftsstelle Berlin
  IT Services Berlin
   Friedrichstraße 187
  phone +49(0)30 2007697-21   10117 Berlin, Germany
  fax   +49(0)30 2007697-11   http://de.atos.net/sc

  S/MIME-Sicherheit:
  http://www.science-computing.de/cacert.crt
  http://www.science-computing.de/cacert-sha512.crt


 --

  pve-user mailing list
  pve-user@pve.proxmox.com
  http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>>
>>>
>>> --
>>>
>>> pve-user mailing list
>>> pve-user@pve.proxmox.com
>>> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>>
>>>
>> --
>> Sent from my Android phone with K-9 Mail. Please excuse my brevity.
>>
>
>
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Szabolcs F.
Hi Michael,

I can change to LACP, sure. Would it be better than simple active-backup? I
haven't got too much experience with LACP though.

On Wed, Oct 26, 2016 at 11:55 AM, Michael Rasmussen  wrote:

> Is it possible to switch to 802.3ad bond mode?
>
> On October 26, 2016 11:12:06 AM GMT+02:00, "Szabolcs F." <
> subc...@gmail.com> wrote:
>
>> Hi Lutz,
>>
>> my bondXX files look like this: http://pastebin.com/GX8x3ZaN
>> and my corosync.conf : http://pastebin.com/2ss0AAEr
>>
>> Mutlicast is enabled on my switches.
>>
>> The problem is I don't have a way to to replicate the problem, it seems to
>> happen randomly, so I'm unsure how to do more tests. At the moment my
>> cluster is working fine for about 16 hours. Any ideas forcing the issue?
>>
>> Thanks,
>> Szabolcs
>>
>> On Wed, Oct 26, 2016 at 9:17 AM, Lutz Willek 
>> wrote:
>>
>>  Am 24.10.2016 um 15:16 schrieb Szabolcs F.:
>>>
>>>  Corosync has a lot of
 these in the /var/logs/daemon.log :
  http://pastebin.com/ajhE8Rb9
>>>
>>>
>>>
>>>  please carefully check your (node/switch/multicast) network configuration,
>>>  and please paste your corosync configuration file and output of
>>>  /proc/net/bonding/bondXX
>>>
>>>  just a guess:
>>>
>>>  * powerdown 1/3 - 1/2 of your nodes, adjust quorum (pvecm expect)
>>>--> Problems still occours?
>>>
>>>  * during "problem time"
>>>--> omping is still ok?
>>>
>>>  https://pve.proxmox.com/wiki/Troubleshooting_multicast,_quor
>>>  um_and_cluster_issues
>>>
>>>
>>>  Freundliche Grüße / Best Regards
>>>
>>>   Lutz Willek
>>>
>>>  --
>>> --
>>> creating IT solutions
>>>  Lutz Willek science + computing ag
>>>  Senior Systems Engineer Geschäftsstelle Berlin
>>>  IT Services Berlin
>>>   Friedrichstraße 187
>>>  phone +49(0)30 2007697-21   10117 Berlin, Germany
>>>  fax   +49(0)30 2007697-11   http://de.atos.net/sc
>>>
>>>  S/MIME-Sicherheit:
>>>  http://www.science-computing.de/cacert.crt
>>>  http://www.science-computing.de/cacert-sha512.crt
>>>
>>>
>>> --
>>>
>>>  pve-user mailing list
>>>  pve-user@pve.proxmox.com
>>>  http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>>
>> --
>>
>> pve-user mailing list
>> pve-user@pve.proxmox.com
>> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>>
> --
> Sent from my Android phone with K-9 Mail. Please excuse my brevity.
>
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Michael Rasmussen
Is it possible to switch to 802.3ad bond mode?

On October 26, 2016 11:12:06 AM GMT+02:00, "Szabolcs F."  
wrote:
>Hi Lutz,
>
>my bondXX files look like this: http://pastebin.com/GX8x3ZaN
>and my corosync.conf : http://pastebin.com/2ss0AAEr
>
>Mutlicast is enabled on my switches.
>
>The problem is I don't have a way to to replicate the problem, it seems
>to
>happen randomly, so I'm unsure how to do more tests. At the moment my
>cluster is working fine for about 16 hours. Any ideas forcing the
>issue?
>
>Thanks,
>Szabolcs
>
>On Wed, Oct 26, 2016 at 9:17 AM, Lutz Willek
>
>wrote:
>
>> Am 24.10.2016 um 15:16 schrieb Szabolcs F.:
>>
>>> Corosync has a lot of these in the /var/logs/daemon.log :
>>> http://pastebin.com/ajhE8Rb9
>>>
>>
>> please carefully check your (node/switch/multicast) network
>configuration,
>> and please paste your corosync configuration file and output of
>> /proc/net/bonding/bondXX
>>
>> just a guess:
>>
>> * powerdown 1/3 - 1/2 of your nodes, adjust quorum (pvecm expect)
>>   --> Problems still occours?
>>
>> * during "problem time"
>>   --> omping is still ok?
>>
>> https://pve.proxmox.com/wiki/Troubleshooting_multicast,_quor
>> um_and_cluster_issues
>>
>>
>> Freundliche Grüße / Best Regards
>>
>>  Lutz Willek
>>
>> --
>> creating IT solutions
>> Lutz Willek science + computing ag
>> Senior Systems Engineer Geschäftsstelle Berlin
>> IT Services Berlin  Friedrichstraße 187
>> phone +49(0)30 2007697-21   10117 Berlin, Germany
>> fax   +49(0)30 2007697-11   http://de.atos.net/sc
>>
>> S/MIME-Sicherheit:
>> http://www.science-computing.de/cacert.crt
>> http://www.science-computing.de/cacert-sha512.crt
>>
>>
>> ___
>> pve-user mailing list
>> pve-user@pve.proxmox.com
>> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>>
>___
>pve-user mailing list
>pve-user@pve.proxmox.com
>http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user

-- 
Sent from my Android phone with K-9 Mail. Please excuse my brevity.
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Szabolcs F.
Hi Lutz,

my bondXX files look like this: http://pastebin.com/GX8x3ZaN
and my corosync.conf : http://pastebin.com/2ss0AAEr

Mutlicast is enabled on my switches.

The problem is I don't have a way to to replicate the problem, it seems to
happen randomly, so I'm unsure how to do more tests. At the moment my
cluster is working fine for about 16 hours. Any ideas forcing the issue?

Thanks,
Szabolcs

On Wed, Oct 26, 2016 at 9:17 AM, Lutz Willek 
wrote:

> Am 24.10.2016 um 15:16 schrieb Szabolcs F.:
>
>> Corosync has a lot of these in the /var/logs/daemon.log :
>> http://pastebin.com/ajhE8Rb9
>>
>
> please carefully check your (node/switch/multicast) network configuration,
> and please paste your corosync configuration file and output of
> /proc/net/bonding/bondXX
>
> just a guess:
>
> * powerdown 1/3 - 1/2 of your nodes, adjust quorum (pvecm expect)
>   --> Problems still occours?
>
> * during "problem time"
>   --> omping is still ok?
>
> https://pve.proxmox.com/wiki/Troubleshooting_multicast,_quor
> um_and_cluster_issues
>
>
> Freundliche Grüße / Best Regards
>
>  Lutz Willek
>
> --
> creating IT solutions
> Lutz Willek science + computing ag
> Senior Systems Engineer Geschäftsstelle Berlin
> IT Services Berlin  Friedrichstraße 187
> phone +49(0)30 2007697-21   10117 Berlin, Germany
> fax   +49(0)30 2007697-11   http://de.atos.net/sc
>
> S/MIME-Sicherheit:
> http://www.science-computing.de/cacert.crt
> http://www.science-computing.de/cacert-sha512.crt
>
>
> ___
> pve-user mailing list
> pve-user@pve.proxmox.com
> http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user
>
___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Lutz Willek

Am 24.10.2016 um 15:16 schrieb Szabolcs F.:

Corosync has a lot of these in the /var/logs/daemon.log :
http://pastebin.com/ajhE8Rb9


please carefully check your (node/switch/multicast) network 
configuration, and please paste your corosync configuration file and 
output of /proc/net/bonding/bondXX


just a guess:

* powerdown 1/3 - 1/2 of your nodes, adjust quorum (pvecm expect)
  --> Problems still occours?

* during "problem time"
  --> omping is still ok?

https://pve.proxmox.com/wiki/Troubleshooting_multicast,_quorum_and_cluster_issues


Freundliche Grüße / Best Regards

 Lutz Willek

--
creating IT solutions
Lutz Willek science + computing ag
Senior Systems Engineer Geschäftsstelle Berlin
IT Services Berlin  Friedrichstraße 187
phone +49(0)30 2007697-21   10117 Berlin, Germany
fax   +49(0)30 2007697-11   http://de.atos.net/sc

S/MIME-Sicherheit:
http://www.science-computing.de/cacert.crt
http://www.science-computing.de/cacert-sha512.crt

___
pve-user mailing list
pve-user@pve.proxmox.com
http://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user


Re: [PVE-User] Promox 4.3 cluster issue

2016-10-26 Thread Szabolcs F.
Hi Alwin,

thanks for the links. Do you mean VLAN tagging on trunk ports or completely
separated, untagged, dedicated ports?

ps: I forgot to ask about the jumbo frames. Should I enable them?

Thanks,
Szabolcs

On Tue, Oct 25, 2016 at 6:09 PM, Alwin Antreich 
wrote:

> Hi Szabolcs,
>
> On 10/25/2016 04:07 PM, Szabolcs F. wrote:
> > Hi Alwin,
> >
> > the Cisco 4948 switches don't have jumbo frames enabled. Global Ethernet
> > MTU is 1500 bytes. Port security is not enabled.
> >
> > When the issue happens the hosts are able to ping each other without any
> > packet loss.
> >
> > On Tue, Oct 25, 2016 at 3:02 PM, Alwin Antreich <
> sysadmin-...@cognitec.com>
> > wrote:
> >
> >> Hi Szabolcs,
> >>
> >> On 10/25/2016 12:24 PM, Szabolcs F. wrote:
> >>> Hi Alwin,
> >>>
> >>> bond0 is on two Cisco 4948 switches and bond1 is on two Cisco
> N3K-3064PQ
> >>> switches. They worked fine for about two months in this setup. But last
> >>> week (after I started to have these issues) I powered down one Cisco
> 4948
> >>> and one N3K-3064PQ switch (in both cases the designated backup switches
> >>> were powered down). This is to make sure all servers use the same
> switch
> >> as
> >>> their active link. After that I stopped the Proxmox cluster (all nodes)
> >> and
> >>> started them again, but the issue occurred again.
> >>
> >> Ok, so one thing less to check. How are your remaining switch
> configured,
> >> especially, where the pve cluster is on? Do
> >> they use jumbo frames? Or some network/port security?
> >>
> >>>
> >>> I've just added the 'bond_primary ethX' option to the interfaces file.
> >> I'll
> >>> reboot everything once again and see if it helps.
> >>
> >> That's only going to be used, when you have all links connected and want
> >> to prefer a link to be the primary, eg. 10GbE
> >> as primary and 1GbE as backup.
> >>
> >>>
> >>> syslog: http://pastebin.com/MsuCcNx8
> >>> dmesg: http://pastebin.com/xUPMKDJR
> >>> pveproxy (I can only see access.log for pveproxy, so this is the
> service
> >>> status): http://pastebin.com/gPPb4F3x
> >>
> >> I couldn't find anything unusual, but that doesn't mean there isn't.
> >>
> >>>
> >>> What other logs should I be reading?
> >>>
> >>> Thanks
> >>>
> >>> On Tue, Oct 25, 2016 at 11:23 AM, Alwin Antreich <
> >> sysadmin-...@cognitec.com>
> >>> wrote:
> >>>
>  Hi Szabolcs,
> 
>  On 10/25/2016 10:01 AM, Szabolcs F. wrote:
> > Hi Alwin,
> >
> > thanks for your hints.
> >
> >> On which interface is proxmox running on? Are these interfaces
> clogged
> > because, there is some heavy network IO going on?
> > I've got my two Intel Gbps network interfaces bonded together (bond0)
> >> as
> > active-backup and vmbr0 is bridged on this bond, then Proxmox is
> >> running
>  on
> > this interface. I.e. http://pastebin.com/WZKQ02Qu
> > All nodes are configured like this. There is no heavy IO on these
> > interfaces, because the storage network uses the separate 10Gbps
> fiber
> > Intel NICs (bond1).
> 
>  Is your bond working properly? Is the bond on the same switch or two
>  different?
> 
>  Usually I add the "bond_primary ethX" option to set the interface that
>  should be primarily used in active-backup
>  configuration - side note. :-)
> 
>  What are the logs on the server showing? You know, syslog, dmesg,
>  pveproxy, etc. ;-)
> 
> >
> >> Another guess, are all servers synchronizing with a NTP server and
> >> have
> > the correct time?
> > Yes, NTP is working properly, the firewall lets all NTP request go
>  through.
> >
> >
> > On Mon, Oct 24, 2016 at 5:19 PM, Alwin Antreich <
>  sysadmin-...@cognitec.com>
> > wrote:
> >
> >> Hello Szabolcs,
> >>
> >> On 10/24/2016 03:16 PM, Szabolcs F. wrote:
> >>> Hello,
> >>>
> >>> I've got a Proxmox VE 4.3 cluster of 12 nodes. All of them are Dell
>  C6220
> >>> sleds. Each has 2x Intel Xeon E5-2670 CPU and 64GB RAM. I've got
> two
> >>> separate networks: 1Gbps LAN (Cisco 4948 switch) and 10Gbps storage
> >> (Cisco
> >>> N3K-3064PQ fiber switch). The Dell nodes use the integrated Intel
> >> Gbit
> >>> adapters for LAN and Intel PCI-E 10Gbps cards for the fiber network
> >> (ixgbe
> >>> driver). The storage servers are separate, they run FreeNAS and
> >> export
> >> the
> >>> shares with NFS. My virtual machines (I've made about 40 of them so
>  far)
> >>> are KVM/QCOW2 and they are stored on the FreeNAS storage. So far so
>  good.
> >>> I've been using this environment as a test and was almost ready to
> >> push
> >>> into production.
> >> On which interface is proxmox running on? Are these interfaces
> clogged
> >> because, there is some heavy network IO going on?
> >>>
> >>> But I have a problem with the cluster. From time to time the
> pveproxy
> >>> service dies on