Re: [Ganglia-general] XSS vulnerabilities in Ganglia web

2015-01-13 Thread Cristovao Jose Domingues Cordeiro
Hi,

sorry for the late reply.
I've finally tried to install the latest web version. It seems indeed that the 
below vulnerabilities have been dealt with.
The only issue which is now being reported after the scan is:
NVT: Missing httpOnly Cookie Attribute
Summary:
The application is missing the 'httpOnly' cookie attribute


Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/R-018
CERN

From: Vladimir Vuksan [vli...@veus.hr]
Sent: 07 November 2014 23:06
To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web

I recall trying this out on 3.6.2 and I couldn't reproduce it so if you could 
run this against 3.6.2 and see if you can reproduce it that would be really 
helpful.

Vladimir

On 11/07/2014 04:50 PM, Cristovao Jose Domingues Cordeiro wrote:
It is implemented on 3.5.12
Is this fixed on the latest version?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/R-018
CERN

From: Vladimir Vuksan [vli...@veus.hrmailto:vli...@veus.hr]
Sent: 07 November 2014 22:31
To: Cristovao Jose Domingues Cordeiro; 
ganglia-general@lists.sourceforge.netmailto:ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web

Hi Cristovao,

what Ganglia Web version was tested ? Is this against latest e.g. 3.6.2 ?

Thanks,

Vladimir

On 04/11/2014 03:35 AM, Cristovao Jose Domingues Cordeiro wrote:
Hi all,

recently I've updated my Ganglia web frontend to the latest version (so I could 
perform HTTP queries) and when I issued the security check with skipfish I got 
these:

Vulnerabilities found: 33

· Severity: 4, Type: File inclusion
..
..
· Severity: 4, Type: Query injection vector
..
..
· Severity: 4, Type: Shell injection vector
..
..
· Severity: 4, Type: Server-side XML injection vector
..
..
· Severity: 3, Type: Directory traversal / file inclusion possible
··
··
· Severity: 3, Type: XSS vector in document body
..
..


Now, these are too many vulnerabilities, but I don't know if they can affect 
the backend of if they just affect the frontend. Do you know?

The XSS vulnerability must be fixed for sure. I've seen some references to this 
in your release notes (e.g. 
http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html
 ) but in fact there if no difference between these last releases and the ones 
before that announcement.

Is there a workaround for this? I can not open this Ganglia machine to the 
outside world if I don't have this fixed.


--
New Year. New Location. New Benefits. New Data Center in Ashburn, VA.
GigeNET is offering a free month of service with a new server in Ashburn.
Choose from 2 high performing configs, both with 100TB of bandwidth.
Higher redundancy.Lower latency.Increased capacity.Completely compliant.
http://p.sf.net/sfu/gigenet___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] XSS vulnerabilities in Ganglia web

2014-11-07 Thread Cristovao Jose Domingues Cordeiro
It is implemented on 3.5.12
Is this fixed on the latest version?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/R-018
CERN

From: Vladimir Vuksan [vli...@veus.hr]
Sent: 07 November 2014 22:31
To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web

Hi Cristovao,

what Ganglia Web version was tested ? Is this against latest e.g. 3.6.2 ?

Thanks,

Vladimir

On 04/11/2014 03:35 AM, Cristovao Jose Domingues Cordeiro wrote:
Hi all,

recently I've updated my Ganglia web frontend to the latest version (so I could 
perform HTTP queries) and when I issued the security check with skipfish I got 
these:

Vulnerabilities found: 33

· Severity: 4, Type: File inclusion
..
..
· Severity: 4, Type: Query injection vector
..
..
· Severity: 4, Type: Shell injection vector
..
..
· Severity: 4, Type: Server-side XML injection vector
..
..
· Severity: 3, Type: Directory traversal / file inclusion possible
··
··
· Severity: 3, Type: XSS vector in document body
..
..


Now, these are too many vulnerabilities, but I don't know if they can affect 
the backend of if they just affect the frontend. Do you know?

The XSS vulnerability must be fixed for sure. I've seen some references to this 
in your release notes (e.g. 
http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html
 ) but in fact there if no difference between these last releases and the ones 
before that announcement.

Is there a workaround for this? I can not open this Ganglia machine to the 
outside world if I don't have this fixed.

--
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] XSS Vulnerability

2014-11-03 Thread Cristovao Jose Domingues Cordeiro
Hi all,

just wondering if there are any plans to take care of this issue in next 
releases of Ganglia web?
Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/R-018
CERN
--
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk

2014-08-28 Thread Cristovao Jose Domingues Cordeiro
Hi again,

can someone point to the code where action are taken upon the gmon 
configuration 'host_tmax' and 'host_dmax'? Because that's where the RRD 
physical hosts should be deleted in my opinion

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Cristovao Jose Domingues Cordeiro
Sent: 26 August 2014 09:22
To: Adam Compton; Ganglia List ‎[ganglia-general@lists.sourceforge.net]‎
Subject: RE: [Ganglia-general] Down hosts remain on disk and flood RAM disk

Hi Adam,

Thanks. Yeah that might be the ultimate solution. But I guess this is a known 
problem Ganglia developers had to deal with before right? It's an unavoidable 
scenario that will break gmetad. There must be some kind of protection...


From: Adam Compton [acomp...@quantcast.com]
Sent: 26 August 2014 00:59
To: Cristovao Jose Domingues Cordeiro; Ganglia List 
‎[ganglia-general@lists.sourceforge.net]‎
Subject: Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk

On 8/25/14 6:53 AM, Cristovao Jose Domingues Cordeiro wrote:
Hi all,

I have a Ganglia setup running with a 4Gb ramdisk.

I'm monitoring several  opportunistic clusters with short-lifetimed VM's which 
are renewed a lot, causing a huge number of down hosts in Ganglia.

From a web interface point of view this is solved with the host_dmax option, 
and it works fine, but the problem is that even though the VM's are down and 
out of the web interface, they still remain on the ramdisk, causing it to 
flood eventually, and consequently breaking gmetad.

Any solutions?

We have a cronjob that finds and purges RRDs that have not been modified in N 
days (currently 90 in most cases). I would also be interested in other peoples' 
solutions to this problem.

- Adam

--
Slashdot TV.  
Video for Nerds.  Stuff that matters.
http://tv.slashdot.org/___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk

2014-08-26 Thread Cristovao Jose Domingues Cordeiro
Hi Adam,

Thanks. Yeah that might be the ultimate solution. But I guess this is a known 
problem Ganglia developers had to deal with before right? It's an unavoidable 
scenario that will break gmetad. There must be some kind of protection...


From: Adam Compton [acomp...@quantcast.com]
Sent: 26 August 2014 00:59
To: Cristovao Jose Domingues Cordeiro; Ganglia List 
‎[ganglia-general@lists.sourceforge.net]‎
Subject: Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk

On 8/25/14 6:53 AM, Cristovao Jose Domingues Cordeiro wrote:
Hi all,

I have a Ganglia setup running with a 4Gb ramdisk.

I'm monitoring several  opportunistic clusters with short-lifetimed VM's which 
are renewed a lot, causing a huge number of down hosts in Ganglia.

From a web interface point of view this is solved with the host_dmax option, 
and it works fine, but the problem is that even though the VM's are down and 
out of the web interface, they still remain on the ramdisk, causing it to 
flood eventually, and consequently breaking gmetad.

Any solutions?

We have a cronjob that finds and purges RRDs that have not been modified in N 
days (currently 90 in most cases). I would also be interested in other peoples' 
solutions to this problem.

- Adam

--
Slashdot TV.  
Video for Nerds.  Stuff that matters.
http://tv.slashdot.org/___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] Huge metrics' size being reported to gmetad

2014-06-24 Thread Cristovao Jose Domingues Cordeiro
Hi,

I have a grid configuration, with several clusters. I am also using RAM disk 
for I/O optimization (4GB).

I've been noticing that sometimes, gmetad breaks, complaining about lack of 
space in this tmpfs partition.

I checked and I saw that for some reason, some clusters, have hosts which 
occupy 3MB, 6MB and even sometimes 150MB All together  makes the cluster 
occupy 2GB and consequently occupy half of the ramdisk space.

We would normally expect these host metrics to have +/- 336k right?

Has anyone experienced this?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/1-010
CERN
--
Open source business process management suite built on Java and Eclipse
Turn processes into business applications with Bonita BPM Community Edition
Quickly connect people, data, and systems into organized workflows
Winner of BOSSIE, CODIE, OW2 and Gartner awards
http://p.sf.net/sfu/Bonitasoft___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] Random blank timeslots in graphs

2014-05-19 Thread Cristovao Jose Domingues Cordeiro
Hi,

this is happening in two completely different (but with the same deployment 
method) Ganglia headnodes.

I'm monitoring about 500 VM's (on each headnode), separated by clusters of 
different sizes. From time to time, the summary graphs over some cluster stop 
reporting, showing zero activity, and then suddenly after a while they come 
back up again.

This is very undesirable since I end up with several white holes per day on 
each cluster.

The information I can give you so far is the following:


  *   The attached image shows what happens
  *   I have a master-slave type of configuration, where the collector gmonds 
are sitting in the same machine (the headnode) as gmetad and ganglia-web, and 
where all the gmond nodes are reporting their metrics through unicast to the 
headnode.
  *   I have the latest Ganglia versions running (both core and web)
  *   All VM's are based on SL6
  *   When I look at /var/log/messages I see a lot of this:
 *   May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular 
data for pkts_out#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_user#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_system#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_idle#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_nice#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_aidle#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_wio#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_steal#012
May 19 16:14:37 gangliamon gmond[22304]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_user#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_system#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_idle#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_nice#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_aidle#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_wio#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_steal#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_free#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_shared#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_buffers#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_cached#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
swap_free#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
bytes_out#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
bytes_in#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
pkts_in#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
pkts_out#012
May 19 16:14:40 gangliamon gmond[10560]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:42 gangliamon gmond[22304]: Error 1 sending the modular data for 
disk_free#012



Which I understand is a known unsolved issue, by looking at other discussions 
like 
https://www.mail-archive.com/ganglia-general@lists.sourceforge.net/msg06602.html
 .


Does anyone know how to solve this?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Accelerate Dev Cycles with Automated Cross-Browser Testing - For FREE
Instantly run your Selenium tests across 300+ browser/OS combos.
Get unparalleled scalability from the best Selenium testing platform available
Simple to use. Nothing to install. Get started now for free.
http://p.sf.net/sfu/SauceLabs___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] Random blank timeslots in graphs

2014-05-19 Thread Cristovao Jose Domingues Cordeiro
Hi,

I am using Ganglia Web Frontend version 3.5.12 and Ganglia Web Backend (gmetad) 
version 3.6.0. The Gmond version on the nodes is not consistent, since they are 
being set by different users, on different environments. But I believe their 
version is not below 3.1.7.

No, I am not using RRDCached...all of my Ganglia configurations are the default 
ones. I'll try to set that up.

Since you believe it is a scaling problem, should I try to store the DB in 
ramdisk?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Vladimir Vuksan [vli...@veus.hr]
Sent: 19 May 2014 16:37
To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] Random blank timeslots in graphs

Error 1 sending messages are a red herring.

If you are seeing gaps it's most likely that storage system is not keeping up. 
What version of ganglia are you using and are you using rrdcached ?

Vladimir

On 05/19/2014 10:20 AM, Cristovao Jose Domingues Cordeiro wrote:
Hi,

this is happening in two completely different (but with the same deployment 
method) Ganglia headnodes.

I'm monitoring about 500 VM's (on each headnode), separated by clusters of 
different sizes. From time to time, the summary graphs over some cluster stop 
reporting, showing zero activity, and then suddenly after a while they come 
back up again.

This is very undesirable since I end up with several white holes per day on 
each cluster.

The information I can give you so far is the following:


  *   The attached image shows what happens
  *   I have a master-slave type of configuration, where the collector gmonds 
are sitting in the same machine (the headnode) as gmetad and ganglia-web, and 
where all the gmond nodes are reporting their metrics through unicast to the 
headnode.
  *   I have the latest Ganglia versions running (both core and web)
  *   All VM's are based on SL6
  *   When I look at /var/log/messages I see a lot of this:
 *   May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular 
data for pkts_out#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_user#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_system#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_idle#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_nice#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_aidle#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_wio#012
May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for 
cpu_steal#012
May 19 16:14:37 gangliamon gmond[22304]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_user#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_system#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_idle#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_nice#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_aidle#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_wio#012
May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for 
cpu_steal#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_free#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_shared#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_buffers#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
mem_cached#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
swap_free#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
bytes_out#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
bytes_in#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
pkts_in#012
May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for 
pkts_out#012
May 19 16:14:40 gangliamon gmond[10560]: Error 1 sending the modular data for 
heartbeat#012
May 19 16:14:42 gangliamon gmond[22304]: Error 1 sending the modular data for 
disk_free#012



Which I understand is a known unsolved issue, by looking at other discussions 
like 
https://www.mail-archive.com/ganglia-general@lists.sourceforge.net/msg06602.html
 .


Does anyone know how to solve this?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro



--
Accelerate Dev Cycles

[Ganglia-general] XSS vulnerabilities in Ganglia web

2014-04-11 Thread Cristovao Jose Domingues Cordeiro
Hi all,

recently I've updated my Ganglia web frontend to the latest version (so I could 
perform HTTP queries) and when I issued the security check with skipfish I got 
these:

Vulnerabilities found: 33

· Severity: 4, Type: File inclusion
..
..
· Severity: 4, Type: Query injection vector
..
..
· Severity: 4, Type: Shell injection vector
..
..
· Severity: 4, Type: Server-side XML injection vector
..
..
· Severity: 3, Type: Directory traversal / file inclusion possible
··
··
· Severity: 3, Type: XSS vector in document body
..
..


Now, these are too many vulnerabilities, but I don't know if they can affect 
the backend of if they just affect the frontend. Do you know?

The XSS vulnerability must be fixed for sure. I've seen some references to this 
in your release notes (e.g. 
http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html
 ) but in fact there if no difference between these last releases and the ones 
before that announcement.

Is there a workaround for this? I can not open this Ganglia machine to the 
outside world if I don't have this fixed.

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Put Bad Developers to Shame
Dominate Development with Jenkins Continuous Integration
Continuously Automate Build, Test  Deployment 
Start a new project now. Try Jenkins in the cloud.
http://p.sf.net/sfu/13600_Cloudbees___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] latency vs response probability in Unicast

2014-03-25 Thread Cristovao Jose Domingues Cordeiro
Hi,

I have a technical question which concerns my current Ganglia deployment.
Basically I have a unicast setup with multiple clusters and one server (no 
intermediate receiving gmonds), which means:

Gmetad runs on Server A and this server also has gmond_cluster1.conf and 
gmond_cluster2.conf;
Cluster 1 has X nodes, reporting back to gmond_cluster1 on Server A;
Cluster 2 has Y nodes, in  VERY remote location (higher ping), reporting back 
to gmond_cluster2 on Server A.

My questions are:
 - Since these are UDP connections, and baring in mind that Gmetad is polling 
the data every 15 sec(default), what's the probability of loosing packets?
 - Do the packets contain history information? Meaning: if I loose one packet 
from machine1, does the next packet contain the monitored information about the 
previous status of the machine?
 - If the packets contain all the node information history, how does the packet 
size influence the time of flight and the probability of getting well received 
at Gmetad?
 - Does changing the polling frequency (from 15 sec to higher of smaller value) 
influence the number of well received packages?
 - How about the transfer volume? How can I check the packets size?

Sorry for the long list of questions, but I'm evaluating different Ganglia 
setups and this matters.

Thanks in advance.

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] multicast not working

2014-03-21 Thread Cristovao Jose Domingues Cordeiro
Hi Khrist,

well that makes sense indeed.
All my VM's are running in a private cloud infrastructure so I have no control 
over their network characteristics.
But still, for instance, in gmetad I am gathering the data from node1, and 
node1's IP is in the same subnet as node2:

 - node1 IP= x.x.x.169
 - node2 IP= x.x.x.96

but in Ganglia frontend, I only get information from node1!!

Nevertheless, do I have to do any extra configuration to put multicast working? 
Because I did nothing in the gmonds. I just installed ganglia and ganglia-gmond 
and I left gmond.conf as default. Is this correct?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Khrist Hansen [khrist.han...@gmail.com]
Sent: 20 March 2014 23:07
To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net
Subject: RE: [Ganglia-general] multicast not working

I had the same problem due to multicast packets not being routed between 
subnets.

Are your nodes on different subnets/VLANs?

There are some tools out there that will test multicast connectivity across 
subnets.  All that comes to mind for the moment is mping on AIX, but I know 
there are more for the various OS platforms.

I had to use unicast in the end, but I am still working on my network 
engineering team to enable multicast routing between subnets.

Hope that helps,

Khrist Hansen


From: Cristovao Jose Domingues Cordeiro [mailto:cristovao.corde...@cern.ch]
Sent: Thursday, March 20, 2014 10:42 AM
To: ganglia-general@lists.sourceforge.net
Subject: [Ganglia-general] multicast not working

Hi,

I'm trying to set the simplest of clusters with mcast. Basically I have one 
gmetad.conf where I just did:
data_source unspecified node1

then, I have in one cloud, node1, node2, node3, node4 and node5.
All of these have default gmond installations with no changes in gmond.conf.

The problem is that in my frontend, I only get node1 and not the others.
Weren't they supposed to talk with each others and have all the metrics about 
each others?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] multicast not working

2014-03-20 Thread Cristovao Jose Domingues Cordeiro
Hi,

I'm trying to set the simplest of clusters with mcast. Basically I have one 
gmetad.conf where I just did:
data_source unspecified node1

then, I have in one cloud, node1, node2, node3, node4 and node5.
All of these have default gmond installations with no changes in gmond.conf.

The problem is that in my frontend, I only get node1 and not the others.
Weren't they supposed to talk with each others and have all the metrics about 
each others?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro

--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] FW: one gmetad gathering from multiple gmetad

2014-03-13 Thread Cristovao Jose Domingues Cordeiro
On the main gmetad I've only set 'all_trusted on' and on the reporting gmetad's 
I've set the trusted_hosts to have the hostname of the machine runnning the 
main gmetad.

It's just one specific gmetad which fails (when scalable is ON). I think it is 
an Apache issue.

'''
Forbidden

You don't have permission to access /ganglia/ on this server.
Apache/2.2.15 (Red Hat) Server at myhostname Port 80
'''

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/1-010
CERN


From: Bernard Li [bern...@vanhpc.org]
Sent: 12 March 2014 18:15
To: Cristovao Jose Domingues Cordeiro
Cc: Ganglia [ganglia-general@lists.sourceforge.net]
Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

Did you set the trusted_hosts option on gmetad.conf?

Cheers,

Bernard

On Wed, Mar 12, 2014 at 7:25 AM, Cristovao Jose Domingues Cordeiro
cristovao.corde...@cern.ch wrote:
 Ok I've made it work. I didn't know 8651 was the way to go.

 I only have one issue now, which is, if I keep the scalable option ON, which
 I want to, I get Permission Denied to one of the remote grids...any idea
 why?


 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

 
 From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch]
 Sent: 11 March 2014 15:02
 To: Bernard Li

 Cc: Ganglia [ganglia-general@lists.sourceforge.net]
 Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

 Hi Bernard,

 thanks for the reply.
 But my question is, in one gmetad I have 5 different data_sources,
 representing different clusters, and each own has its own port. In the MAIN
 gmetad (the one gathering the other gmetads), which port do I use for the
 multicluster one?

 What do you mean by collapse? I wanted to keep a clean separation of which
 gmetad I am monitoring...

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

 
 From: Bernard Li [bern...@vanhpc.org]
 Sent: 07 March 2014 17:37
 To: Cristovao Jose Domingues Cordeiro
 Cc: Ganglia [ganglia-general@lists.sourceforge.net]
 Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

 Hi Cristovao:

 If you are trying to accomplish a grid of grids, you'll have to setup a
 separate gmetad server and have a data_source entry for each gmetad server
 you have.  Make sure you put in the corresponding port for each gmetad
 server and that should be it.

 Alternatively, can you collapse all your gmetads into one?

 Cheers,

 Bernard

 On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro
 cristovao.corde...@cern.ch wrote:

 Hi,

 I have several gmetad headnodes corresponding each one to one grid of
 several clusters reporting metrics through unicast (meaning that in each of
 my gmeta headnodes I have several data_sources and several gmond files - one
 for each cluster).

 I would like now to aggregate all of these metric in just on MAIN
 headnode, like a Grid of grids. How can I do this?

 Thanks

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] FW: one gmetad gathering from multiple gmetad

2014-03-13 Thread Cristovao Jose Domingues Cordeiro
Hi Alexander,

thanks for the hint.
It was actually an Apache problem as I suspected. Just gave access to main 
monitor and now it works fine.

Thanks all.

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/1-010
CERN

From: Alexander Karner [a...@de.ibm.com]
Sent: 13 March 2014 09:31
To: Cristovao Jose Domingues Cordeiro
Cc: ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] FW: one gmetad gathering from multiple gmetad

Hi Christovao!

You should doublecheck the name resolution on both gmetad systems in both 
directions.
I use a multi-grid setup with actually five child gmetad's of which one failed 
to establish a connection with the central gmetad.

Interestingly the behaviour changed over time: For some time I had to use the 
short name, then the FQN and then the IP (and back again).

So for this specific system I set this child gmetad to all_trusted on but we 
have communication limited by the firewall.

Mit freundlichen Grüßen / Kind regards

Alexander Karner

Program Manager Health Check
Senior Certified IBM Architect
Senior IT Specialist
Global Technology Services
B141
















From:Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch
To:ganglia-general@lists.sourceforge.net 
ganglia-general@lists.sourceforge.net,
Date:13.03.2014 09:23
Subject:[Ganglia-general] FW:  one gmetad gathering from multiple gmetad




On the main gmetad I've only set 'all_trusted on' and on the reporting gmetad's 
I've set the trusted_hosts to have the hostname of the machine runnning the 
main gmetad.

It's just one specific gmetad which fails (when scalable is ON). I think it is 
an Apache issue.

'''
Forbidden

You don't have permission to access /ganglia/ on this server.
Apache/2.2.15 (Red Hat) Server at myhostname Port 80
'''

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/1-010
CERN


From: Bernard Li [bern...@vanhpc.org]
Sent: 12 March 2014 18:15
To: Cristovao Jose Domingues Cordeiro
Cc: Ganglia [ganglia-general@lists.sourceforge.net]
Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

Did you set the trusted_hosts option on gmetad.conf?

Cheers,

Bernard

On Wed, Mar 12, 2014 at 7:25 AM, Cristovao Jose Domingues Cordeiro
cristovao.corde...@cern.ch wrote:
 Ok I've made it work. I didn't know 8651 was the way to go.

 I only have one issue now, which is, if I keep the scalable option ON, which
 I want to, I get Permission Denied to one of the remote grids...any idea
 why?


 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

 
 From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch]
 Sent: 11 March 2014 15:02
 To: Bernard Li

 Cc: Ganglia [ganglia-general@lists.sourceforge.net]
 Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

 Hi Bernard,

 thanks for the reply.
 But my question is, in one gmetad I have 5 different data_sources,
 representing different clusters, and each own has its own port. In the MAIN
 gmetad (the one gathering the other gmetads), which port do I use for the
 multicluster one?

 What do you mean by collapse? I wanted to keep a clean separation of which
 gmetad I am monitoring...

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

 
 From: Bernard Li [bern...@vanhpc.org]
 Sent: 07 March 2014 17:37
 To: Cristovao Jose Domingues Cordeiro
 Cc: Ganglia [ganglia-general@lists.sourceforge.net]
 Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

 Hi Cristovao:

 If you are trying to accomplish a grid of grids, you'll have to setup a
 separate gmetad server and have a data_source entry for each gmetad server
 you have.  Make sure you put in the corresponding port for each gmetad
 server and that should be it.

 Alternatively, can you collapse all your gmetads into one?

 Cheers,

 Bernard

 On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro
 cristovao.corde...@cern.ch wrote:

 Hi,

 I have several gmetad headnodes corresponding each one to one grid of
 several clusters reporting metrics through unicast (meaning that in each of
 my gmeta headnodes I have several data_sources and several gmond files - one
 for each cluster).

 I would like now to aggregate all of these metric in just on MAIN
 headnode, like a Grid of grids. How can I do this?

 Thanks

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net

Re: [Ganglia-general] one gmetad gathering from multiple gmetad

2014-03-12 Thread Cristovao Jose Domingues Cordeiro
Ok I've made it work. I didn't know 8651 was the way to go.

I only have one issue now, which is, if I keep the scalable option ON, which I 
want to, I get Permission Denied to one of the remote grids...any idea why?

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch]
Sent: 11 March 2014 15:02
To: Bernard Li
Cc: Ganglia [ganglia-general@lists.sourceforge.net]
Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

Hi Bernard,

thanks for the reply.
But my question is, in one gmetad I have 5 different data_sources, representing 
different clusters, and each own has its own port. In the MAIN gmetad (the one 
gathering the other gmetads), which port do I use for the multicluster one?

What do you mean by collapse? I wanted to keep a clean separation of which 
gmetad I am monitoring...

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Bernard Li [bern...@vanhpc.org]
Sent: 07 March 2014 17:37
To: Cristovao Jose Domingues Cordeiro
Cc: Ganglia [ganglia-general@lists.sourceforge.net]
Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

Hi Cristovao:

If you are trying to accomplish a grid of grids, you'll have to setup a 
separate gmetad server and have a data_source entry for each gmetad server you 
have.  Make sure you put in the corresponding port for each gmetad server and 
that should be it.

Alternatively, can you collapse all your gmetads into one?

Cheers,

Bernard

On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro 
cristovao.corde...@cern.chmailto:cristovao.corde...@cern.ch wrote:
Hi,

I have several gmetad headnodes corresponding each one to one grid of several 
clusters reporting metrics through unicast (meaning that in each of my gmeta 
headnodes I have several data_sources and several gmond files - one for each 
cluster).

I would like now to aggregate all of these metric in just on MAIN headnode, 
like a Grid of grids. How can I do this?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] one gmetad gathering from multiple gmetad

2014-03-11 Thread Cristovao Jose Domingues Cordeiro
Hi Bernard,

thanks for the reply.
But my question is, in one gmetad I have 5 different data_sources, representing 
different clusters, and each own has its own port. In the MAIN gmetad (the one 
gathering the other gmetads), which port do I use for the multicluster one?

What do you mean by collapse? I wanted to keep a clean separation of which 
gmetad I am monitoring...

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Bernard Li [bern...@vanhpc.org]
Sent: 07 March 2014 17:37
To: Cristovao Jose Domingues Cordeiro
Cc: Ganglia [ganglia-general@lists.sourceforge.net]
Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad

Hi Cristovao:

If you are trying to accomplish a grid of grids, you'll have to setup a 
separate gmetad server and have a data_source entry for each gmetad server you 
have.  Make sure you put in the corresponding port for each gmetad server and 
that should be it.

Alternatively, can you collapse all your gmetads into one?

Cheers,

Bernard

On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro 
cristovao.corde...@cern.chmailto:cristovao.corde...@cern.ch wrote:
Hi,

I have several gmetad headnodes corresponding each one to one grid of several 
clusters reporting metrics through unicast (meaning that in each of my gmeta 
headnodes I have several data_sources and several gmond files - one for each 
cluster).

I would like now to aggregate all of these metric in just on MAIN headnode, 
like a Grid of grids. How can I do this?

Thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
--
Learn Graph Databases - Download FREE O'Reilly Book
Graph Databases is the definitive new guide to graph databases and their
applications. Written by three acclaimed leaders in the field,
this first edition is now available. Download your free book today!
http://p.sf.net/sfu/13534_NeoTech___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] multiple clusters with just one collector

2014-01-27 Thread Cristovao Jose Domingues Cordeiro
Hi,

sorry for the late reply.
Oh ok those were the problematic ports. Good to know thanks!
Unfortunately and at least for now this monitoring page will only be accessible 
inside the private network. The only ports open for the outside world are the 
gmond ones. Sorry. If this changes I'll let you know.

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Bernard Li [bern...@vanhpc.org]
Sent: 24 January 2014 19:49
To: Cristovao Jose Domingues Cordeiro
Cc: ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] multiple clusters with just one collector

Yes, 8651 and 8652 are reserved ports for gmetad, you can start from
port 9000 to avoid this issue :-)

Glad that you got it working...  are you guys planning to make your
monitoring page public?  It would be cool if we can reference it
somewhere :-)

Cheers,

Bernard

On Fri, Jan 24, 2014 at 12:47 AM, Cristovao Jose Domingues Cordeiro
cristovao.corde...@cern.ch wrote:
 Thank you for your replies Bernard and Shekar.

 Actually in the meanwhile I managed to make it work. For those who might want 
 to do the same here's what I did:

  · Stop gmond and gmetad, and clean the RRD database in the collector (I was 
 getting unspecified clusters and cached information so this was a clean start)

  · I have 5 clusters so I've created 5 gmond.conf files - 
 gmond-cluster1.conf, gmond-cluster2.conf...etc. In these files I've set up 
 the cluster name, both udp and tcp ports to the same number, and the hostname 
 in the udp_send_channel to 'localhost'

  · In gmetad.conf I set up a grid name (not necessary but better to have a 
 cleaner UI)

  · Still in gmetad.conf I have created 5 data_sources, one for each cluster:
   data_source Cluster A localhost:portA
   data_source Cluster B localhost:portB
   ...
Note that these ports and cluster names match those in the gmond files 
 I've created earlier.

I had some problems with the ports, some clusters were not appearing if 
 they were using port 8651 for instance, but I just changed the port and they 
 worked...I am not sure why, I don't think this port numbers are 
 restricted...but well, if someone knows why, tell me please

  · Then I just daemonized the 5 gmonds like this:
# deamon gmond -c /etc/ganglia/gmond-clusterA.conf
# .

  · Finally, I just started gmetad and now it's running fine. To add more 
 clusters I just need to, in this order, daemonized the new gmond for that 
 cluster and restart gmetad.


 Hope this helps someone in the future ;)

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro

 
 From: Bernard Li [bern...@vanhpc.org]
 Sent: 24 January 2014 04:49
 To: Cristovao Jose Domingues Cordeiro
 Cc: ganglia-general@lists.sourceforge.net
 Subject: Re: [Ganglia-general] multiple clusters with just one collector

 Hello:

 To accomplish what you want to do, I believe you have two options:

 1) Run multiple gmonds, on the master collector, each one listening on
 different ports
 2) Have one gmond but have multiple udp_recv_channels listening on
 different ports

 It might help if you also post your configuration somewhere and
 reference it here (eg. pastebin or some such).

 Please let us know if you have any other questions.

 Cheers,

 Bernard

 On Thu, Jan 23, 2014 at 5:08 AM, Cristovao Jose Domingues Cordeiro
 cristovao.corde...@cern.ch wrote:
 Hi,

 I found some similar posts in this mailing list but unfortunately none can
 solve my situation.

 Basically, I want to have just one collector to multiple clusters.
 As far as I understood I need to have multiple gmonds in my headnode, and so
 I do.

 I have in my collector:
 gmetad.conf:
   data_source clusterA localhost
   data_source clusterB localhost:8650

 gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports,
 hosts and cluster names defined, pointing to localhost.

 I also issued the daemon:
 gmond -c gmond-clusterA.conf
 gmond -c gmond-clusterB.conf

 But  this doesn't work. I end up with my web interface just showing an
 'unspecified' cluster and 'clusterA'... not clusterB.

 Any thoughts? thanks

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro


 --
 CenturyLink Cloud: The Leader in Enterprise Cloud Services.
 Learn Why More Businesses Are Choosing CenturyLink Cloud For
 Critical Workloads, Development Environments  Everything In Between.
 Get a Quote or Start a Free Trial Today.
 http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk
 ___
 Ganglia-general mailing list
 Ganglia-general@lists.sourceforge.net
 https://lists.sourceforge.net/lists/listinfo/ganglia-general


--
CenturyLink Cloud: The Leader in Enterprise

Re: [Ganglia-general] multiple clusters with just one collector

2014-01-27 Thread Cristovao Jose Domingues Cordeiro
Hi Adrian,

I saw that the other guys are already helping you so there's nothing that I can 
add which would help you I think.
Nevertheless I must say that I don't have an aggregation gmond, that was the 
point of my question, not having one. I receive all the metric directly in the 
gmetad node.

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro



From: Adrian Sevcenco [adrian.sevce...@cern.ch]
Sent: 25 January 2014 13:19
To: ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] multiple clusters with just one collector

On 01/24/2014 10:47 AM, Cristovao Jose Domingues Cordeiro wrote:
 Thank you for your replies Bernard and Shekar.

 Actually in the meanwhile I managed to make it work. For those who
 might want to do the same here's what I did:

 · Stop gmond and gmetad, and clean the RRD database in the collector
 (I was getting unspecified clusters and cached information so this
 was a clean start)

 · I have 5 clusters so I've created 5 gmond.conf files -
 gmond-cluster1.conf, gmond-cluster2.conf...etc. In these files I've
 set up the cluster name, both udp and tcp ports to the same number,
 and the hostname in the udp_send_channel to 'localhost'
Hi! Is it possible to post your gmond settings?
How did you solve the problem that the reporting node (where aggregation
gmond sits) will also report the self metrics?

My problem is a little different : i try to make a fake gmond for
gmetric input for UPS/RAID/Switches metrics.

Thanks a lot!
Adrian


--
CenturyLink Cloud: The Leader in Enterprise Cloud Services.
Learn Why More Businesses Are Choosing CenturyLink Cloud For
Critical Workloads, Development Environments  Everything In Between.
Get a Quote or Start a Free Trial Today. 
http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] multiple clusters with just one collector

2014-01-24 Thread Cristovao Jose Domingues Cordeiro
Thank you for your replies Bernard and Shekar.

Actually in the meanwhile I managed to make it work. For those who might want 
to do the same here's what I did:

 · Stop gmond and gmetad, and clean the RRD database in the collector (I was 
getting unspecified clusters and cached information so this was a clean start)

 · I have 5 clusters so I've created 5 gmond.conf files - gmond-cluster1.conf, 
gmond-cluster2.conf...etc. In these files I've set up the cluster name, both 
udp and tcp ports to the same number, and the hostname in the udp_send_channel 
to 'localhost'

 · In gmetad.conf I set up a grid name (not necessary but better to have a 
cleaner UI)

 · Still in gmetad.conf I have created 5 data_sources, one for each cluster:
  data_source Cluster A localhost:portA
  data_source Cluster B localhost:portB
  ...
   Note that these ports and cluster names match those in the gmond files I've 
created earlier. 

   I had some problems with the ports, some clusters were not appearing if they 
were using port 8651 for instance, but I just changed the port and they 
worked...I am not sure why, I don't think this port numbers are 
restricted...but well, if someone knows why, tell me please

 · Then I just daemonized the 5 gmonds like this:
   # deamon gmond -c /etc/ganglia/gmond-clusterA.conf
   # .

 · Finally, I just started gmetad and now it's running fine. To add more 
clusters I just need to, in this order, daemonized the new gmond for that 
cluster and restart gmetad.


Hope this helps someone in the future ;)

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro


From: Bernard Li [bern...@vanhpc.org]
Sent: 24 January 2014 04:49
To: Cristovao Jose Domingues Cordeiro
Cc: ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] multiple clusters with just one collector

Hello:

To accomplish what you want to do, I believe you have two options:

1) Run multiple gmonds, on the master collector, each one listening on
different ports
2) Have one gmond but have multiple udp_recv_channels listening on
different ports

It might help if you also post your configuration somewhere and
reference it here (eg. pastebin or some such).

Please let us know if you have any other questions.

Cheers,

Bernard

On Thu, Jan 23, 2014 at 5:08 AM, Cristovao Jose Domingues Cordeiro
cristovao.corde...@cern.ch wrote:
 Hi,

 I found some similar posts in this mailing list but unfortunately none can
 solve my situation.

 Basically, I want to have just one collector to multiple clusters.
 As far as I understood I need to have multiple gmonds in my headnode, and so
 I do.

 I have in my collector:
 gmetad.conf:
   data_source clusterA localhost
   data_source clusterB localhost:8650

 gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports,
 hosts and cluster names defined, pointing to localhost.

 I also issued the daemon:
 gmond -c gmond-clusterA.conf
 gmond -c gmond-clusterB.conf

 But  this doesn't work. I end up with my web interface just showing an
 'unspecified' cluster and 'clusterA'... not clusterB.

 Any thoughts? thanks

 Cumprimentos / Best regards,
 Cristóvão José Domingues Cordeiro


 --
 CenturyLink Cloud: The Leader in Enterprise Cloud Services.
 Learn Why More Businesses Are Choosing CenturyLink Cloud For
 Critical Workloads, Development Environments  Everything In Between.
 Get a Quote or Start a Free Trial Today.
 http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk
 ___
 Ganglia-general mailing list
 Ganglia-general@lists.sourceforge.net
 https://lists.sourceforge.net/lists/listinfo/ganglia-general


--
CenturyLink Cloud: The Leader in Enterprise Cloud Services.
Learn Why More Businesses Are Choosing CenturyLink Cloud For
Critical Workloads, Development Environments  Everything In Between.
Get a Quote or Start a Free Trial Today. 
http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


[Ganglia-general] multiple clusters with just one collector

2014-01-23 Thread Cristovao Jose Domingues Cordeiro
Hi,

I found some similar posts in this mailing list but unfortunately none can 
solve my situation.

Basically, I want to have just one collector to multiple clusters.
As far as I understood I need to have multiple gmonds in my headnode, and so I 
do.

I have in my collector:
gmetad.conf:
  data_source clusterA localhost
  data_source clusterB localhost:8650

gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports, hosts 
and cluster names defined, pointing to localhost.

I also issued the daemon:
gmond -c gmond-clusterA.conf
gmond -c gmond-clusterB.conf

But  this doesn't work. I end up with my web interface just showing an 
'unspecified' cluster and 'clusterA'... not clusterB.

Any thoughts? thanks

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro

--
CenturyLink Cloud: The Leader in Enterprise Cloud Services.
Learn Why More Businesses Are Choosing CenturyLink Cloud For
Critical Workloads, Development Environments  Everything In Between.
Get a Quote or Start a Free Trial Today. 
http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general


Re: [Ganglia-general] multiple clusters with just one collector

2014-01-23 Thread Cristovao Jose Domingues Cordeiro
Hi Gowtham,

thanks but that makes use of two frontends, and I only have one. In that 
example server B is running in one of the clusters and my clusters have a 
random number of VM's which only live for a few days, top.

My master server needs to be collecting everything from every cluster. I am 
reading solutions about having several gmonds in the collector, but I am not 
succeeding. 

Cumprimentos / Best regards,
Cristóvão José Domingues Cordeiro
IT Department - 28/1-010
CERN


From: Gowtham [sgowt...@mtu.edu]
Sent: 23 January 2014 15:04
To: Cristovao Jose Domingues Cordeiro
Cc: ganglia-general@lists.sourceforge.net
Subject: Re: [Ganglia-general] multiple clusters with just one collector

A working example of following the instructions in the link below is here:

  http://hpcmonitor.it.mtu.edu/

Best regards,
g

--
Gowtham, PhD
HPC Research Scientist, ITS
Adj. Asst. Professor, Physics/ECE
Michigan Technological University

(906) 487/3593
http://it.mtu.edu


On Thu, 23 Jan 2014, Gowtham wrote:

|
| Instructions here might be of use to you:
|
|   http://central6.rocksclusters.org/roll-documentation/ganglia/6.1/x111.html
|
| Best regards,
| g
|
| --
| Gowtham, PhD
| HPC Research Scientist, ITS
| Adj. Asst. Professor, Physics/ECE
| Michigan Technological University
|
| (906) 487/3593
| http://it.mtu.edu
|
|
| On Thu, 23 Jan 2014, Cristovao Jose Domingues Cordeiro wrote:
|
| | Hi,
| |
| | I found some similar posts in this mailing list but unfortunately none can
| | solve my situation.
| |
| | Basically, I want to have just one collector to multiple clusters.
| | As far as I understood I need to have multiple gmonds in my headnode, and so
| | I do.
| |
| | I have in my collector:
| | gmetad.conf:
| |   data_source clusterA localhost
| |   data_source clusterB localhost:8650
| |
| | gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports,
| | hosts and cluster names defined, pointing to localhost.
| |
| | I also issued the daemon:
| | gmond -c gmond-clusterA.conf
| | gmond -c gmond-clusterB.conf
| |
| | But  this doesn't work. I end up with my web interface just showing an
| | 'unspecified' cluster and 'clusterA'... not clusterB.
| |
| | Any thoughts? thanks
| |
| | Cumprimentos / Best regards,
| | Cristóvão José Domingues Cordeiro
| |
| |
| |

--
CenturyLink Cloud: The Leader in Enterprise Cloud Services.
Learn Why More Businesses Are Choosing CenturyLink Cloud For
Critical Workloads, Development Environments  Everything In Between.
Get a Quote or Start a Free Trial Today. 
http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk
___
Ganglia-general mailing list
Ganglia-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/ganglia-general