Re: [Ganglia-general] XSS vulnerabilities in Ganglia web
Hi, sorry for the late reply. I've finally tried to install the latest web version. It seems indeed that the below vulnerabilities have been dealt with. The only issue which is now being reported after the scan is: NVT: Missing httpOnly Cookie Attribute Summary: The application is missing the 'httpOnly' cookie attribute Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/R-018 CERN From: Vladimir Vuksan [vli...@veus.hr] Sent: 07 November 2014 23:06 To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web I recall trying this out on 3.6.2 and I couldn't reproduce it so if you could run this against 3.6.2 and see if you can reproduce it that would be really helpful. Vladimir On 11/07/2014 04:50 PM, Cristovao Jose Domingues Cordeiro wrote: It is implemented on 3.5.12 Is this fixed on the latest version? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/R-018 CERN From: Vladimir Vuksan [vli...@veus.hrmailto:vli...@veus.hr] Sent: 07 November 2014 22:31 To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.netmailto:ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web Hi Cristovao, what Ganglia Web version was tested ? Is this against latest e.g. 3.6.2 ? Thanks, Vladimir On 04/11/2014 03:35 AM, Cristovao Jose Domingues Cordeiro wrote: Hi all, recently I've updated my Ganglia web frontend to the latest version (so I could perform HTTP queries) and when I issued the security check with skipfish I got these: Vulnerabilities found: 33 · Severity: 4, Type: File inclusion .. .. · Severity: 4, Type: Query injection vector .. .. · Severity: 4, Type: Shell injection vector .. .. · Severity: 4, Type: Server-side XML injection vector .. .. · Severity: 3, Type: Directory traversal / file inclusion possible ·· ·· · Severity: 3, Type: XSS vector in document body .. .. Now, these are too many vulnerabilities, but I don't know if they can affect the backend of if they just affect the frontend. Do you know? The XSS vulnerability must be fixed for sure. I've seen some references to this in your release notes (e.g. http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html ) but in fact there if no difference between these last releases and the ones before that announcement. Is there a workaround for this? I can not open this Ganglia machine to the outside world if I don't have this fixed. -- New Year. New Location. New Benefits. New Data Center in Ashburn, VA. GigeNET is offering a free month of service with a new server in Ashburn. Choose from 2 high performing configs, both with 100TB of bandwidth. Higher redundancy.Lower latency.Increased capacity.Completely compliant. http://p.sf.net/sfu/gigenet___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] XSS vulnerabilities in Ganglia web
It is implemented on 3.5.12 Is this fixed on the latest version? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/R-018 CERN From: Vladimir Vuksan [vli...@veus.hr] Sent: 07 November 2014 22:31 To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] XSS vulnerabilities in Ganglia web Hi Cristovao, what Ganglia Web version was tested ? Is this against latest e.g. 3.6.2 ? Thanks, Vladimir On 04/11/2014 03:35 AM, Cristovao Jose Domingues Cordeiro wrote: Hi all, recently I've updated my Ganglia web frontend to the latest version (so I could perform HTTP queries) and when I issued the security check with skipfish I got these: Vulnerabilities found: 33 · Severity: 4, Type: File inclusion .. .. · Severity: 4, Type: Query injection vector .. .. · Severity: 4, Type: Shell injection vector .. .. · Severity: 4, Type: Server-side XML injection vector .. .. · Severity: 3, Type: Directory traversal / file inclusion possible ·· ·· · Severity: 3, Type: XSS vector in document body .. .. Now, these are too many vulnerabilities, but I don't know if they can affect the backend of if they just affect the frontend. Do you know? The XSS vulnerability must be fixed for sure. I've seen some references to this in your release notes (e.g. http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html ) but in fact there if no difference between these last releases and the ones before that announcement. Is there a workaround for this? I can not open this Ganglia machine to the outside world if I don't have this fixed. -- ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] XSS Vulnerability
Hi all, just wondering if there are any plans to take care of this issue in next releases of Ganglia web? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/R-018 CERN -- ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk
Hi again, can someone point to the code where action are taken upon the gmon configuration 'host_tmax' and 'host_dmax'? Because that's where the RRD physical hosts should be deleted in my opinion Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Cristovao Jose Domingues Cordeiro Sent: 26 August 2014 09:22 To: Adam Compton; Ganglia List [ganglia-general@lists.sourceforge.net] Subject: RE: [Ganglia-general] Down hosts remain on disk and flood RAM disk Hi Adam, Thanks. Yeah that might be the ultimate solution. But I guess this is a known problem Ganglia developers had to deal with before right? It's an unavoidable scenario that will break gmetad. There must be some kind of protection... From: Adam Compton [acomp...@quantcast.com] Sent: 26 August 2014 00:59 To: Cristovao Jose Domingues Cordeiro; Ganglia List [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk On 8/25/14 6:53 AM, Cristovao Jose Domingues Cordeiro wrote: Hi all, I have a Ganglia setup running with a 4Gb ramdisk. I'm monitoring several opportunistic clusters with short-lifetimed VM's which are renewed a lot, causing a huge number of down hosts in Ganglia. From a web interface point of view this is solved with the host_dmax option, and it works fine, but the problem is that even though the VM's are down and out of the web interface, they still remain on the ramdisk, causing it to flood eventually, and consequently breaking gmetad. Any solutions? We have a cronjob that finds and purges RRDs that have not been modified in N days (currently 90 in most cases). I would also be interested in other peoples' solutions to this problem. - Adam -- Slashdot TV. Video for Nerds. Stuff that matters. http://tv.slashdot.org/___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk
Hi Adam, Thanks. Yeah that might be the ultimate solution. But I guess this is a known problem Ganglia developers had to deal with before right? It's an unavoidable scenario that will break gmetad. There must be some kind of protection... From: Adam Compton [acomp...@quantcast.com] Sent: 26 August 2014 00:59 To: Cristovao Jose Domingues Cordeiro; Ganglia List [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] Down hosts remain on disk and flood RAM disk On 8/25/14 6:53 AM, Cristovao Jose Domingues Cordeiro wrote: Hi all, I have a Ganglia setup running with a 4Gb ramdisk. I'm monitoring several opportunistic clusters with short-lifetimed VM's which are renewed a lot, causing a huge number of down hosts in Ganglia. From a web interface point of view this is solved with the host_dmax option, and it works fine, but the problem is that even though the VM's are down and out of the web interface, they still remain on the ramdisk, causing it to flood eventually, and consequently breaking gmetad. Any solutions? We have a cronjob that finds and purges RRDs that have not been modified in N days (currently 90 in most cases). I would also be interested in other peoples' solutions to this problem. - Adam -- Slashdot TV. Video for Nerds. Stuff that matters. http://tv.slashdot.org/___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] Huge metrics' size being reported to gmetad
Hi, I have a grid configuration, with several clusters. I am also using RAM disk for I/O optimization (4GB). I've been noticing that sometimes, gmetad breaks, complaining about lack of space in this tmpfs partition. I checked and I saw that for some reason, some clusters, have hosts which occupy 3MB, 6MB and even sometimes 150MB All together makes the cluster occupy 2GB and consequently occupy half of the ramdisk space. We would normally expect these host metrics to have +/- 336k right? Has anyone experienced this? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/1-010 CERN -- Open source business process management suite built on Java and Eclipse Turn processes into business applications with Bonita BPM Community Edition Quickly connect people, data, and systems into organized workflows Winner of BOSSIE, CODIE, OW2 and Gartner awards http://p.sf.net/sfu/Bonitasoft___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] Random blank timeslots in graphs
Hi, this is happening in two completely different (but with the same deployment method) Ganglia headnodes. I'm monitoring about 500 VM's (on each headnode), separated by clusters of different sizes. From time to time, the summary graphs over some cluster stop reporting, showing zero activity, and then suddenly after a while they come back up again. This is very undesirable since I end up with several white holes per day on each cluster. The information I can give you so far is the following: * The attached image shows what happens * I have a master-slave type of configuration, where the collector gmonds are sitting in the same machine (the headnode) as gmetad and ganglia-web, and where all the gmond nodes are reporting their metrics through unicast to the headnode. * I have the latest Ganglia versions running (both core and web) * All VM's are based on SL6 * When I look at /var/log/messages I see a lot of this: * May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for pkts_out#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_user#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_system#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_idle#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_nice#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_aidle#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_wio#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_steal#012 May 19 16:14:37 gangliamon gmond[22304]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_user#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_system#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_idle#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_nice#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_aidle#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_wio#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_steal#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_free#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_shared#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_buffers#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_cached#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for swap_free#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for bytes_out#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for bytes_in#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for pkts_in#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for pkts_out#012 May 19 16:14:40 gangliamon gmond[10560]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:42 gangliamon gmond[22304]: Error 1 sending the modular data for disk_free#012 Which I understand is a known unsolved issue, by looking at other discussions like https://www.mail-archive.com/ganglia-general@lists.sourceforge.net/msg06602.html . Does anyone know how to solve this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Accelerate Dev Cycles with Automated Cross-Browser Testing - For FREE Instantly run your Selenium tests across 300+ browser/OS combos. Get unparalleled scalability from the best Selenium testing platform available Simple to use. Nothing to install. Get started now for free. http://p.sf.net/sfu/SauceLabs___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] Random blank timeslots in graphs
Hi, I am using Ganglia Web Frontend version 3.5.12 and Ganglia Web Backend (gmetad) version 3.6.0. The Gmond version on the nodes is not consistent, since they are being set by different users, on different environments. But I believe their version is not below 3.1.7. No, I am not using RRDCached...all of my Ganglia configurations are the default ones. I'll try to set that up. Since you believe it is a scaling problem, should I try to store the DB in ramdisk? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Vladimir Vuksan [vli...@veus.hr] Sent: 19 May 2014 16:37 To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] Random blank timeslots in graphs Error 1 sending messages are a red herring. If you are seeing gaps it's most likely that storage system is not keeping up. What version of ganglia are you using and are you using rrdcached ? Vladimir On 05/19/2014 10:20 AM, Cristovao Jose Domingues Cordeiro wrote: Hi, this is happening in two completely different (but with the same deployment method) Ganglia headnodes. I'm monitoring about 500 VM's (on each headnode), separated by clusters of different sizes. From time to time, the summary graphs over some cluster stop reporting, showing zero activity, and then suddenly after a while they come back up again. This is very undesirable since I end up with several white holes per day on each cluster. The information I can give you so far is the following: * The attached image shows what happens * I have a master-slave type of configuration, where the collector gmonds are sitting in the same machine (the headnode) as gmetad and ganglia-web, and where all the gmond nodes are reporting their metrics through unicast to the headnode. * I have the latest Ganglia versions running (both core and web) * All VM's are based on SL6 * When I look at /var/log/messages I see a lot of this: * May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for pkts_out#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_user#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_system#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_idle#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_nice#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_aidle#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_wio#012 May 19 16:14:36 gangliamon gmond[22292]: Error 1 sending the modular data for cpu_steal#012 May 19 16:14:37 gangliamon gmond[22304]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_user#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_system#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_idle#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_nice#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_aidle#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_wio#012 May 19 16:14:38 gangliamon gmond[10560]: Error 1 sending the modular data for cpu_steal#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_free#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_shared#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_buffers#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for mem_cached#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for swap_free#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for bytes_out#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for bytes_in#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for pkts_in#012 May 19 16:14:39 gangliamon gmond[22300]: Error 1 sending the modular data for pkts_out#012 May 19 16:14:40 gangliamon gmond[10560]: Error 1 sending the modular data for heartbeat#012 May 19 16:14:42 gangliamon gmond[22304]: Error 1 sending the modular data for disk_free#012 Which I understand is a known unsolved issue, by looking at other discussions like https://www.mail-archive.com/ganglia-general@lists.sourceforge.net/msg06602.html . Does anyone know how to solve this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Accelerate Dev Cycles
[Ganglia-general] XSS vulnerabilities in Ganglia web
Hi all, recently I've updated my Ganglia web frontend to the latest version (so I could perform HTTP queries) and when I issued the security check with skipfish I got these: Vulnerabilities found: 33 · Severity: 4, Type: File inclusion .. .. · Severity: 4, Type: Query injection vector .. .. · Severity: 4, Type: Shell injection vector .. .. · Severity: 4, Type: Server-side XML injection vector .. .. · Severity: 3, Type: Directory traversal / file inclusion possible ·· ·· · Severity: 3, Type: XSS vector in document body .. .. Now, these are too many vulnerabilities, but I don't know if they can affect the backend of if they just affect the frontend. Do you know? The XSS vulnerability must be fixed for sure. I've seen some references to this in your release notes (e.g. http://www.mail-archive.com/ganglia-general%40lists.sourceforge.net/msg08004.html ) but in fact there if no difference between these last releases and the ones before that announcement. Is there a workaround for this? I can not open this Ganglia machine to the outside world if I don't have this fixed. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Put Bad Developers to Shame Dominate Development with Jenkins Continuous Integration Continuously Automate Build, Test Deployment Start a new project now. Try Jenkins in the cloud. http://p.sf.net/sfu/13600_Cloudbees___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] latency vs response probability in Unicast
Hi, I have a technical question which concerns my current Ganglia deployment. Basically I have a unicast setup with multiple clusters and one server (no intermediate receiving gmonds), which means: Gmetad runs on Server A and this server also has gmond_cluster1.conf and gmond_cluster2.conf; Cluster 1 has X nodes, reporting back to gmond_cluster1 on Server A; Cluster 2 has Y nodes, in VERY remote location (higher ping), reporting back to gmond_cluster2 on Server A. My questions are: - Since these are UDP connections, and baring in mind that Gmetad is polling the data every 15 sec(default), what's the probability of loosing packets? - Do the packets contain history information? Meaning: if I loose one packet from machine1, does the next packet contain the monitored information about the previous status of the machine? - If the packets contain all the node information history, how does the packet size influence the time of flight and the probability of getting well received at Gmetad? - Does changing the polling frequency (from 15 sec to higher of smaller value) influence the number of well received packages? - How about the transfer volume? How can I check the packets size? Sorry for the long list of questions, but I'm evaluating different Ganglia setups and this matters. Thanks in advance. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] multicast not working
Hi Khrist, well that makes sense indeed. All my VM's are running in a private cloud infrastructure so I have no control over their network characteristics. But still, for instance, in gmetad I am gathering the data from node1, and node1's IP is in the same subnet as node2: - node1 IP= x.x.x.169 - node2 IP= x.x.x.96 but in Ganglia frontend, I only get information from node1!! Nevertheless, do I have to do any extra configuration to put multicast working? Because I did nothing in the gmonds. I just installed ganglia and ganglia-gmond and I left gmond.conf as default. Is this correct? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Khrist Hansen [khrist.han...@gmail.com] Sent: 20 March 2014 23:07 To: Cristovao Jose Domingues Cordeiro; ganglia-general@lists.sourceforge.net Subject: RE: [Ganglia-general] multicast not working I had the same problem due to multicast packets not being routed between subnets. Are your nodes on different subnets/VLANs? There are some tools out there that will test multicast connectivity across subnets. All that comes to mind for the moment is mping on AIX, but I know there are more for the various OS platforms. I had to use unicast in the end, but I am still working on my network engineering team to enable multicast routing between subnets. Hope that helps, Khrist Hansen From: Cristovao Jose Domingues Cordeiro [mailto:cristovao.corde...@cern.ch] Sent: Thursday, March 20, 2014 10:42 AM To: ganglia-general@lists.sourceforge.net Subject: [Ganglia-general] multicast not working Hi, I'm trying to set the simplest of clusters with mcast. Basically I have one gmetad.conf where I just did: data_source unspecified node1 then, I have in one cloud, node1, node2, node3, node4 and node5. All of these have default gmond installations with no changes in gmond.conf. The problem is that in my frontend, I only get node1 and not the others. Weren't they supposed to talk with each others and have all the metrics about each others? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] multicast not working
Hi, I'm trying to set the simplest of clusters with mcast. Basically I have one gmetad.conf where I just did: data_source unspecified node1 then, I have in one cloud, node1, node2, node3, node4 and node5. All of these have default gmond installations with no changes in gmond.conf. The problem is that in my frontend, I only get node1 and not the others. Weren't they supposed to talk with each others and have all the metrics about each others? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] FW: one gmetad gathering from multiple gmetad
On the main gmetad I've only set 'all_trusted on' and on the reporting gmetad's I've set the trusted_hosts to have the hostname of the machine runnning the main gmetad. It's just one specific gmetad which fails (when scalable is ON). I think it is an Apache issue. ''' Forbidden You don't have permission to access /ganglia/ on this server. Apache/2.2.15 (Red Hat) Server at myhostname Port 80 ''' Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/1-010 CERN From: Bernard Li [bern...@vanhpc.org] Sent: 12 March 2014 18:15 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Did you set the trusted_hosts option on gmetad.conf? Cheers, Bernard On Wed, Mar 12, 2014 at 7:25 AM, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Ok I've made it work. I didn't know 8651 was the way to go. I only have one issue now, which is, if I keep the scalable option ON, which I want to, I get Permission Denied to one of the remote grids...any idea why? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch] Sent: 11 March 2014 15:02 To: Bernard Li Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Bernard, thanks for the reply. But my question is, in one gmetad I have 5 different data_sources, representing different clusters, and each own has its own port. In the MAIN gmetad (the one gathering the other gmetads), which port do I use for the multicluster one? What do you mean by collapse? I wanted to keep a clean separation of which gmetad I am monitoring... Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 07 March 2014 17:37 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Cristovao: If you are trying to accomplish a grid of grids, you'll have to setup a separate gmetad server and have a data_source entry for each gmetad server you have. Make sure you put in the corresponding port for each gmetad server and that should be it. Alternatively, can you collapse all your gmetads into one? Cheers, Bernard On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Hi, I have several gmetad headnodes corresponding each one to one grid of several clusters reporting metrics through unicast (meaning that in each of my gmeta headnodes I have several data_sources and several gmond files - one for each cluster). I would like now to aggregate all of these metric in just on MAIN headnode, like a Grid of grids. How can I do this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] FW: one gmetad gathering from multiple gmetad
Hi Alexander, thanks for the hint. It was actually an Apache problem as I suspected. Just gave access to main monitor and now it works fine. Thanks all. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/1-010 CERN From: Alexander Karner [a...@de.ibm.com] Sent: 13 March 2014 09:31 To: Cristovao Jose Domingues Cordeiro Cc: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] FW: one gmetad gathering from multiple gmetad Hi Christovao! You should doublecheck the name resolution on both gmetad systems in both directions. I use a multi-grid setup with actually five child gmetad's of which one failed to establish a connection with the central gmetad. Interestingly the behaviour changed over time: For some time I had to use the short name, then the FQN and then the IP (and back again). So for this specific system I set this child gmetad to all_trusted on but we have communication limited by the firewall. Mit freundlichen Grüßen / Kind regards Alexander Karner Program Manager Health Check Senior Certified IBM Architect Senior IT Specialist Global Technology Services B141 From:Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch To:ganglia-general@lists.sourceforge.net ganglia-general@lists.sourceforge.net, Date:13.03.2014 09:23 Subject:[Ganglia-general] FW: one gmetad gathering from multiple gmetad On the main gmetad I've only set 'all_trusted on' and on the reporting gmetad's I've set the trusted_hosts to have the hostname of the machine runnning the main gmetad. It's just one specific gmetad which fails (when scalable is ON). I think it is an Apache issue. ''' Forbidden You don't have permission to access /ganglia/ on this server. Apache/2.2.15 (Red Hat) Server at myhostname Port 80 ''' Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/1-010 CERN From: Bernard Li [bern...@vanhpc.org] Sent: 12 March 2014 18:15 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Did you set the trusted_hosts option on gmetad.conf? Cheers, Bernard On Wed, Mar 12, 2014 at 7:25 AM, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Ok I've made it work. I didn't know 8651 was the way to go. I only have one issue now, which is, if I keep the scalable option ON, which I want to, I get Permission Denied to one of the remote grids...any idea why? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch] Sent: 11 March 2014 15:02 To: Bernard Li Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Bernard, thanks for the reply. But my question is, in one gmetad I have 5 different data_sources, representing different clusters, and each own has its own port. In the MAIN gmetad (the one gathering the other gmetads), which port do I use for the multicluster one? What do you mean by collapse? I wanted to keep a clean separation of which gmetad I am monitoring... Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 07 March 2014 17:37 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Cristovao: If you are trying to accomplish a grid of grids, you'll have to setup a separate gmetad server and have a data_source entry for each gmetad server you have. Make sure you put in the corresponding port for each gmetad server and that should be it. Alternatively, can you collapse all your gmetads into one? Cheers, Bernard On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Hi, I have several gmetad headnodes corresponding each one to one grid of several clusters reporting metrics through unicast (meaning that in each of my gmeta headnodes I have several data_sources and several gmond files - one for each cluster). I would like now to aggregate all of these metric in just on MAIN headnode, like a Grid of grids. How can I do this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net
Re: [Ganglia-general] one gmetad gathering from multiple gmetad
Ok I've made it work. I didn't know 8651 was the way to go. I only have one issue now, which is, if I keep the scalable option ON, which I want to, I get Permission Denied to one of the remote grids...any idea why? Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Cristovao Jose Domingues Cordeiro [cristovao.corde...@cern.ch] Sent: 11 March 2014 15:02 To: Bernard Li Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Bernard, thanks for the reply. But my question is, in one gmetad I have 5 different data_sources, representing different clusters, and each own has its own port. In the MAIN gmetad (the one gathering the other gmetads), which port do I use for the multicluster one? What do you mean by collapse? I wanted to keep a clean separation of which gmetad I am monitoring... Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 07 March 2014 17:37 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Cristovao: If you are trying to accomplish a grid of grids, you'll have to setup a separate gmetad server and have a data_source entry for each gmetad server you have. Make sure you put in the corresponding port for each gmetad server and that should be it. Alternatively, can you collapse all your gmetads into one? Cheers, Bernard On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.chmailto:cristovao.corde...@cern.ch wrote: Hi, I have several gmetad headnodes corresponding each one to one grid of several clusters reporting metrics through unicast (meaning that in each of my gmeta headnodes I have several data_sources and several gmond files - one for each cluster). I would like now to aggregate all of these metric in just on MAIN headnode, like a Grid of grids. How can I do this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] one gmetad gathering from multiple gmetad
Hi Bernard, thanks for the reply. But my question is, in one gmetad I have 5 different data_sources, representing different clusters, and each own has its own port. In the MAIN gmetad (the one gathering the other gmetads), which port do I use for the multicluster one? What do you mean by collapse? I wanted to keep a clean separation of which gmetad I am monitoring... Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 07 March 2014 17:37 To: Cristovao Jose Domingues Cordeiro Cc: Ganglia [ganglia-general@lists.sourceforge.net] Subject: Re: [Ganglia-general] one gmetad gathering from multiple gmetad Hi Cristovao: If you are trying to accomplish a grid of grids, you'll have to setup a separate gmetad server and have a data_source entry for each gmetad server you have. Make sure you put in the corresponding port for each gmetad server and that should be it. Alternatively, can you collapse all your gmetads into one? Cheers, Bernard On Friday, 7 March 2014, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.chmailto:cristovao.corde...@cern.ch wrote: Hi, I have several gmetad headnodes corresponding each one to one grid of several clusters reporting metrics through unicast (meaning that in each of my gmeta headnodes I have several data_sources and several gmond files - one for each cluster). I would like now to aggregate all of these metric in just on MAIN headnode, like a Grid of grids. How can I do this? Thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- Learn Graph Databases - Download FREE O'Reilly Book Graph Databases is the definitive new guide to graph databases and their applications. Written by three acclaimed leaders in the field, this first edition is now available. Download your free book today! http://p.sf.net/sfu/13534_NeoTech___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] multiple clusters with just one collector
Hi, sorry for the late reply. Oh ok those were the problematic ports. Good to know thanks! Unfortunately and at least for now this monitoring page will only be accessible inside the private network. The only ports open for the outside world are the gmond ones. Sorry. If this changes I'll let you know. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 24 January 2014 19:49 To: Cristovao Jose Domingues Cordeiro Cc: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] multiple clusters with just one collector Yes, 8651 and 8652 are reserved ports for gmetad, you can start from port 9000 to avoid this issue :-) Glad that you got it working... are you guys planning to make your monitoring page public? It would be cool if we can reference it somewhere :-) Cheers, Bernard On Fri, Jan 24, 2014 at 12:47 AM, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Thank you for your replies Bernard and Shekar. Actually in the meanwhile I managed to make it work. For those who might want to do the same here's what I did: · Stop gmond and gmetad, and clean the RRD database in the collector (I was getting unspecified clusters and cached information so this was a clean start) · I have 5 clusters so I've created 5 gmond.conf files - gmond-cluster1.conf, gmond-cluster2.conf...etc. In these files I've set up the cluster name, both udp and tcp ports to the same number, and the hostname in the udp_send_channel to 'localhost' · In gmetad.conf I set up a grid name (not necessary but better to have a cleaner UI) · Still in gmetad.conf I have created 5 data_sources, one for each cluster: data_source Cluster A localhost:portA data_source Cluster B localhost:portB ... Note that these ports and cluster names match those in the gmond files I've created earlier. I had some problems with the ports, some clusters were not appearing if they were using port 8651 for instance, but I just changed the port and they worked...I am not sure why, I don't think this port numbers are restricted...but well, if someone knows why, tell me please · Then I just daemonized the 5 gmonds like this: # deamon gmond -c /etc/ganglia/gmond-clusterA.conf # . · Finally, I just started gmetad and now it's running fine. To add more clusters I just need to, in this order, daemonized the new gmond for that cluster and restart gmetad. Hope this helps someone in the future ;) Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 24 January 2014 04:49 To: Cristovao Jose Domingues Cordeiro Cc: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] multiple clusters with just one collector Hello: To accomplish what you want to do, I believe you have two options: 1) Run multiple gmonds, on the master collector, each one listening on different ports 2) Have one gmond but have multiple udp_recv_channels listening on different ports It might help if you also post your configuration somewhere and reference it here (eg. pastebin or some such). Please let us know if you have any other questions. Cheers, Bernard On Thu, Jan 23, 2014 at 5:08 AM, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Hi, I found some similar posts in this mailing list but unfortunately none can solve my situation. Basically, I want to have just one collector to multiple clusters. As far as I understood I need to have multiple gmonds in my headnode, and so I do. I have in my collector: gmetad.conf: data_source clusterA localhost data_source clusterB localhost:8650 gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports, hosts and cluster names defined, pointing to localhost. I also issued the daemon: gmond -c gmond-clusterA.conf gmond -c gmond-clusterB.conf But this doesn't work. I end up with my web interface just showing an 'unspecified' cluster and 'clusterA'... not clusterB. Any thoughts? thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general -- CenturyLink Cloud: The Leader in Enterprise
Re: [Ganglia-general] multiple clusters with just one collector
Hi Adrian, I saw that the other guys are already helping you so there's nothing that I can add which would help you I think. Nevertheless I must say that I don't have an aggregation gmond, that was the point of my question, not having one. I receive all the metric directly in the gmetad node. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Adrian Sevcenco [adrian.sevce...@cern.ch] Sent: 25 January 2014 13:19 To: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] multiple clusters with just one collector On 01/24/2014 10:47 AM, Cristovao Jose Domingues Cordeiro wrote: Thank you for your replies Bernard and Shekar. Actually in the meanwhile I managed to make it work. For those who might want to do the same here's what I did: · Stop gmond and gmetad, and clean the RRD database in the collector (I was getting unspecified clusters and cached information so this was a clean start) · I have 5 clusters so I've created 5 gmond.conf files - gmond-cluster1.conf, gmond-cluster2.conf...etc. In these files I've set up the cluster name, both udp and tcp ports to the same number, and the hostname in the udp_send_channel to 'localhost' Hi! Is it possible to post your gmond settings? How did you solve the problem that the reporting node (where aggregation gmond sits) will also report the self metrics? My problem is a little different : i try to make a fake gmond for gmetric input for UPS/RAID/Switches metrics. Thanks a lot! Adrian -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] multiple clusters with just one collector
Thank you for your replies Bernard and Shekar. Actually in the meanwhile I managed to make it work. For those who might want to do the same here's what I did: · Stop gmond and gmetad, and clean the RRD database in the collector (I was getting unspecified clusters and cached information so this was a clean start) · I have 5 clusters so I've created 5 gmond.conf files - gmond-cluster1.conf, gmond-cluster2.conf...etc. In these files I've set up the cluster name, both udp and tcp ports to the same number, and the hostname in the udp_send_channel to 'localhost' · In gmetad.conf I set up a grid name (not necessary but better to have a cleaner UI) · Still in gmetad.conf I have created 5 data_sources, one for each cluster: data_source Cluster A localhost:portA data_source Cluster B localhost:portB ... Note that these ports and cluster names match those in the gmond files I've created earlier. I had some problems with the ports, some clusters were not appearing if they were using port 8651 for instance, but I just changed the port and they worked...I am not sure why, I don't think this port numbers are restricted...but well, if someone knows why, tell me please · Then I just daemonized the 5 gmonds like this: # deamon gmond -c /etc/ganglia/gmond-clusterA.conf # . · Finally, I just started gmetad and now it's running fine. To add more clusters I just need to, in this order, daemonized the new gmond for that cluster and restart gmetad. Hope this helps someone in the future ;) Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro From: Bernard Li [bern...@vanhpc.org] Sent: 24 January 2014 04:49 To: Cristovao Jose Domingues Cordeiro Cc: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] multiple clusters with just one collector Hello: To accomplish what you want to do, I believe you have two options: 1) Run multiple gmonds, on the master collector, each one listening on different ports 2) Have one gmond but have multiple udp_recv_channels listening on different ports It might help if you also post your configuration somewhere and reference it here (eg. pastebin or some such). Please let us know if you have any other questions. Cheers, Bernard On Thu, Jan 23, 2014 at 5:08 AM, Cristovao Jose Domingues Cordeiro cristovao.corde...@cern.ch wrote: Hi, I found some similar posts in this mailing list but unfortunately none can solve my situation. Basically, I want to have just one collector to multiple clusters. As far as I understood I need to have multiple gmonds in my headnode, and so I do. I have in my collector: gmetad.conf: data_source clusterA localhost data_source clusterB localhost:8650 gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports, hosts and cluster names defined, pointing to localhost. I also issued the daemon: gmond -c gmond-clusterA.conf gmond -c gmond-clusterB.conf But this doesn't work. I end up with my web interface just showing an 'unspecified' cluster and 'clusterA'... not clusterB. Any thoughts? thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
[Ganglia-general] multiple clusters with just one collector
Hi, I found some similar posts in this mailing list but unfortunately none can solve my situation. Basically, I want to have just one collector to multiple clusters. As far as I understood I need to have multiple gmonds in my headnode, and so I do. I have in my collector: gmetad.conf: data_source clusterA localhost data_source clusterB localhost:8650 gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports, hosts and cluster names defined, pointing to localhost. I also issued the daemon: gmond -c gmond-clusterA.conf gmond -c gmond-clusterB.conf But this doesn't work. I end up with my web interface just showing an 'unspecified' cluster and 'clusterA'... not clusterB. Any thoughts? thanks Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general
Re: [Ganglia-general] multiple clusters with just one collector
Hi Gowtham, thanks but that makes use of two frontends, and I only have one. In that example server B is running in one of the clusters and my clusters have a random number of VM's which only live for a few days, top. My master server needs to be collecting everything from every cluster. I am reading solutions about having several gmonds in the collector, but I am not succeeding. Cumprimentos / Best regards, Cristóvão José Domingues Cordeiro IT Department - 28/1-010 CERN From: Gowtham [sgowt...@mtu.edu] Sent: 23 January 2014 15:04 To: Cristovao Jose Domingues Cordeiro Cc: ganglia-general@lists.sourceforge.net Subject: Re: [Ganglia-general] multiple clusters with just one collector A working example of following the instructions in the link below is here: http://hpcmonitor.it.mtu.edu/ Best regards, g -- Gowtham, PhD HPC Research Scientist, ITS Adj. Asst. Professor, Physics/ECE Michigan Technological University (906) 487/3593 http://it.mtu.edu On Thu, 23 Jan 2014, Gowtham wrote: | | Instructions here might be of use to you: | | http://central6.rocksclusters.org/roll-documentation/ganglia/6.1/x111.html | | Best regards, | g | | -- | Gowtham, PhD | HPC Research Scientist, ITS | Adj. Asst. Professor, Physics/ECE | Michigan Technological University | | (906) 487/3593 | http://it.mtu.edu | | | On Thu, 23 Jan 2014, Cristovao Jose Domingues Cordeiro wrote: | | | Hi, | | | | I found some similar posts in this mailing list but unfortunately none can | | solve my situation. | | | | Basically, I want to have just one collector to multiple clusters. | | As far as I understood I need to have multiple gmonds in my headnode, and so | | I do. | | | | I have in my collector: | | gmetad.conf: | | data_source clusterA localhost | | data_source clusterB localhost:8650 | | | | gmond-clusterA.conf and gmond-clusterB.conf, each one with their ports, | | hosts and cluster names defined, pointing to localhost. | | | | I also issued the daemon: | | gmond -c gmond-clusterA.conf | | gmond -c gmond-clusterB.conf | | | | But this doesn't work. I end up with my web interface just showing an | | 'unspecified' cluster and 'clusterA'... not clusterB. | | | | Any thoughts? thanks | | | | Cumprimentos / Best regards, | | Cristóvão José Domingues Cordeiro | | | | | | -- CenturyLink Cloud: The Leader in Enterprise Cloud Services. Learn Why More Businesses Are Choosing CenturyLink Cloud For Critical Workloads, Development Environments Everything In Between. Get a Quote or Start a Free Trial Today. http://pubads.g.doubleclick.net/gampad/clk?id=119420431iu=/4140/ostg.clktrk ___ Ganglia-general mailing list Ganglia-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/ganglia-general