Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change 
notification.

The following page has been changed by andyk:
http://wiki.apache.org/hadoop/Chukwa_Quick_Start

------------------------------------------------------------------------------
   1. In the chukwa root directory, say ``bash bin/processSinkFiles.sh'' 
   1. (Document TODO: This script has a hard-coded 54310 port.  Can you confirm 
that you must be running hdfs from 54310?)
  
- == Running Chukwa -- Networked ==
+ == Running Chukwa on a Cluster ==
+ The cluster deployment process is still under active development, thus it is 
possible that the following instructions may not work yet, but they will soon, 
so please don't delete them. Eventually, even the single machine setup (for 
newcomers to Chukwa who want to try it out of the box on their) above will be 
replaced by the below process, renaming the conf/slaves.template and 
conf/collectors.template files (to remove the .template suffix) for the 
defaults of localhost for the collector and agent.
  
- Running Chukwa in a networked context is essentially similar to the 
single-machine deployment discussed above.  However, in a network context, you 
would also need to tell the local agent where the collector[s] live, by listing 
them in conf/collectors.
+ 1. Like in Hadoop, you can specify a set of nodes on which you want to run 
Chukwa agents (similar to conf/slaves in Hadoop) using a conf/slaves file.
  
+ 2. Similarly, collectors should be specified using the conf/collectors file. 
These can be run using bin/start-collectors.sh
+ 
+ 3. The local agents on each machine will also reference the conf/collectors 
file, selecting a collector at random from this list to talk to. Thus, like 
Hadoop, it is common to run Chukwa from a shared file system where all of the 
agents (i.e. slaves) can access the same conf files.
+ 

Reply via email to