On Oct 16, 2010, at 1:08 PM, Bruce Williams wrote:
> I am doing a student Independent Study  Project and Harvery Mudd has given
> me 13 Sun Netra X1 I can use as a dedicated Hadoop cluster. Right now they
> are without an OS.
> 
> If anyone with experience with Hadoop and Solaris can contact me off list,
> even to just say I am doing it and it is OK it would be appreciated.



        That's my cue! :)

        We have a few grids that are running Solaris. It mostly works out of 
the box as long as you are aware of three things:

                - There are some settings in hadoop-env.sh and in the path that 
need to be dealt with.  Rather than re-quote, these were added to the Hadoop 
FAQ a week or two ago so definitely take a look at that.

                - The native compression libraries will need to be compiled.  
Depending upon what you are doing/how performant the machines are, this may or 
may not make a big difference.  Compiling under Solaris with gcc will work fine 
(but it is gcc... ugh!).  Only a few minor changes are required to compile it 
with SUNWspro.  [I have patches laying around here somewhere if anyone wants to 
play with them.]

                - The Solaris JRE is a mixed-mode implementation.  So keep in 
mind that -d32 and -d64 have meaning and do work as advertised.  You'll likely 
want to pick a bitsize and use that for all your Hadoop daemons and tasks, 
especially if you plan on using any JNI like the compression libraries.

Reply via email to