Yes, you would have to distribute ruby (though. it's typically installed by default) as well as the wukong and json libraries to all the nodes in the cluster. Unfortunately this isn't something wukong gives you for free at the moment though it is planned.

As far as I know Pig doesn't do anything more complex than launch a hadoop streaming job and use the output in the subsequent steps

btw I write 90% of my mr jobs using either wukong or Pig. Only when it's absolutely required do I use a language with as much overhead as java :)

--jacob
@thedatachef

Sent from my iPhone

On Jan 30, 2011, at 2:09 PM, Alex McLintock <[email protected]> wrote:

On 29 January 2011 13:43, Jacob Perkins <[email protected]> wrote:

Write a map only wukong script that parses the json as you want it. See
the example here:


http://thedatachef.blogspot.com/2011/01/processing-json-records-with-hadoop-and.html


Hi Jacob,

Thanks very much for helping me out. I haven't heard of Wukong before.
I am a bit concerned though by adding Ruby into my tool stack as well as
Pig. It seems like a step too far.
Presumably I have to distribute Ruby and Wukong across all my job nodes in
the same way as if I were writing perl or C++ streaming programs.

With STREAMing - the script is launched once per file, right, not once per
record?

Alex

Reply via email to