Hmmm....this thread is very interesting - I didn't know most of the stuff mentioned here. Ted, when you say "copy in the distro" do you need to include the configuration files from the running grid? You don't need to actually start HDFS on this node do you? If I'm following this approach correctly, I would want to have an "xfer server" whose job it is to essentially run dfs -copyFromLocal on all inbound-to-HDFS data. Once I'm certain that my data has copied correctly, I can delete the local files on the xfer server. This is great news, as my current system wastes a lot of time copying data from data acquisition servers to the master node. If I can copy to HDFS directly from ny acquisition servers then I am a happy guy.... Thanks, C G
Ted Dunning <[EMAIL PROTECTED]> wrote: Just copy the hadoop distro directory to the other machine and use whatever command you were using before. A program that uses hadoop just have to have access to all of the nodes across the net. It doesn't assume anything else. On 12/20/07 2:35 PM, "Jeff Eastman" wrote: > .... Can you give me a pointer on how to accomplish this (upload from other > machine)? --------------------------------- Be a better friend, newshound, and know-it-all with Yahoo! Mobile. Try it now.