I checked for the usages of df in Hadoop for you. DF is used by many
parts of the system including dfs and mapred, but it will not run on a
node that does not run TaskTracker of Datanode. Could you please check
your configuration if you're missing smt. Can you confirm that you just
only submit the job from the windows machine?
Tim Gautier wrote:
Thanks for the reply, but I think you missed my point. I've been
running nutch and Hadoop through eclipse under Windows for several
months now. It never called Linux shell commands before, now it is.
Maybe it called df from some path I never hit, I don't know. What I
do know is that I could do an entire Nutch crawl in eclipse on Windows
with Hadoop 0.13 and I can't even inject with the latest version in
trunk which is using Hadoop 0.15 because it calls df. I realize I can
run it from the command line under cygwin, I just don't want to, I
want to use eclipse like I have been for months.
I feel like I'm missing something simple, but I can't figure out what
it is. Anyone else have any ideas?
On Nov 9, 2007 7:00 AM, Enis Soztutar <[EMAIL PROTECTED]> wrote:
Hadoop has been running df for a long time way before 0.13. You can run
hadoop under cygwin ın windows. Please refer to Hadoop's documentation.
Tim Gautier wrote:
I do my nutch development and debugging on a Windows XP machine before
transferring my jar files to a Linux cluster for actual production
runs. This has worked fine in the past, but there seems to be a
problem now that we're using Hadoop .15. Now when I run the injector
(presumably other classes as well) from Eclipse, Hadoop makes a shell
command call to "df" and of course doesn't find it on a Windows
machine so the job fails. There has to be a way around this so I can
debug from Eclipse, anybody know what that might be? Configuration
setting or something to tell it that I'm running on Windows?