If the path is for a file in HDFS, I have had problems where the
hadoop tasks think the path is a local path, not an HDFS path. I think
this happened to me because my HADOOP_CONF_PATH and HADOOP_PATH
weren't set to a place that had HDFS as default. If that doesn't work,
you can always try the fully qualified path
(hdfs://server:port/path/to/file).
On 05/23/2011 04:59 PM, Keith Thompson wrote:
I am trying to run the arff.vector algorithm but am getting a File Not Found
Exception on my input directory. I have checked and double-checked. The
input path is definitely correct. Someone suggested it could be a problem
that my Hadoop folder and my Mahout folder aren't in the same folder, but I
was able to run the Dirchlet clustering algorithm successfully specifying
both the --input and --output paths so I don't think that is the problem.
What other problems could cause this error to pop up?