Hi all- I am having trouble with SequenceFileOutputFormat.getReaders on a hadoop 17.2 cluster. I am trying to open a set of SequenceFiles that was created in one map process that has completed from within a second map process by passing in the job configuration for the running map process (not of the map process that created the set of sequence files) and the path to the output. When I run locally, this works fine, but when I run remotely on the cluster (using HDFS on the cluster), I get the following IOException:
java.io.IOException: Cannot open filename /user/redpony/Model1.data.0/_logs However, the following works: hadoop dfs -ls /user/redpony/Model1.data.0/_logs Found 1 items /user/redpony/Model1.data.0/_logs/history <dir> 2008-09-18 00:43 rwxrwxrwx redpony supergroup This is probably something dumb, and quite likely related to me not having my settings configured properly, but I'm completely at a loss for how to proceed. Any ideas? Thanks! Chris
