Jay,

What does your job do? Create files directly on HDFS? If so, do you
follow this method?:
http://wiki.apache.org/hadoop/FAQ#Can_I_write_create.2BAC8-write-to_hdfs_files_directly_from_map.2BAC8-reduce_tasks.3F

A local filesystem may not complain if you re-create an existent file.
HDFS' behavior here is different. This simple Python test is what I
mean:
>>> a = open('a', 'w')
>>> a.write('f')
>>> b = open('a', 'w')
>>> b.write('s')
>>> a.close(), b.close()
>>> open('a').read()
's'

Hence it is best to use the FileOutputCommitter framework as detailed
in the mentioned link.

On Mon, Apr 2, 2012 at 7:09 PM, Jay Vyas <jayunit...@gmail.com> wrote:
> Hi guys:
>
> I have a map reduce job that runs normally on local file system from
> eclipse, *but* it fails on HDFS running in psuedo distributed mode.
>
> The exception I see is
>
> *org.apache.hadoop.ipc.RemoteException:
> org.apache.hadoop.hdfs.protocol.AlreadyBeingCreatedException:*
>
>
> Any thoughts on why this might occur in psuedo distributed mode, but not in
> regular file system ?



-- 
Harsh J

Reply via email to