I'm using the Hadoop FS commands to move files from my local machine into the Hadoop dfs. I'd like a way to force a write to the dfs even if a file of the same name exists. Ideally I'd like to use a "-force" switch or some such; e.g., hadoop dfs -copyFromLocal -force adirectory s3n://wholeinthebucket/
Is there a way to do this or does anyone know if this is in the future Hadoop plans? Thanks John SD