[ 
https://issues.apache.org/jira/browse/HDFS-12688?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16212035#comment-16212035
 ] 

Shriya Gupta commented on HDFS-12688:
-------------------------------------

I suspected that same asynchronous behaviour when this issue was brought to my 
attention by a user. However, at no point in time am I running two copies of 
the job at the same time. Each time, I have launched my script only after the 
previous execution would end.

I suggested running it multiple times only because the error occurs randomly -- 
it has also shown up at the very first run of the script sometimes.

> HDFS File Not Removed Despite Successful "Moved to .Trash" Message
> ------------------------------------------------------------------
>
>                 Key: HDFS-12688
>                 URL: https://issues.apache.org/jira/browse/HDFS-12688
>             Project: Hadoop HDFS
>          Issue Type: Bug
>          Components: hdfs
>    Affects Versions: 2.6.0
>            Reporter: Shriya Gupta
>            Priority: Critical
>
> Wrote a simple script to delete and create a file and ran it multiple times. 
> However, some executions of the script randomly threw a FileAlreadyExists 
> error while the others succeeded despite successful hdfs dfs -rm command. The 
> script is as below, I have reproduced it in two different environments -- 
> hdfs dfs -ls  /user/shriya/shell_test/
> echo "starting hdfs remove **************" 
> hdfs dfs -rm -r -f /user/shriya/shell_test/wordcountOutput
>  echo "hdfs compeleted!"
> hdfs dfs -ls  /user/shriya/shell_test/
> echo "starting mapReduce*******************************"
> mapred job -libjars 
> /data/home/shriya/shell_test/hadoop-mapreduce-client-jobclient-2.7.1.jar 
> -submit /data/home/shriya/shell_test/wordcountJob.xml
> The message confirming successful move -- 
> 17/10/19 14:49:12 INFO fs.TrashPolicyDefault: Moved: 
> 'hdfs://nameservice1/user/shriya/shell_test/wordcountOutput' to trash at: 
> hdfs://nameservice1/user/shriya/.Trash/Current/user/shriya/shell_test/wordcountOutput1508438952728
> The contents of subsequent -ls after -rm also showed that the file still 
> existed)
> The error I got when my MapReduce job tried to create the file -- 
> 17/10/19 14:50:00 WARN security.UserGroupInformation: 
> PriviledgedActionException as:<REDACTED> (auth:KERBEROS) 
> cause:org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory 
> hdfs://nameservice1/user/shriya/shell_test/wordcountOutput already exists
> Exception in thread "main" 
> org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory 
> hdfs://nameservice1/user/shriya/shell_test/wordcountOutput already exists
>         at 
> org.apache.hadoop.mapred.FileOutputFormat.checkOutputSpecs(FileOutputFormat.java:131)
>         at 
> org.apache.hadoop.mapreduce.JobSubmitter.checkSpecs(JobSubmitter.java:272)
>         at 
> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:143)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1307)
>         at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1304)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.security.auth.Subject.doAs(Subject.java:422)
>         at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1920)
>         at org.apache.hadoop.mapreduce.Job.submit(Job.java:1304)
>         at org.apache.hadoop.mapreduce.tools.CLI.run(CLI.java:315)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70)
>         at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:84)
>         at org.apache.hadoop.mapred.JobClient.main(JobClient.java:1277)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org

Reply via email to