[ 
https://issues.apache.org/jira/browse/HCATALOG-451?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13449146#comment-13449146
 ] 

Francis Liu commented on HCATALOG-451:
--------------------------------------

I'm not up for covering up another bug in another product as much as possible. 
Trunk is not released yet, so there should be no rush getting things to work 
with existing releases. 

The antlr issue is a debt that has to be paid. Doesn't anyone find it weird how 
we're running our unit tests against an incompatible version of pig? We can't 
even run our unit tests against 0.23 since pig 0.8 does not have an 0.23 build. 
Let's a least try shoot for that and worry about adding hacks for compatibility 
when the next release is close. 

As for the pig issue, I can take a crack at it since the pig community is not 
interested.




                
> Partitions are created even when Jobs are aborted
> -------------------------------------------------
>
>                 Key: HCATALOG-451
>                 URL: https://issues.apache.org/jira/browse/HCATALOG-451
>             Project: HCatalog
>          Issue Type: Bug
>          Components: mapreduce
>    Affects Versions: 0.4, 0.5
>         Environment: Hadoop 1.0.2, non-dynamic partitions.
>            Reporter: Mithun Radhakrishnan
>            Assignee: Vandana Ayyalasomayajula
>             Fix For: 0.4.1
>
>         Attachments: HCAT-451-trunk.02.patch, HCATALOG-451.0.patch, 
> HCATALOG-451-branch-0.4.02.patch, HCATALOG-451-branch-0.4.03.patch, 
> HCATALOG-451-branch-0.4.patch
>
>
> If an MR job using HCatOutputFormat fails, and 
> FileOutputCommitterContainer::abortJob() is called, one would expect that 
> partitions aren't created/registered with HCatalog.
> When using dynamic-partitions, one sees that this behaves correctly. But when 
> static-partitions are used, partitions are created regardless of whether the 
> Job succeeded or failed.
> (This manifested as a failure when the job is repeated. The retry-job fails 
> to launch since the partitions already exist from the last failed run.)
> This is a result of bad code in FileOutputCommitter::cleanupJob(), which 
> seems to do an unconditional partition-add. This can be fixed by adding a 
> check for the output directory before adding partitions (in the 
> !dynamicParititoning case), since the directory is removed in abortJob().
> We'll have a patch for this shortly. As an aside, we ought to move the 
> partition-creation into commitJob(), where it logically belongs. cleanupJob() 
> is deprecated and common to both success and failure code paths.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to