GitHub user manishgupta88 opened a pull request:

    https://github.com/apache/carbondata/pull/1368

    [CARBONDATA-1486] Fixed issue of table status updation on insert overwrite 
failure and exception thrown while deletion of stale folders

    Issues Fixed as part of this PR:
    
    1. Fixed issue of HDFS file system throwing exception while deleting a 
non-existent folder. This happens because on multiple run of insert overwrite 
operation, same folder is getting added to the stale folder list even though it 
has been deleted during first time insert overwrite operation run and when the 
same folder is tried to be deleted again HDFS file system throws an IO 
exception.
    
    2. Insert or Load operation should throw exception if:
         a. If insert overwrite is in progress and any other load or insert 
operation is triggered
         b. If load or insert into operation is in progress and insert 
overwrite operation is triggered
    
    3. Fixed issue for table status not getting updated after failure of 
insert/load job. In case of insert overwrite job status in table status file is 
"Overwrite in Progress". In this case no other load for same table can run in 
parallel and if this status is not changes on success of failure jobs, the 
system will still assume that insert overwrite is in progress and fail all next 
insert/load jobs
    
    4. Removed Unused Code

You can merge this pull request into a Git repository by running:

    $ git pull https://github.com/manishgupta88/carbondata carbondata_issues

Alternatively you can review and apply these changes as the patch at:

    https://github.com/apache/carbondata/pull/1368.patch

To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:

    This closes #1368
    
----
commit 9a1baaad8f2ff3ec53c4b546a944bb488aaf0a49
Author: manishgupta88 <[email protected]>
Date:   2017-09-18T12:33:04Z

    Fixed issue of HDFS file system throwing exception while deleting a 
non-existent folder. this happens because on multiple run of insert overwrite 
operation, same folder is getting added to the stale folder list even though it 
has been deleted during first time insert overwrite operation run and when the 
same folder is tried to be deleted again HDFS file system throws an exception.
    
    Insert or Load operation should throw exception if:
    1. If insert overwrite is in progress and any other load or insert 
operation is triggered
    2. If load or insert into operation is in progress and insert overwrite 
operation is triggered
    
    Fixed issue for table status not getting updated after failure of 
insert/load job. In case of insert overwrite job status in table status file is 
"Overwrite in Progress". In this case no other load for same table can run in 
parallel and if this status is not changes on success of failure jobs, the 
system will still assume that insert overwrite is in progress and fail all next 
insert/load jobs
    
    Fixed fortify issue for the category "Identical code for different branches"

----


---

Reply via email to