GitHub user manishgupta88 opened a pull request:
https://github.com/apache/carbondata/pull/1368
[CARBONDATA-1486] Fixed issue of table status updation on insert overwrite
failure and exception thrown while deletion of stale folders
Issues Fixed as part of this PR:
1. Fixed issue of HDFS file system throwing exception while deleting a
non-existent folder. This happens because on multiple run of insert overwrite
operation, same folder is getting added to the stale folder list even though it
has been deleted during first time insert overwrite operation run and when the
same folder is tried to be deleted again HDFS file system throws an IO
exception.
2. Insert or Load operation should throw exception if:
a. If insert overwrite is in progress and any other load or insert
operation is triggered
b. If load or insert into operation is in progress and insert
overwrite operation is triggered
3. Fixed issue for table status not getting updated after failure of
insert/load job. In case of insert overwrite job status in table status file is
"Overwrite in Progress". In this case no other load for same table can run in
parallel and if this status is not changes on success of failure jobs, the
system will still assume that insert overwrite is in progress and fail all next
insert/load jobs
4. Removed Unused Code
You can merge this pull request into a Git repository by running:
$ git pull https://github.com/manishgupta88/carbondata carbondata_issues
Alternatively you can review and apply these changes as the patch at:
https://github.com/apache/carbondata/pull/1368.patch
To close this pull request, make a commit to your master/trunk branch
with (at least) the following in the commit message:
This closes #1368
----
commit 9a1baaad8f2ff3ec53c4b546a944bb488aaf0a49
Author: manishgupta88 <[email protected]>
Date: 2017-09-18T12:33:04Z
Fixed issue of HDFS file system throwing exception while deleting a
non-existent folder. this happens because on multiple run of insert overwrite
operation, same folder is getting added to the stale folder list even though it
has been deleted during first time insert overwrite operation run and when the
same folder is tried to be deleted again HDFS file system throws an exception.
Insert or Load operation should throw exception if:
1. If insert overwrite is in progress and any other load or insert
operation is triggered
2. If load or insert into operation is in progress and insert overwrite
operation is triggered
Fixed issue for table status not getting updated after failure of
insert/load job. In case of insert overwrite job status in table status file is
"Overwrite in Progress". In this case no other load for same table can run in
parallel and if this status is not changes on success of failure jobs, the
system will still assume that insert overwrite is in progress and fail all next
insert/load jobs
Fixed fortify issue for the category "Identical code for different branches"
----
---