Repository: sqoop Updated Branches: refs/heads/branch-1.4.6 08c98bd7c -> 1c8263393 refs/heads/trunk bbfa59a61 -> eaf03baab
SQOOP-2257: Parquet target for imports with Hive overwrite option does not work (Qian Xu via Jarek Jarcec Cecho) (cherry picked from commit fa8518998f527653587c8f74ebf9990c22842dbb) Project: http://git-wip-us.apache.org/repos/asf/sqoop/repo Commit: http://git-wip-us.apache.org/repos/asf/sqoop/commit/847f3f13 Tree: http://git-wip-us.apache.org/repos/asf/sqoop/tree/847f3f13 Diff: http://git-wip-us.apache.org/repos/asf/sqoop/diff/847f3f13 Branch: refs/heads/branch-1.4.6 Commit: 847f3f13c144051a9ee183265ba8ee1ec84e8e4e Parents: 08c98bd Author: Jarek Jarcec Cecho <[email protected]> Authored: Tue Mar 24 06:43:30 2015 -0700 Committer: Gwen Shapira <[email protected]> Committed: Tue Mar 24 16:52:11 2015 -0700 ---------------------------------------------------------------------- src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java | 4 +++- src/java/org/apache/sqoop/mapreduce/ParquetJob.java | 4 ++-- 2 files changed, 5 insertions(+), 3 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/sqoop/blob/847f3f13/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java ---------------------------------------------------------------------- diff --git a/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java b/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java index 708da5b..bcf6611 100644 --- a/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java +++ b/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java @@ -103,7 +103,9 @@ public class DataDrivenImportJob extends ImportJobBase { // Parquet data records. The import will fail, if schema is invalid. Schema schema = generateAvroSchema(tableName); String uri = getKiteUri(conf, tableName); - ParquetJob.configureImportJob(conf, schema, uri, options.isAppendMode()); + boolean reuseExistingDataset = options.isAppendMode() || + (options.doHiveImport() && options.doOverwriteHiveTable()); + ParquetJob.configureImportJob(conf, schema, uri, reuseExistingDataset); } job.setMapperClass(getMapperClass()); http://git-wip-us.apache.org/repos/asf/sqoop/blob/847f3f13/src/java/org/apache/sqoop/mapreduce/ParquetJob.java ---------------------------------------------------------------------- diff --git a/src/java/org/apache/sqoop/mapreduce/ParquetJob.java b/src/java/org/apache/sqoop/mapreduce/ParquetJob.java index bea74c3..85d88ff 100644 --- a/src/java/org/apache/sqoop/mapreduce/ParquetJob.java +++ b/src/java/org/apache/sqoop/mapreduce/ParquetJob.java @@ -71,9 +71,9 @@ public final class ParquetJob { * {@link org.apache.avro.generic.GenericRecord}. */ public static void configureImportJob(Configuration conf, Schema schema, - String uri, boolean doAppend) throws IOException { + String uri, boolean reuseExistingDataset) throws IOException { Dataset dataset; - if (doAppend) { + if (reuseExistingDataset) { try { dataset = Datasets.load(uri); } catch (DatasetNotFoundException ex) {
