Repository: sqoop
Updated Branches:
  refs/heads/trunk 41724c31d -> e23d1571b


SQOOP-1494: Fix generateArvoSchema in DataDrivenImportJob

(Qian Xu via Jarek Jarcec Cecho)


Project: http://git-wip-us.apache.org/repos/asf/sqoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/sqoop/commit/e23d1571
Tree: http://git-wip-us.apache.org/repos/asf/sqoop/tree/e23d1571
Diff: http://git-wip-us.apache.org/repos/asf/sqoop/diff/e23d1571

Branch: refs/heads/trunk
Commit: e23d1571bd24d79583b832d22658fe2d557de65f
Parents: 41724c3
Author: Jarek Jarcec Cecho <[email protected]>
Authored: Fri Sep 5 08:44:56 2014 +0200
Committer: Jarek Jarcec Cecho <[email protected]>
Committed: Fri Sep 5 08:44:56 2014 +0200

----------------------------------------------------------------------
 src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java | 6 +++---
 1 file changed, 3 insertions(+), 3 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/sqoop/blob/e23d1571/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java
----------------------------------------------------------------------
diff --git a/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java 
b/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java
index 19ec542..938948f 100644
--- a/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java
+++ b/src/java/org/apache/sqoop/mapreduce/DataDrivenImportJob.java
@@ -88,7 +88,7 @@ public class DataDrivenImportJob extends ImportJobBase {
       job.setOutputValueClass(NullWritable.class);
     } else if (options.getFileLayout()
         == SqoopOptions.FileLayout.AvroDataFile) {
-      Schema schema = generateArvoSchema(tableName);
+      Schema schema = generateAvroSchema(tableName);
       try {
         writeAvroSchema(schema);
       } catch (final IOException e) {
@@ -101,7 +101,7 @@ public class DataDrivenImportJob extends ImportJobBase {
       Configuration conf = job.getConfiguration();
       // An Avro schema is required for creating a dataset that manages
       // Parquet data records. The import will fail, if schema is invalid.
-      Schema schema = generateArvoSchema(tableName);
+      Schema schema = generateAvroSchema(tableName);
       String uri;
       if (options.doHiveImport()) {
         uri = "dataset:hive?dataset=" + options.getTableName();
@@ -115,7 +115,7 @@ public class DataDrivenImportJob extends ImportJobBase {
     job.setMapperClass(getMapperClass());
   }
 
-  private Schema generateArvoSchema(String tableName) throws IOException {
+  private Schema generateAvroSchema(String tableName) throws IOException {
     ConnManager connManager = getContext().getConnManager();
     AvroSchemaGenerator generator = new AvroSchemaGenerator(options,
         connManager, tableName);

Reply via email to