[ https://issues.apache.org/jira/browse/HADOOP-5887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Aaron Kimball updated HADOOP-5887: ---------------------------------- Status: Patch Available (was: Open) > Sqoop should create tables in Hive metastore after importing to HDFS > -------------------------------------------------------------------- > > Key: HADOOP-5887 > URL: https://issues.apache.org/jira/browse/HADOOP-5887 > Project: Hadoop Core > Issue Type: New Feature > Reporter: Aaron Kimball > Assignee: Aaron Kimball > Attachments: HADOOP-5887.2.patch, HADOOP-5887.patch > > > Sqoop (HADOOP-5815) imports tables into HDFS; it is a straightforward > enhancement to then generate a Hive DDL statement to recreate the table > definition in the Hive metastore and move the imported table into the Hive > warehouse directory from its upload target. > This feature enhancement makes this process automatic. An import is performed > with sqoop in the usual way; providing the argument "--hive-import" will > cause it to then issue a CREATE TABLE .. LOAD DATA INTO statement to a Hive > shell. It generates a script file and then attempts to run > "$HIVE_HOME/bin/hive" on it, or failing that, any "hive" on the $PATH; > $HIVE_HOME can be overridden with --hive-home. As a result, no direct linking > against Hive is necessary. > The unit tests provided with this enhancement use a mock implementation of > 'bin/hive' that compares the script it's fed with one from a directory full > of "expected" scripts. The exact script file referenced is controlled via an > environment variable. It doesn't actually load into a proper Hive metastore, > but manual testing has shown that this process works in practice, so the mock > implementation is a reasonable unit testing tool. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.