Great! So what is the interim fix you are implementing
Richa On Mar 1, 2016 4:06 PM, "Biswajit Nayak" <biswa...@altiscale.com> wrote: > Thanks Richa. > > The issue was suppose to be fixed in Hive 0.12 version as per the jira > https://issues.apache.org/jira/browse/HIVE-7164. > > Even raised a ticket in sqoop jira [SQOOP-2840] for this . > > Thanks > Biswa > > > > > > On Tue, Mar 1, 2016 at 9:56 AM, Richa Sharma <mailtorichasha...@gmail.com> > wrote: > >> Hi, >> >> The values should still persist if partition column data type in Hive is >> a string. >> >> I am checking HCatalog documentation for support of int data type in >> partition column. >> >> Cheers >> Richa >> >> On Tue, Mar 1, 2016 at 3:06 PM, Biswajit Nayak <biswa...@altiscale.com> >> wrote: >> >>> Hi Richa, >>> >>> Thats a work around. But how to handle the columns with INT type. >>> Changing the type will be the last option for me. >>> >>> Regards >>> Biswa >>> >>> >>> >>> On Tue, Mar 1, 2016 at 9:31 AM, Richa Sharma < >>> mailtorichasha...@gmail.com> wrote: >>> >>>> Hi Biswajit >>>> >>>> The answer is in the last line of the error message. Change the data >>>> type of partition column to string in hive and try again. >>>> >>>> Hope it helps ! >>>> >>>> Richa >>>> >>>> 16/02/12 08:04:12 ERROR tool.ExportTool: Encountered IOException running >>>> export job: java.io.IOException: The table provided default.emp_details1 >>>> uses unsupported partitioning key type for column salary : int. Only >>>> string fields are allowed in partition columns in Catalog >>>> >>>> >>>> On Tue, Mar 1, 2016 at 2:19 PM, Biswajit Nayak <biswa...@altiscale.com> >>>> wrote: >>>> >>>>> Hi All, >>>>> >>>>> I am trying to do a SQOOP export from hive( integer type partition) to >>>>> mysql through HCAT and it fails with the following error. >>>>> >>>>> Versions:- >>>>> >>>>> Hadoop :- 2.7.1 >>>>> Hive :- 1.2.0 >>>>> Sqoop :- 1.4.5 >>>>> >>>>> Table in Hive :- >>>>> >>>>> >>>>> hive> use default; >>>>> OK >>>>> Time taken: 0.028 seconds >>>>> hive> describe emp_details1; >>>>> OK >>>>> id int >>>>> name string >>>>> deg string >>>>> dept string >>>>> salary int >>>>> >>>>> # Partition Information >>>>> # col_name data_type comment >>>>> >>>>> salary int >>>>> Time taken: 0.125 seconds, Fetched: 10 row(s) >>>>> hive> >>>>> >>>>> hive> select * from emp_details1; >>>>> OK >>>>> 1201 gopal 50000 >>>>> 1202 manisha 50000 >>>>> 1203 kalil 50000 >>>>> 1204 prasanth 50000 >>>>> 1205 kranthi 50000 >>>>> 1206 satish 50000 >>>>> Time taken: 0.195 seconds, Fetched: 6 row(s) >>>>> hive> >>>>> >>>>> >>>>> Conf added to Hive metastore site.xml >>>>> >>>>> >>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$ grep -A5 -B2 -i >>>>> "hive.metastore.integral.jdo.pushdown" /etc/hive-metastore/hive-site.xml >>>>> </property> >>>>> <property> >>>>> <name>hive.metastore.integral.jdo.pushdown</name> >>>>> <value>TRUE</value> >>>>> </property> >>>>> >>>>> </configuration> >>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$ >>>>> >>>>> >>>>> The issue remains same >>>>> >>>>> >>>>> [alti-test-01@hdpnightly271-ci-91-services ~]$ /opt/sqoop-1.4.5/bin/sqoop >>>>> export --connect jdbc:mysql://localhost:3306/test --username hive >>>>> --password ********* --table employee --hcatalog-database default >>>>> --hcatalog-table emp_details1 >>>>> Warning: /opt/sqoop-1.4.5/bin/../../hbase does not exist! HBase imports >>>>> will fail. >>>>> Please set $HBASE_HOME to the root of your HBase installation. >>>>> Warning: /opt/sqoop-1.4.5/bin/../../accumulo does not exist! Accumulo >>>>> imports will fail. >>>>> Please set $ACCUMULO_HOME to the root of your Accumulo installation. >>>>> Warning: /opt/sqoop-1.4.5/bin/../../zookeeper does not exist! Accumulo >>>>> imports will fail. >>>>> Please set $ZOOKEEPER_HOME to the root of your Zookeeper installation. >>>>> 16/02/12 08:04:00 INFO sqoop.Sqoop: Running Sqoop version: 1.4.5 >>>>> 16/02/12 08:04:00 WARN tool.BaseSqoopTool: Setting your password on the >>>>> command-line is insecure. Consider using -P instead. >>>>> 16/02/12 08:04:00 INFO manager.MySQLManager: Preparing to use a MySQL >>>>> streaming resultset. >>>>> 16/02/12 08:04:00 INFO tool.CodeGenTool: Beginning code generation >>>>> 16/02/12 08:04:01 INFO manager.SqlManager: Executing SQL statement: >>>>> SELECT t.* FROM `employee` AS t LIMIT 1 >>>>> 16/02/12 08:04:01 INFO manager.SqlManager: Executing SQL statement: >>>>> SELECT t.* FROM `employee` AS t LIMIT 1 >>>>> 16/02/12 08:04:01 INFO orm.CompilationManager: HADOOP_MAPRED_HOME is >>>>> /opt/hadoop >>>>> Note: >>>>> /tmp/sqoop-alti-test-01/compile/1b0d4b1c30f167eb57ef488232ab49c8/employee.java >>>>> uses or overrides a deprecated API. >>>>> Note: Recompile with -Xlint:deprecation for details. >>>>> 16/02/12 08:04:07 INFO orm.CompilationManager: Writing jar file: >>>>> /tmp/sqoop-alti-test-01/compile/1b0d4b1c30f167eb57ef488232ab49c8/employee.jar >>>>> 16/02/12 08:04:07 INFO mapreduce.ExportJobBase: Beginning export of >>>>> employee >>>>> 16/02/12 08:04:08 INFO mapreduce.ExportJobBase: Configuring HCatalog for >>>>> export job >>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Configuring HCatalog >>>>> specific details for job >>>>> 16/02/12 08:04:08 INFO manager.SqlManager: Executing SQL statement: >>>>> SELECT t.* FROM `employee` AS t LIMIT 1 >>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Database column names >>>>> projected : [id, name, deg, salary, dept] >>>>> 16/02/12 08:04:08 INFO hcat.SqoopHCatUtilities: Database column name - >>>>> info map : >>>>> id : [Type : 4,Precision : 11,Scale : 0] >>>>> name : [Type : 12,Precision : 20,Scale : 0] >>>>> deg : [Type : 12,Precision : 20,Scale : 0] >>>>> salary : [Type : 4,Precision : 11,Scale : 0] >>>>> dept : [Type : 12,Precision : 10,Scale : 0] >>>>> >>>>> 16/02/12 08:04:10 INFO hive.metastore: Trying to connect to metastore >>>>> with URI thrift://hive-hdpnightly271-ci-91.test.altiscale.com:9083 >>>>> 16/02/12 08:04:10 INFO hive.metastore: Connected to metastore. >>>>> 16/02/12 08:04:11 INFO hcat.SqoopHCatUtilities: HCatalog full table >>>>> schema fields = [id, name, deg, dept, salary] >>>>> 16/02/12 08:04:12 ERROR tool.ExportTool: Encountered IOException running >>>>> export job: java.io.IOException: The table provided default.emp_details1 >>>>> uses unsupported partitioning key type for column salary : int. Only >>>>> string fields are allowed in partition columns in Catalog >>>>> >>>>> >>>>> Stuck with this issue. Any one had conquered this before. >>>>> >>>>> Regards >>>>> Biswa >>>>> >>>>> >>>> >>> >> >