Hi Mengwei After seeing my table schema, do you have any advanced idea about this problem?
On Fri, Sep 6, 2013 at 9:02 AM, Yanting Chen <[email protected]> wrote: > TABLE DS_MSG_LOG > ( > MESSAGE_ID VARCHAR2(23) NOT NULL, > LOGIN_ID VARCHAR2(30), > ACPT_DTS TIMESTAMP(6), > DLVR_DTS TIMESTAMP(6), > SENDER_ID VARCHAR2(30), > SENDER_VAC_ID VARCHAR2(39), > RECEIVER_ID VARCHAR2(30), > RECEIVER_VAC_ID VARCHAR2(39), > STATUS VARCHAR2(1), > MESSAGE_TYPE VARCHAR2(8), > FLOW_TYPE VARCHAR2(5), > SERVICE_TYPE VARCHAR2(1), > SOURCE_FILE_NAME VARCHAR2(150), > ARCHIVE_FILE_NAME VARCHAR2(250), > ARCHIVE_CHAR_COUNT NUMBER, > DECRYPT_FILE_NAME VARCHAR2(250), > DECRYPT_CHAR_COUNT NUMBER, > RESP_FILE_NAME VARCHAR2(250), > RESP_CHAR_COUNT NUMBER, > RESP_FLAG VARCHAR2(1), > RTG_SEQ VARCHAR2(8), > RESENT_FLAG VARCHAR2(1) DEFAULT 'N', > TOTAL_INV_COUNT NUMBER, > CONSTRAINT PK_DS_MSG_LOG PRIMARY KEY (MESSAGE_ID) > ) > > > On Fri, Sep 6, 2013 at 1:55 AM, Abraham Elmahrek <[email protected]> wrote: > >> Could you provide your schema from PostGreSQL? Mengwei is likely right. >> >> >> On Wed, Sep 4, 2013 at 7:36 PM, Yanting Chen <[email protected]>wrote: >> >>> Actually the schema comes from Oracle. However, I try to modify it to >>> let it fit into PostgreSQL database. So, now the database I use is >>> PostgreSQL. >>> >>> >>> On Thu, Sep 5, 2013 at 10:33 AM, Abraham Elmahrek <[email protected]>wrote: >>> >>>> Yanting, >>>> >>>> Also, it seems like the schema you've provided is for an Oracle >>>> database. i.e. VARCHAR2 and NUMBER are datatypes specific to Oracle. Could >>>> you please use an oracle connection string and driver? i.e. >>>> oracle.jdbc.driver.OracleDriver >>>> and jdbc:oracle:thin:@host:port:SID. >>>> >>>> -abe >>>> >>>> >>>> On Wed, Sep 4, 2013 at 7:32 PM, Mengwei Ding <[email protected]>wrote: >>>> >>>>> Hmm... would you mind showing us your most updated job configuration >>>>> by typing "show job --jid 3"? I just want to make sure that you provide >>>>> the >>>>> partition column correctly. >>>>> >>>>> Also, I notice that the primary key for this table is "VARCHAR(23)" >>>>> type, this might be the problem. >>>>> >>>>> Mengwei >>>>> >>>>> >>>>> On Wed, Sep 4, 2013 at 10:23 PM, Yanting Chen >>>>> <[email protected]>wrote: >>>>> >>>>>> Hi Mengwei >>>>>> >>>>>> I try to set primary key into partition column. But still the same >>>>>> error! >>>>>> >>>>>> >>>>>> On Thu, Sep 5, 2013 at 10:17 AM, Mengwei Ding <[email protected] >>>>>> > wrote: >>>>>> >>>>>>> Hi Yanting, >>>>>>> >>>>>>> It seems like you did not specify the 'partition column' for the >>>>>>> job. Generally, the primary key of the table is a good choice for >>>>>>> 'parition column'. >>>>>>> >>>>>>> You could use 'update job -jid 3' to update the job configuration. >>>>>>> >>>>>>> Mengwei >>>>>>> >>>>>>> >>>>>>> On Wed, Sep 4, 2013 at 9:37 PM, Yanting Chen <[email protected] >>>>>>> > wrote: >>>>>>> >>>>>>>> Abraham, >>>>>>>> >>>>>>>> Thanks for you answer. I review my database. >>>>>>>> I think the database name is invoice and the schema name is public, >>>>>>>> just like the picture below. >>>>>>>> http://imgur.com/ns0iNLi >>>>>>>> >>>>>>>> So, I change the schema name with a new value, "public" >>>>>>>> Then, I run this job and get a different error. >>>>>>>> >>>>>>>> Status: FAILURE_ON_SUBMIT >>>>>>>> Creation date: 2013-09-05 09:30:44 CST >>>>>>>> Last update date: 2013-09-05 09:30:44 CST >>>>>>>> Exception: org.apache.sqoop.common.SqoopException: >>>>>>>> GENERIC_JDBC_CONNECTOR_0011:The type is not supported - 12 >>>>>>>> Stack trace: org.apache.sqoop.common.SqoopException: >>>>>>>> GENERIC_JDBC_CONNECTOR_0011:The type is not supported - 12 >>>>>>>> at >>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportPartitioner.getPartitions(GenericJdbcImportPartitioner.java:87) >>>>>>>> at >>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportPartitioner.getPartitions(GenericJdbcImportPartitioner.java:32) >>>>>>>> at >>>>>>>> org.apache.sqoop.job.mr.SqoopInputFormat.getSplits(SqoopInputFormat.java:71) >>>>>>>> at >>>>>>>> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:452) >>>>>>>> at >>>>>>>> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:469) >>>>>>>> at >>>>>>>> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:366) >>>>>>>> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1269) >>>>>>>> at org.apache.hadoop.mapreduce.Job$11.run(Job.java:1266) >>>>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>>>>>> at >>>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1408) >>>>>>>> at org.apache.hadoop.mapreduce.Job.submit(Job.java:1266) >>>>>>>> at >>>>>>>> org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine.submit(MapreduceSubmissionEngine.java:265) >>>>>>>> at >>>>>>>> org.apache.sqoop.framework.FrameworkManager.submit(FrameworkManager.java:480) >>>>>>>> at >>>>>>>> org.apache.sqoop.handler.SubmissionRequestHandler.submissionSubmit(SubmissionRequestHandler.java:112) >>>>>>>> at >>>>>>>> org.apache.sqoop.handler.SubmissionRequestHandler.handleActionEvent(SubmissionRequestHandler.java:98) >>>>>>>> at >>>>>>>> org.apache.sqoop.handler.SubmissionRequestHandler.handleEvent(SubmissionRequestHandler.java:68) >>>>>>>> at >>>>>>>> org.apache.sqoop.server.v1.SubmissionServlet.handlePostRequest(SubmissionServlet.java:44) >>>>>>>> at >>>>>>>> org.apache.sqoop.server.SqoopProtocolServlet.doPost(SqoopProtocolServlet.java:63) >>>>>>>> at javax.servlet.http.HttpServlet.service(HttpServlet.java:637) >>>>>>>> at javax.servlet.http.HttpServlet.service(HttpServlet.java:717) >>>>>>>> at >>>>>>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290) >>>>>>>> at >>>>>>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) >>>>>>>> at >>>>>>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) >>>>>>>> at >>>>>>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) >>>>>>>> at >>>>>>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) >>>>>>>> at >>>>>>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) >>>>>>>> at >>>>>>>> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) >>>>>>>> at >>>>>>>> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293) >>>>>>>> at >>>>>>>> org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) >>>>>>>> at >>>>>>>> org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:602) >>>>>>>> at >>>>>>>> org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) >>>>>>>> at java.lang.Thread.run(Thread.java:724) >>>>>>>> >>>>>>>> Also, I try to remove the schema name, but I get the same error as >>>>>>>> above. >>>>>>>> >>>>>>>> On Thu, Sep 5, 2013 at 6:11 AM, Abraham Elmahrek <[email protected]> >>>>>>>> wrote: >>>>>>>> >>>>>>>> Yanting, >>>>>>>>> >>>>>>>>> I'm sorry, I'm a bit confused. The database you are using here is >>>>>>>>> called "invoice". Do you also have a schema called "invoice"? Try >>>>>>>>> removing >>>>>>>>> the "Schema name" from your job and see if that works for you. >>>>>>>>> >>>>>>>>> http://www.postgresql.org/docs/9.1/static/ddl-schemas.html >>>>>>>>> >>>>>>>>> -Abe >>>>>>>>> >>>>>>>>> >>>>>>>>> On Tue, Sep 3, 2013 at 6:34 PM, Yanting Chen < >>>>>>>>> [email protected]> wrote: >>>>>>>>> >>>>>>>>>> However, with Sqoop 1.4.4, all schema and table names are >>>>>>>>>> lower-case, and it is successful to make import. >>>>>>>>>> >>>>>>>>>> As a result, I am not sure if case sensitivity will be a problem >>>>>>>>>> in Sqoop 2. >>>>>>>>>> >>>>>>>>>> *This is the connection I created:* >>>>>>>>>> 1 connection(s) to show: >>>>>>>>>> Connection with id 1 and name ds_msg_log (Created 9/2/13 10:54 >>>>>>>>>> AM, Updated 9/3/13 6:35 PM) >>>>>>>>>> Using Connector id 1 >>>>>>>>>> Connection configuration >>>>>>>>>> JDBC Driver Class: org.postgresql.Driver >>>>>>>>>> JDBC Connection String: >>>>>>>>>> jdbc:postgresql://MY_IP:MY_PORT/invoice >>>>>>>>>> Username: hadoop >>>>>>>>>> Password: >>>>>>>>>> JDBC Connection Properties: >>>>>>>>>> Security related configuration options >>>>>>>>>> Max connections: 123 >>>>>>>>>> >>>>>>>>>> *This is the job I created:* >>>>>>>>>> Job with id 3 and name ds_msg_log (Created 9/2/13 11:37 AM, >>>>>>>>>> Updated 9/3/13 5:57 PM) >>>>>>>>>> Using Connection id 1 and Connector id 1 >>>>>>>>>> Database configuration >>>>>>>>>> Schema name: invoice >>>>>>>>>> Table name: ds_msg_log >>>>>>>>>> Table SQL statement: >>>>>>>>>> Table column names: >>>>>>>>>> Partition column name: >>>>>>>>>> Boundary query: >>>>>>>>>> Output configuration >>>>>>>>>> Storage type: HDFS >>>>>>>>>> Output format: TEXT_FILE >>>>>>>>>> Output directory: /user/root/ds_msg_log_sqoop2 >>>>>>>>>> Throttling resources >>>>>>>>>> Extractors: >>>>>>>>>> Loaders: >>>>>>>>>> >>>>>>>>>> *The command I use to start this job is:* >>>>>>>>>> submission start --jid 3 >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Wed, Sep 4, 2013 at 9:24 AM, Abraham Elmahrek < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Sorry, I missed that you are using PostGreSQL. I think case >>>>>>>>>>> sensitivity will be a problem with that database. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Tue, Sep 3, 2013 at 6:23 PM, Abraham Elmahrek < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> What database are you using? Are table names or schema names >>>>>>>>>>>> case sensitive in your database? Sqoop quotes all names, so case >>>>>>>>>>>> sensitivity makes a difference here. >>>>>>>>>>>> >>>>>>>>>>>> Also, could you provide the corresponding Sqoop 1.4.4 command >>>>>>>>>>>> that worked for you? >>>>>>>>>>>> >>>>>>>>>>>> -Abe >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Tue, Sep 3, 2013 at 6:06 PM, 陳彥廷(Yanting Chen) < >>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> I am pretty sure that the schema "invoice" exists because I >>>>>>>>>>>>> successfully imported data to this schema using Sqoop 1.4.4 >>>>>>>>>>>>> >>>>>>>>>>>>> Also, I already remove "*" when creating job. >>>>>>>>>>>>> >>>>>>>>>>>>> The following is the table in invoice schema: >>>>>>>>>>>>> TABLE DS_MSG_LOG >>>>>>>>>>>>> ( >>>>>>>>>>>>> MESSAGE_ID VARCHAR2(23) NOT NULL, >>>>>>>>>>>>> LOGIN_ID VARCHAR2(30), >>>>>>>>>>>>> ACPT_DTS TIMESTAMP(6), >>>>>>>>>>>>> DLVR_DTS TIMESTAMP(6), >>>>>>>>>>>>> SENDER_ID VARCHAR2(30), >>>>>>>>>>>>> SENDER_VAC_ID VARCHAR2(39), >>>>>>>>>>>>> RECEIVER_ID VARCHAR2(30), >>>>>>>>>>>>> RECEIVER_VAC_ID VARCHAR2(39), >>>>>>>>>>>>> STATUS VARCHAR2(1), >>>>>>>>>>>>> MESSAGE_TYPE VARCHAR2(8), >>>>>>>>>>>>> FLOW_TYPE VARCHAR2(5), >>>>>>>>>>>>> SERVICE_TYPE VARCHAR2(1), >>>>>>>>>>>>> SOURCE_FILE_NAME VARCHAR2(150), >>>>>>>>>>>>> ARCHIVE_FILE_NAME VARCHAR2(250), >>>>>>>>>>>>> ARCHIVE_CHAR_COUNT NUMBER, >>>>>>>>>>>>> DECRYPT_FILE_NAME VARCHAR2(250), >>>>>>>>>>>>> DECRYPT_CHAR_COUNT NUMBER, >>>>>>>>>>>>> RESP_FILE_NAME VARCHAR2(250), >>>>>>>>>>>>> RESP_CHAR_COUNT NUMBER, >>>>>>>>>>>>> RESP_FLAG VARCHAR2(1), >>>>>>>>>>>>> RTG_SEQ VARCHAR2(8), >>>>>>>>>>>>> RESENT_FLAG VARCHAR2(1) DEFAULT 'N', >>>>>>>>>>>>> TOTAL_INV_COUNT NUMBER, >>>>>>>>>>>>> CONSTRAINT PK_DS_MSG_LOG PRIMARY KEY (MESSAGE_ID) >>>>>>>>>>>>> ) >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Sep 4, 2013 at 1:46 AM, Abraham Elmahrek < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hey User, >>>>>>>>>>>>>> >>>>>>>>>>>>>> It looks like the schema "invoice" does not exist in your >>>>>>>>>>>>>> database. Could you please provide your database schema? Also, >>>>>>>>>>>>>> "*" is >>>>>>>>>>>>>> unnecessary when specifying "Table column names". If you leave >>>>>>>>>>>>>> it blank it >>>>>>>>>>>>>> will import all columns by default. >>>>>>>>>>>>>> >>>>>>>>>>>>>> -Abe >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Tue, Sep 3, 2013 at 3:03 AM, 陳彥廷(Yanting Chen) < >>>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> According to the official guide, >>>>>>>>>>>>>>> http://sqoop.apache.org/docs/1.99.2/Sqoop5MinutesDemo.html , >>>>>>>>>>>>>>> I successfully created a job. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> However, when I executed the command, submission start --jid >>>>>>>>>>>>>>> 1, I got this error message: "Exception has occurred during >>>>>>>>>>>>>>> processing >>>>>>>>>>>>>>> command Server has returned exception: Exception: >>>>>>>>>>>>>>> java.lang.Throwable >>>>>>>>>>>>>>> Message: GENERIC_JDBC_CONNECTOR_0002:Unable to execute the SQL >>>>>>>>>>>>>>> statement" >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is the information of my job. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Database configuration >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Schema name: invoice >>>>>>>>>>>>>>> Table name: ds_msg_log >>>>>>>>>>>>>>> Table SQL statement: >>>>>>>>>>>>>>> Table column names: * >>>>>>>>>>>>>>> Partition column name: >>>>>>>>>>>>>>> Boundary query: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Output configuration >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Storage type: HDFS >>>>>>>>>>>>>>> Output format: TEXT_FILE >>>>>>>>>>>>>>> Output directory: /user/root/ds_msg_log >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Throttling resources Extractors: Loaders: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Since there is no information in the official guide talking >>>>>>>>>>>>>>> about how to set the values above, does any know anything wrong >>>>>>>>>>>>>>> in my job >>>>>>>>>>>>>>> setting? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is the log: Stack trace: at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcExecutor >>>>>>>>>>>>>>> (GenericJdbcExecutor.java:59) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:155) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:48) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:37) >>>>>>>>>>>>>>> at org.apache.sqoop.framework.FrameworkManager >>>>>>>>>>>>>>> (FrameworkManager.java:447) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:112) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:98) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:68) >>>>>>>>>>>>>>> at org.apache.sqoop.server.v1.SubmissionServlet >>>>>>>>>>>>>>> (SubmissionServlet.java:44) >>>>>>>>>>>>>>> at org.apache.sqoop.server.SqoopProtocolServlet >>>>>>>>>>>>>>> (SqoopProtocolServlet.java:63) >>>>>>>>>>>>>>> at javax.servlet.http.HttpServlet (HttpServlet.java:637) >>>>>>>>>>>>>>> at javax.servlet.http.HttpServlet (HttpServlet.java:717) >>>>>>>>>>>>>>> at org.apache.catalina.core.ApplicationFilterChain >>>>>>>>>>>>>>> (ApplicationFilterChain.java:290) >>>>>>>>>>>>>>> at org.apache.catalina.core.ApplicationFilterChain >>>>>>>>>>>>>>> (ApplicationFilterChain.java:206) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardWrapperValve >>>>>>>>>>>>>>> (StandardWrapperValve.java:233) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardContextValve >>>>>>>>>>>>>>> (StandardContextValve.java:191) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardHostValve >>>>>>>>>>>>>>> (StandardHostValve.java:127) >>>>>>>>>>>>>>> at org.apache.catalina.valves.ErrorReportValve >>>>>>>>>>>>>>> (ErrorReportValve.java:102) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardEngineValve >>>>>>>>>>>>>>> (StandardEngineValve.java:109) >>>>>>>>>>>>>>> at org.apache.catalina.connector.CoyoteAdapter >>>>>>>>>>>>>>> (CoyoteAdapter.java:293) >>>>>>>>>>>>>>> at org.apache.coyote.http11.Http11Processor >>>>>>>>>>>>>>> (Http11Processor.java:859) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler >>>>>>>>>>>>>>> (Http11Protocol.java:602) >>>>>>>>>>>>>>> at org.apache.tomcat.util.net.JIoEndpoint$Worker >>>>>>>>>>>>>>> (JIoEndpoint.java:489) >>>>>>>>>>>>>>> at java.lang.Thread (Thread.java:724) >>>>>>>>>>>>>>> Caused by: Exception: java.lang.Throwable Message: ERROR: >>>>>>>>>>>>>>> schema "invoice" does not exist Position: 46 Stack trace: at >>>>>>>>>>>>>>> org.postgresql.core.v3.QueryExecutorImpl >>>>>>>>>>>>>>> (QueryExecutorImpl.java:2102) >>>>>>>>>>>>>>> at org.postgresql.core.v3.QueryExecutorImpl >>>>>>>>>>>>>>> (QueryExecutorImpl.java:1835) >>>>>>>>>>>>>>> at org.postgresql.core.v3.QueryExecutorImpl >>>>>>>>>>>>>>> (QueryExecutorImpl.java:257) >>>>>>>>>>>>>>> at org.postgresql.jdbc2.AbstractJdbc2Statement >>>>>>>>>>>>>>> (AbstractJdbc2Statement.java:500) >>>>>>>>>>>>>>> at org.postgresql.jdbc2.AbstractJdbc2Statement >>>>>>>>>>>>>>> (AbstractJdbc2Statement.java:374) >>>>>>>>>>>>>>> at org.postgresql.jdbc2.AbstractJdbc2Statement >>>>>>>>>>>>>>> (AbstractJdbc2Statement.java:254) >>>>>>>>>>>>>>> at org.apache.sqoop.connector.jdbc.GenericJdbcExecutor >>>>>>>>>>>>>>> (GenericJdbcExecutor.java:56) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:155) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:48) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.sqoop.connector.jdbc.GenericJdbcImportInitializer >>>>>>>>>>>>>>> (GenericJdbcImportInitializer.java:37) >>>>>>>>>>>>>>> at org.apache.sqoop.framework.FrameworkManager >>>>>>>>>>>>>>> (FrameworkManager.java:447) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:112) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:98) >>>>>>>>>>>>>>> at org.apache.sqoop.handler.SubmissionRequestHandler >>>>>>>>>>>>>>> (SubmissionRequestHandler.java:68) >>>>>>>>>>>>>>> at org.apache.sqoop.server.v1.SubmissionServlet >>>>>>>>>>>>>>> (SubmissionServlet.java:44) >>>>>>>>>>>>>>> at org.apache.sqoop.server.SqoopProtocolServlet >>>>>>>>>>>>>>> (SqoopProtocolServlet.java:63) >>>>>>>>>>>>>>> at javax.servlet.http.HttpServlet (HttpServlet.java:637) >>>>>>>>>>>>>>> at javax.servlet.http.HttpServlet (HttpServlet.java:717) >>>>>>>>>>>>>>> at org.apache.catalina.core.ApplicationFilterChain >>>>>>>>>>>>>>> (ApplicationFilterChain.java:290) >>>>>>>>>>>>>>> at org.apache.catalina.core.ApplicationFilterChain >>>>>>>>>>>>>>> (ApplicationFilterChain.java:206) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardWrapperValve >>>>>>>>>>>>>>> (StandardWrapperValve.java:233) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardContextValve >>>>>>>>>>>>>>> (StandardContextValve.java:191) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardHostValve >>>>>>>>>>>>>>> (StandardHostValve.java:127) >>>>>>>>>>>>>>> at org.apache.catalina.valves.ErrorReportValve >>>>>>>>>>>>>>> (ErrorReportValve.java:102) >>>>>>>>>>>>>>> at org.apache.catalina.core.StandardEngineValve >>>>>>>>>>>>>>> (StandardEngineValve.java:109) >>>>>>>>>>>>>>> at org.apache.catalina.connector.CoyoteAdapter >>>>>>>>>>>>>>> (CoyoteAdapter.java:293) >>>>>>>>>>>>>>> at org.apache.coyote.http11.Http11Processor >>>>>>>>>>>>>>> (Http11Processor.java:859) >>>>>>>>>>>>>>> at >>>>>>>>>>>>>>> org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler >>>>>>>>>>>>>>> (Http11Protocol.java:602) >>>>>>>>>>>>>>> at org.apache.tomcat.util.net.JIoEndpoint$Worker >>>>>>>>>>>>>>> (JIoEndpoint.java:489) >>>>>>>>>>>>>>> at java.lang.Thread (Thread.java:724) >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> >
