And I got the below version from after using sqoop version. ubuntu@ip-172-31-1-201:~$ sqoop version Warning: /opt/cloudera/parcels/CDH-5.4.2-1.cdh5.4.2.p0.2/bin/../lib/sqoop/../accumulo does not exist! Accumulo imports will fail. Please set $ACCUMULO_HOME to the root of your Accumulo installation. 15/05/31 02:19:57 INFO sqoop.Sqoop: Running Sqoop version: 1.4.5-cdh5.4.2 Sqoop 1.4.5-cdh5.4.2 git commit id Compiled by on Tue May 19 16:54:41 PDT 2015
Kind Regards, Sankara Telukutla +1 510 936 0999 On Sat, May 30, 2015 at 12:43 PM, SHANKAR REDDY <[email protected] > wrote: > Hi Abe, > I am using 5.4.2-1.cdh5.4.2.p0.2 > > Is there any way to find the version no directly from this distribution. > > Thanks, > Shankar > > > On Friday, May 29, 2015, Abraham Elmahrek <[email protected]> wrote: > >> What version of Sqoop2 are you using? >> >> On Fri, May 29, 2015 at 1:52 AM, SHANKAR REDDY < >> [email protected]> >> wrote: >> >> > Team, >> > I am facing the below error while running Sqoop 2 job where the data is >> > importing from MySQL. Please let me know if you need any more >> information. >> > >> > 2015-05-29 08:46:44,773 INFO >> > org.apache.sqoop.repository.JdbcRepositoryTransaction: Attempting >> > transaction commit >> > 2015-05-29 08:46:45,025 ERROR >> > org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine: Error >> in > > > >> > submitting job >> > java.lang.ArithmeticException: / by zero >> > at >> > >> > >> org.apache.sqoop.connector.jdbc.GenericJdbcPartitioner.partitionIntegerColumn(GenericJdbcPartitioner.java:317) >> > at >> > >> > >> org.apache.sqoop.connector.jdbc.GenericJdbcPartitioner.getPartitions(GenericJdbcPartitioner.java:86) >> > at >> > >> > >> org.apache.sqoop.connector.jdbc.GenericJdbcPartitioner.getPartitions(GenericJdbcPartitioner.java:38) >> > at >> > >> > >> org.apache.sqoop.job.mr.SqoopInputFormat.getSplits(SqoopInputFormat.java:74) >> > at >> > >> > >> org.apache.hadoop.mapreduce.JobSubmitter.writeNewSplits(JobSubmitter.java:597) >> > at >> > >> org.apache.hadoop.mapreduce.JobSubmitter.writeSplits(JobSubmitter.java:614) >> > at >> > >> > >> org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:492) >> > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1306) >> > at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1303) >> > at java.security.AccessController.doPrivileged(Native Method) >> > at javax.security.auth.Subject.doAs(Subject.java:415) >> > at >> > >> > >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1671) >> > at org.apache.hadoop.mapreduce.Job.submit(Job.java:1303) >> > at >> > >> > >> org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine.submitToCluster(MapreduceSubmissionEngine.java:274) >> > at >> > >> > >> org.apache.sqoop.submission.mapreduce.MapreduceSubmissionEngine.submit(MapreduceSubmissionEngine.java:255) >> > at org.apache.sqoop.driver.JobManager.start(JobManager.java:288) >> > at >> > >> > >> org.apache.sqoop.handler.JobRequestHandler.startJob(JobRequestHandler.java:379) >> > at >> > >> > >> org.apache.sqoop.handler.JobRequestHandler.handleEvent(JobRequestHandler.java:115) >> > at >> > >> org.apache.sqoop.server.v1.JobServlet.handlePutRequest(JobServlet.java:96) >> > at >> > >> > >> org.apache.sqoop.server.SqoopProtocolServlet.doPut(SqoopProtocolServlet.java:79) >> > at javax.servlet.http.HttpServlet.service(HttpServlet.java:646) >> > at javax.servlet.http.HttpServlet.service(HttpServlet.java:723) >> > at >> > >> > >> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:290) >> > at >> > >> > >> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) >> > at >> > >> > >> org.apache.hadoop.security.authentication.server.AuthenticationFilter.doFilter(AuthenticationFilter.java:592) >> > at >> > >> > >> org.apache.hadoop.security.token.delegation.web.DelegationTokenAuthenticationFilter.doFilter(DelegationTokenAuthenticationFilter.java:277) >> > at >> > >> > >> org.apache.hadoop.security.authentication.server.AuthenticationFilter.doFilter(AuthenticationFilter.java:555) >> > at >> > >> > >> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) >> > at >> > >> > >> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) >> > at >> > >> > >> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) >> > at >> > >> > >> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) >> > at >> > >> > >> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) >> > at >> > >> > >> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103) >> > at >> > >> > >> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) >> > at >> > >> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293) >> > at >> > >> org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:861) >> > at >> > >> > >> org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:620) >> > at >> > org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) >> > at java.lang.Thread.run(Thread.java:745) >> > 2015-05-29 08:46:45,025 INFO >> > org.apache.sqoop.connector.jdbc.GenericJdbcFromDestroyer: Running >> generic >> > JDBC connector destroyer >> > 2015-05-29 08:46:45,027 INFO >> > org.apache.sqoop.repository.JdbcRepositoryTransaction: Attempting >> > transaction commit >> > >> > >> > - Shankar >> > >> > > > -- > Regards, > Sankara Reddy Telukutla >
