Hi Sarah, this JIRA might be related to your problem with Oracle Date versus Timestamp field.
https://issues.apache.org/jira/browse/SQOOP-456 Another idea that I just got is to override Sqoop's default type mapping: http://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_controlling_type_mapping Jarcec On Fri, Oct 26, 2012 at 08:15:08PM +0530, Sarath wrote: > Sure will check that part. But I'm still struck making export work. > I'm facing issues exporting data to Oracle database. > > The issue is with the Timestamp fields. Sqoop generated code uses > java.sql.Timestamp which expects date field value in a particular > format. But in our data, available on hadoop, the format of the date > field is not guaranteed as the data is placed in this location by > multiple sources. > > I even tried putting hadoop data in the format expected by Timestamp > but still it complains with IllegalArgumentException. > > Is there any workaround? > > ~Sarath. > > On Friday 26 October 2012 08:06 PM, Jarek Jarcec Cecho wrote: > >Hi Sarath, > >I'm glad that Sqoop has started working for you. > > > >The internet advise with parameter -Dhadoopversion=100 should indeed do the > >trick, so I'm not sure what has went wrong with your build. Maybe some > >previous classes were found and not recompiled, would you mind trying "ant > >clean package -Dhadoopversion=100" to see if that helps? > > > >Jarcec > > > >On Fri, Oct 26, 2012 at 10:30:06AM +0530, Sarath wrote: > >>Thanks Jarcec. I downloaded the binary artifact and it's working now. > >> > >>I actually built my previous sqoop binary from the sources using the > >>option -Dhadoopversion=100 (since my hadoop version was 1.0.3) after > >>reading some blogs on the net. Not sure why it was still giving me > >>that exception. > >> > >>Sarath. > >> > >>On Thursday 25 October 2012 09:35 PM, Jarek Jarcec Cecho wrote: > >>>Hi Sarah, > >>>this exception is very typical when someone is messing together > >>>incompatible hadoop binaries and applications (for example sqoop compiled > >>>for hadoop 2 running on hadoop 1). Would you mind checking that you've > >>>downloaded appropriate binary distribution for your cluster? You have to > >>>use binary artifact sqoop-1.4.2.bin__hadoop-1.0.0.tar.gz for hadoop 1.0.3. > >>> > >>>Jarcec > >>> > >>>On Thu, Oct 25, 2012 at 07:23:49PM +0530, Sarath wrote: > >>>>Hi, > >>>> > >>>>I'm new to Sqoop. I have sqoop 1.4.2 with hadoop 1.0.3. I have both > >>>>hadoop and sqoop home environment variables set. > >>>> > >>>>I'm trying to export a file on HDFS to a table in Oracle database. I > >>>>included all the required parameters inside a file and then ran - > >>>>/sqoop --options-file export_params/ > >>>> > >>>>I got the below exception - > >>>>/Exception in thread "main" java.lang.IncompatibleClassChangeError: > >>>>Found class org.apache.hadoop.mapreduce.JobContext, but interface > >>>>was expected// > >>>>// at > >>>>org.apache.sqoop.mapreduce.ExportOutputFormat.checkOutputSpecs(ExportOutputFormat.java:57)// > >>>>// at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:887)// > >>>>// at org.apache.hadoop.mapred.JobClient$2.run(JobClient.java:850)// > >>>>// at java.security.AccessController.doPrivileged(Native Method)// > >>>>// ..../ > >>>> > >>>>Is there anything more to be configured? > >>>> > >>>>Regards, > >>>>Sarath.
signature.asc
Description: Digital signature
