Akash,

   Instead of adding the connector jar to $HADOOP_HOME/lib, when
running your map-reduce job using "hadoop jar" you can pass your
connector jar using the "-libjars" flag.

For example: hadoop jar hadoop-examples.jar wordcount -files
cachefile.txt -libjars mylib.jar input output

   ~ Minh

On Tue, Jun 26, 2012 at 9:52 PM, John Hancock <jhancock1...@gmail.com> wrote:
> Akash,
>
> I forgot to mention you will need to make sure the connector for your
> oracle db is available to Hadoop.  There are many ways to do this, but what
> works for me is I copy the library to <hadoop-install>/lib in all nodes in
> my cluster.
>
> I've written a similar program to do what you are asking about but using a
> mysql db, so I have mysql-connector-java-5.0.8-bin.jar in the lib directory
> on all my hadoop nodes.
>
> -John
>
> On Tue, Jun 26, 2012 at 8:38 AM, Akash Sharma <sharma...@hotmail.com> wrote:
>
>>
>> Thanks for the response John,
>>
>> This is exactly what I was looking for. I will pursue this direction and
>> create a Java program to connect with Oracle DB, extract relevant data and
>> push it into HDFS.
>>
>> I am guessing the Hadoop/HDFS API's will contain all the relevant calls
>> required to execute these tasks.
>>
>> Thanks,
>>
>> -Akash
>>
>> > Date: Tue, 26 Jun 2012 05:36:27 -0400
>> > Subject: Re: Sqoop Issue
>> > From: jhancock1...@gmail.com
>> > To: common-user@hadoop.apache.org
>> >
>> > Akash,
>> >
>> > You can write a simple Java program that queries your Oracle DB and uses
>> > whatever kind of file output object from java.io that you like to write
>> the
>> > data to a file.
>> >
>> > Compile the program and package it into a jar file.
>> >
>> > Then run the program using <path-to-hadoop>/'hadoop jar <jar-file-name>
>> > <output-file-name> in your Hadoop cluster.
>> >
>> > The result will be in HDFS.
>> >
>> > -John
>> >
>> > On Mon, Jun 25, 2012 at 8:02 AM, Akash Sharma <sharma...@hotmail.com>
>> wrote:
>> >
>> > >
>> > >
>> > > Hi,
>> > >
>> > > Seeking some advise/options on sqoop of data from Oracle DB to HDFS. We
>> > > are on Cloudera 3.
>> > >
>> > > Sqoop/JDBC connection to Oracle RAC fails within the Integration
>> cluster.
>> > > In the Oracle RAC there  is an additional layer between the Edge
>> > > Node/Server and Database. Routing the connection occurs to  appropriate
>> > > database listener. The specific issue is ORA-12516, TNS:listener could
>> not
>> > > find available handler with matching protocol stack. This issue is
>> > > primarily a network configuration issue which is stopping the team to
>> > > progress further.
>> > >
>> > > I would like to know other than sqoop, what are the other  options to
>> > > bring the data into HDFS. Our primary goal is to land data from Oracle
>> DB
>> > > into HDFS. Since sqoop does not work, we are thinking of using Oracle
>> > > unload, ftp, and then put command to import the data into HDFS. Please
>> > > guide if there  is any other option other than this. Will it help in
>> anyway
>> > > to code these options in java using HDFS/FTP Api within Java itself
>> rather
>> > > than using tools?
>> > >
>> > > Any help is appreciative.
>> > >
>> > > -Akash
>> > >
>> > >
>> > >
>>
>>

Reply via email to