Hi Eran,

That is what I am trying to figure out - how can I put jar in the class
path? I tried adding export var in setup script as mentioned below and it
doesn't seem to add in the classpath. Or did I misunderstood anything?

Thanks,
Dhaval

On Fri, Jul 31, 2015 at 5:14 PM, IT CTO <goi....@gmail.com> wrote:

> As a simple hack you can put the jar in the class path and tgen set the
> jdbc parameters in the hive interpreter parameters. Then use %hive and just
> write sql aginst teradata.
> Eran
>
> בתאריך שבת, 1 באוג׳ 2015, 00:08 מאת Dhaval Patel <dhaval1...@gmail.com>:
>
>> Hi,
>>
>> I am trying to connect to Teradata from spark and getting below error for
>> not finding suitable drivers.
>>
>> : java.sql.SQLException: No suitable driver found for
>> jdbc:teradata://XXXXXX
>>
>>
>> I have tried adding jar files using %dep, as well as in zeppelin-env.sh
>> setting up SPARK_CLASSPATH variable but instead of adding under
>> classpaths, it adds under spark.driver.extraClassPath.
>> SPARK_CLASSPATH=/...path/terajdbc4.jar:/..path/tdgssconfig.jar
>>
>>
>> Below is code I tried from Z :
>>
>> %pyspark
>> df = sqlContext.load(source="jdbc", url="jdbc:teradata://XXXXX,
>> user=XXXXXX, password=XXXXXXX", dbtable="XXXXXXX")
>>
>> I have tried from shell adding the driver and connecting from there and
>> it worked like charm.
>>
>> Thanks in advance!
>>
>> -Dhaval
>>
>

Reply via email to