I was just saying that you don'y need the %pyspark, just override the hive
connection string and use it. We did that for Oracle.
Eran

בתאריך שבת, 1 באוג׳ 2015, 00:33 מאת Dhaval Patel <dhaval1...@gmail.com>:

> Hi Eran,
>
> That is what I am trying to figure out - how can I put jar in the class
> path? I tried adding export var in setup script as mentioned below and it
> doesn't seem to add in the classpath. Or did I misunderstood anything?
>
> Thanks,
> Dhaval
>
> On Fri, Jul 31, 2015 at 5:14 PM, IT CTO <goi....@gmail.com> wrote:
>
>> As a simple hack you can put the jar in the class path and tgen set the
>> jdbc parameters in the hive interpreter parameters. Then use %hive and just
>> write sql aginst teradata.
>> Eran
>>
>> בתאריך שבת, 1 באוג׳ 2015, 00:08 מאת Dhaval Patel <dhaval1...@gmail.com>:
>>
>>> Hi,
>>>
>>> I am trying to connect to Teradata from spark and getting below error
>>> for not finding suitable drivers.
>>>
>>> : java.sql.SQLException: No suitable driver found for
>>> jdbc:teradata://XXXXXX
>>>
>>>
>>> I have tried adding jar files using %dep, as well as in zeppelin-env.sh
>>> setting up SPARK_CLASSPATH variable but instead of adding under
>>> classpaths, it adds under spark.driver.extraClassPath.
>>> SPARK_CLASSPATH=/...path/terajdbc4.jar:/..path/tdgssconfig.jar
>>>
>>>
>>> Below is code I tried from Z :
>>>
>>> %pyspark
>>> df = sqlContext.load(source="jdbc", url="jdbc:teradata://XXXXX,
>>> user=XXXXXX, password=XXXXXXX", dbtable="XXXXXXX")
>>>
>>> I have tried from shell adding the driver and connecting from there and
>>> it worked like charm.
>>>
>>> Thanks in advance!
>>>
>>> -Dhaval
>>>
>>
>

Reply via email to