Ahh. Sorry ignore the typo in my stacktrace. That is due to me trying to
remove hostnames from stacktrace.

The url is jdbc:phoenix:remoteclusterZkQuorum:2181

--Siddhi

On Tue, Aug 4, 2015 at 11:35 AM, Samarth Jain <sama...@apache.org> wrote:

> The jdbc url doesn't look correct - jdbc:phoenix:jdbc:phoenix:
> remoteclusterZkQuorum:2181
>
> It should be jdbc:phoenix:remoteclusterZkQuorum:2181
>
> Do you have the phoneix.mapreduce.output.cluster.quorum configured (take
> note of the typo)? Or hbase.zookeeper.quorum? If yes, what are the values
> set as?
>
>
>
>
>
> On Tue, Aug 4, 2015 at 11:19 AM, Siddhi Mehta <sm26...@gmail.com> wrote:
>
> > Hello,
> >
> > I am trying to run a pig job with mapreduce mode that tries to write to
> > Hbase using PhoenixHBaseStorage.
> >
> > I am seeing the reduce task fail with no suitable driver found for the
> > connection.
> >
> > AttemptID:attempt_1436998373852_1140_r_000000_1 Info:Error:
> > java.lang.RuntimeException: java.sql.SQLException: No suitable driver
> > found for jdbc:phoenix:remoteclusterZkQuorum:2181;
> >         at
> >
> org.apache.phoenix.mapreduce.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:58)
> >         at
> >
> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:88)
> >         at
> >
> org.apache.hadoop.mapred.ReduceTask$NewTrackingRecordWriter.<init>(ReduceTask.java:540)
> >         at
> > org.apache.hadoop.mapred.ReduceTask.runNewReducer(ReduceTask.java:614)
> >         at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:389)
> >         at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
> >         at java.security.AccessController.doPrivileged(Native Method)
> >         at javax.security.auth.Subject.doAs(Subject.java:422)
> >         at
> >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1576)
> >         at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
> > Caused by: java.sql.SQLException: No suitable driver found for
> > jdbc:phoenix:jdbc:phoenix:remoteclusterZkQuorum:2181;
> >
> >         at java.sql.DriverManager.getConnection(DriverManager.java:689)
> >         at java.sql.DriverManager.getConnection(DriverManager.java:208)
> >         at
> >
> org.apache.phoenix.mapreduce.util.ConnectionUtil.getConnection(ConnectionUtil.java:93)
> >         at
> >
> org.apache.phoenix.mapreduce.util.ConnectionUtil.getOutputConnection(ConnectionUtil.java:80)
> >         at
> >
> org.apache.phoenix.mapreduce.util.ConnectionUtil.getOutputConnection(ConnectionUtil.java:68)
> >         at
> >
> org.apache.phoenix.mapreduce.PhoenixRecordWriter.<init>(PhoenixRecordWriter.java:49)
> >         at
> >
> org.apache.phoenix.mapreduce.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:55)
> >
> >
> > I checked that the PhoenixDriver is on the classpath for the reduce task
> by
> > adding a Class.forName("org.apache.phoenix.jdbc.PhoenixDriver") but the
> > write still fails.
> >
> > Anyone else encountered an issue while trying Hbase writes via Pig in
> > mapreduce mode.
> >
> > --Siddhi
> >
>

Reply via email to