Yup, that looks like an issue to me :-/ ------------------- Jesse Yates @jesse_yates jyates.github.com
On Tue, Aug 19, 2014 at 2:06 PM, Russell Jurney <[email protected]> wrote: > Running against any version would be ok, but it does not work. I get this > error: > > 2014-08-19 14:03:46,904 FATAL org.apache.hadoop.mapred.Child: Error running > child : java.lang.IncompatibleClassChangeError: Found interface > org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected > > ᐧ > I went back and checked out the Phoenix commit before the pom.xml was > upgraded to 0.98.4 > <https://github.com/apache/phoenix/commit/36a41c86a4cdb16ab91b95882dbdb02e6006831d>, > and I get the same error. > > Not sure what to do. Stuck. I imagine most users must be too, as CDH 5 is > probably the most popular installation there is? > > On Tue, Aug 19, 2014 at 11:50 AM, James Taylor <[email protected]> > wrote: > >> The dependencies on HBase 0.98.4 are *compile time* dependencies. Is it >> necessary for you to compile against CDH 5.1 or just run against it? >> >> >> On Tuesday, August 19, 2014, Russell Jurney <[email protected]> >> wrote: >> >>> Thats really bad. That means... CDH 5.x can't run Phoenix? How can this >>> be fixed? I'm not sure what to do. We're in limbo on our new cluster now. >>> >>> >>> On Mon, Aug 18, 2014 at 11:57 PM, Ravi Kiran <[email protected]> >>> wrote: >>> >>>> Hi Russel, >>>> Apparently, Phoenix 4.0.0 leverages few API methods of HBase 0.98.4 >>>> v which aren't present within 0.98.1 that comes with CDH 5.1 . That's the >>>> primary cause for the build issues. >>>> >>>> Regards >>>> Ravi >>>> >>>> >>>> >>>> On Mon, Aug 18, 2014 at 5:56 PM, Russell Jurney < >>>> [email protected]> wrote: >>>> >>>>> Talking to myself, but hopefully creating good docs. Replacing the >>>>> previous hadoop version with one I found here: >>>>> https://repository.cloudera.com/artifactory/cloudera-repos/org/apache/hadoop/hadoop-core/, >>>>> 2.3.0-mr1-cdh5.1.0, >>>>> makes things get a little further. >>>>> >>>>> I can't get past some build errors, however. Has anyone done this >>>>> before me who can help? >>>>> >>>>> [ERROR] >>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27] >>>>> cannot find symbol >>>>> >>>>> symbol: method valueOf(java.lang.String,int,int) >>>>> >>>>> location: class org.apache.hadoop.hbase.ServerName >>>>> >>>>> >>>>> [ERROR] Failed to execute goal >>>>> org.apache.maven.plugins:maven-compiler-plugin:3.0:compile >>>>> (default-compile) on project phoenix-core: Compilation failure >>>>> >>>>> [ERROR] >>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27] >>>>> cannot find symbol >>>>> >>>>> [ERROR] symbol: method valueOf(java.lang.String,int,int) >>>>> >>>>> [ERROR] location: class org.apache.hadoop.hbase.ServerName >>>>> >>>>> [ERROR] -> [Help 1] >>>>> >>>>> >>>>> >>>>> On Mon, Aug 18, 2014 at 5:41 PM, Russell Jurney < >>>>> [email protected]> wrote: >>>>> >>>>>> Ok, so it is clear to me what I have to do. I have to edit my pom.xml >>>>>> to point at CDH 5.1, which translates into: >>>>>> >>>>>> Add the cloudera repo: >>>>>> >>>>>> <repository> >>>>>> <id>cloudera</id> >>>>>> <url> >>>>>> https://repository.cloudera.com/artifactory/cloudera-repos/</url> >>>>>> </repository> >>>>>> >>>>>> >>>>>> Then change the hadoop and hbase versions: >>>>>> >>>>>> <!-- Hadoop Versions --> >>>>>> <hbase-hadoop1.version>0.98.1-cdh5.1.0</hbase-hadoop1.version> >>>>>> <hbase-hadoop2.version>0.98.1-cdh5.1.0</hbase-hadoop2.version> >>>>>> <hadoop-one.version>2.3.0-cdh5.1.0</hadoop-one.version> >>>>>> <hadoop-two.version>2.3.0-cdh5.1.0</hadoop-two.version> >>>>>> >>>>>> >>>>>> However, I get this error when I build, which tells me there is more >>>>>> complex POM surgery required. >>>>>> >>>>>> [ERROR] Failed to execute goal on project phoenix-core: Could not >>>>>> resolve dependencies for project >>>>>> org.apache.phoenix:phoenix-core:jar:4.0.0-incubating: The following >>>>>> artifacts could not be resolved: >>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0, >>>>>> org.apache.hadoop:hadoop-test:jar:2.3.0-cdh5.1.0: Could not find artifact >>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0 in apache release ( >>>>>> https://repository.apache.org/content/repositories/releases/) -> >>>>>> [Help 1] >>>>>> >>>>>> Beyond changing the versions, I do not know how to fix this. Can >>>>>> anyone help? >>>>>> >>>>>> In general, is it possible to actually handle different CDH versions >>>>>> in this project? One shouldn't have to do pom surgery to build Phoenix >>>>>> for >>>>>> the most common platform. >>>>>> >>>>>> >>>>>> On Mon, Aug 18, 2014 at 5:15 PM, Russell Jurney < >>>>>> [email protected]> wrote: >>>>>> >>>>>>> When I try to store data into Phoenix from Pig, I get this error. I >>>>>>> am on CDH 5.1, and Phoenix 4.0. >>>>>>> >>>>>>> Anyone know how to resolve this issue? >>>>>>> >>>>>>> 2014-08-18 17:11:25,165 INFO >>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader: >>>>>>> Current split being processed >>>>>>> hdfs://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+845 >>>>>>> 2014-08-18 >>>>>>> <http://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+8452014-08-18> >>>>>>> 17:11:25,173 INFO org.apache.hadoop.mapred.TaskLogsTruncater: >>>>>>> Initializing logs' truncater with mapRetainSize=-1 and >>>>>>> reduceRetainSize=-1 >>>>>>> 2014-08-18 17:11:25,175 FATAL org.apache.hadoop.mapred.Child: Error >>>>>>> running child : java.lang.IncompatibleClassChangeError: Found interface >>>>>>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected >>>>>>> at >>>>>>> org.apache.phoenix.pig.hadoop.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:65) >>>>>>> at >>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:84) >>>>>>> at >>>>>>> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:548) >>>>>>> at >>>>>>> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:653) >>>>>>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:330) >>>>>>> at org.apache.hadoop.mapred.Child$4.run(Child.java:268) >>>>>>> at java.security.AccessController.doPrivileged(Native Method) >>>>>>> at javax.security.auth.Subject.doAs(Subject.java:415) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554) >>>>>>> at org.apache.hadoop.mapred.Child.main(Child.java:262) >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Russell Jurney twitter.com/rjurney [email protected] >>>>>>> datasyndrome.com >>>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Russell Jurney twitter.com/rjurney [email protected] >>>>>> datasyndrome.com >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Russell Jurney twitter.com/rjurney [email protected] >>>>> datasyndrome.com >>>>> >>>> >>>> >>> >>> >>> -- >>> Russell Jurney twitter.com/rjurney [email protected] datasyndrome >>> .com >>> >> > > > -- > Russell Jurney twitter.com/rjurney [email protected] datasyndrome. > com >
