First of all, I apologize if you feel like I was picking on you. I was not
trying to do that.

My understanding is that Salesforce pays people to work on Phoenix. Is that
not the case? I'm hoping one of them will add spark-like support for CDH
and HDP to advance the project.

And I don't mention the POM thing to pick on someone. The majority of HBase
users are not going to be able to use Phoenix because they run a commercial
distribution of Hadoop and aren't pom wizards. That seems kind of important
for the well being of the project.
ᐧ


On Tue, Aug 19, 2014 at 3:26 PM, Andrew Purtell <[email protected]> wrote:

> I don't think an Apache project should spend precious bandwidth tracking
> the various and sundry redistributors of Apache ecosystem projects. This is
> putting the cart before the horse. The horse is the Apache upstream
> projects. The cart is the commercial distributions leveraging the Apache
> ecosystem for profit. Spark is not a good example, it is supported by a
> commercial concern, Databricks. What commercial company supports Phoenix?
> Why do you think it is appropriate to pick on volunteers because editing
> POM files is too much work?
>
>
> On Tue, Aug 19, 2014 at 3:09 PM, Russell Jurney <[email protected]>
> wrote:
>
>> I also created https://issues.apache.org/jira/browse/PHOENIX-1185
>> because requiring users to hand-edit the pom.xml just to build against CDH
>> and HDP is nuts.
>> ᐧ
>>
>>
>> On Tue, Aug 19, 2014 at 3:03 PM, Russell Jurney <[email protected]
>> > wrote:
>>
>>> I have created https://issues.apache.org/jira/browse/PHOENIX-1183 and
>>> https://issues.apache.org/jira/browse/PHOENIX-1184
>>>
>>> I hope this can be resolved soon, we are pretty screwed right now.
>>> ᐧ
>>>
>>>
>>> On Tue, Aug 19, 2014 at 2:08 PM, Jesse Yates <[email protected]>
>>> wrote:
>>>
>>>> Yup, that looks like an issue to me :-/
>>>>
>>>> -------------------
>>>> Jesse Yates
>>>> @jesse_yates
>>>> jyates.github.com
>>>>
>>>>
>>>> On Tue, Aug 19, 2014 at 2:06 PM, Russell Jurney <
>>>> [email protected]> wrote:
>>>>
>>>>> Running against any version would be ok, but it does not work. I get
>>>>> this error:
>>>>>
>>>>> 2014-08-19 14:03:46,904 FATAL org.apache.hadoop.mapred.Child: Error 
>>>>> running child : java.lang.IncompatibleClassChangeError: Found interface 
>>>>> org.apache.hadoop.mapreduce.TaskAttemptContext, but class was expected
>>>>>
>>>>> ᐧ
>>>>> I went back and checked out the Phoenix commit before the pom.xml was
>>>>> upgraded to 0.98.4
>>>>> <https://github.com/apache/phoenix/commit/36a41c86a4cdb16ab91b95882dbdb02e6006831d>,
>>>>> and I get the same error.
>>>>>
>>>>> Not sure what to do. Stuck. I imagine most users must be too, as CDH 5
>>>>> is probably the most popular installation there is?
>>>>>
>>>>> On Tue, Aug 19, 2014 at 11:50 AM, James Taylor <[email protected]
>>>>> > wrote:
>>>>>
>>>>>>  The dependencies on HBase 0.98.4 are *compile time* dependencies. Is
>>>>>> it necessary for you to compile against CDH 5.1 or just run against it?
>>>>>>
>>>>>>
>>>>>> On Tuesday, August 19, 2014, Russell Jurney <[email protected]>
>>>>>> wrote:
>>>>>>
>>>>>>> Thats really bad. That means... CDH 5.x can't run Phoenix? How can
>>>>>>> this be fixed? I'm not sure what to do. We're in limbo on our new 
>>>>>>> cluster
>>>>>>> now.
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Aug 18, 2014 at 11:57 PM, Ravi Kiran <
>>>>>>> [email protected]> wrote:
>>>>>>>
>>>>>>>> Hi Russel,
>>>>>>>>     Apparently, Phoenix 4.0.0 leverages few API methods of HBase
>>>>>>>> 0.98.4 v  which aren't present within 0.98.1 that comes with CDH 5.1 .
>>>>>>>> That's the primary cause for the build issues.
>>>>>>>>
>>>>>>>> Regards
>>>>>>>> Ravi
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Aug 18, 2014 at 5:56 PM, Russell Jurney <
>>>>>>>> [email protected]> wrote:
>>>>>>>>
>>>>>>>>> Talking to myself, but hopefully creating good docs. Replacing the
>>>>>>>>> previous hadoop version with one I found here:
>>>>>>>>> https://repository.cloudera.com/artifactory/cloudera-repos/org/apache/hadoop/hadoop-core/,
>>>>>>>>>  2.3.0-mr1-cdh5.1.0,
>>>>>>>>> makes things get a little further.
>>>>>>>>>
>>>>>>>>> I can't get past some build errors, however. Has anyone done this
>>>>>>>>> before me who can help?
>>>>>>>>>
>>>>>>>>> [ERROR]
>>>>>>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27]
>>>>>>>>> cannot find symbol
>>>>>>>>>
>>>>>>>>>   symbol:   method valueOf(java.lang.String,int,int)
>>>>>>>>>
>>>>>>>>>   location: class org.apache.hadoop.hbase.ServerName
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>> org.apache.maven.plugins:maven-compiler-plugin:3.0:compile
>>>>>>>>> (default-compile) on project phoenix-core: Compilation failure
>>>>>>>>>
>>>>>>>>> [ERROR]
>>>>>>>>> /Users/rjurney/Software/phoenix4/phoenix-4.0.0-incubating-src/phoenix-core/src/main/java/org/apache/phoenix/query/ConnectionlessQueryServicesImpl.java:[143,27]
>>>>>>>>> cannot find symbol
>>>>>>>>>
>>>>>>>>> [ERROR] symbol:   method valueOf(java.lang.String,int,int)
>>>>>>>>>
>>>>>>>>> [ERROR] location: class org.apache.hadoop.hbase.ServerName
>>>>>>>>>
>>>>>>>>> [ERROR] -> [Help 1]
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Mon, Aug 18, 2014 at 5:41 PM, Russell Jurney <
>>>>>>>>> [email protected]> wrote:
>>>>>>>>>
>>>>>>>>>> Ok, so it is clear to me what I have to do. I have to edit my
>>>>>>>>>> pom.xml to point at CDH 5.1, which translates into:
>>>>>>>>>>
>>>>>>>>>> Add the cloudera repo:
>>>>>>>>>>
>>>>>>>>>>     <repository>
>>>>>>>>>>       <id>cloudera</id>
>>>>>>>>>>       <url>
>>>>>>>>>> https://repository.cloudera.com/artifactory/cloudera-repos/</url>
>>>>>>>>>>     </repository>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Then change the hadoop and hbase versions:
>>>>>>>>>>
>>>>>>>>>>      <!-- Hadoop Versions -->
>>>>>>>>>>     <hbase-hadoop1.version>0.98.1-cdh5.1.0</hbase-hadoop1.version>
>>>>>>>>>>     <hbase-hadoop2.version>0.98.1-cdh5.1.0</hbase-hadoop2.version>
>>>>>>>>>>     <hadoop-one.version>2.3.0-cdh5.1.0</hadoop-one.version>
>>>>>>>>>>     <hadoop-two.version>2.3.0-cdh5.1.0</hadoop-two.version>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> However, I get this error when I build, which tells me there is
>>>>>>>>>> more complex POM surgery required.
>>>>>>>>>>
>>>>>>>>>> [ERROR] Failed to execute goal on project phoenix-core: Could not
>>>>>>>>>> resolve dependencies for project
>>>>>>>>>> org.apache.phoenix:phoenix-core:jar:4.0.0-incubating: The following
>>>>>>>>>> artifacts could not be resolved:
>>>>>>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0,
>>>>>>>>>> org.apache.hadoop:hadoop-test:jar:2.3.0-cdh5.1.0: Could not find 
>>>>>>>>>> artifact
>>>>>>>>>> org.apache.hadoop:hadoop-core:jar:2.3.0-cdh5.1.0 in apache release (
>>>>>>>>>> https://repository.apache.org/content/repositories/releases/) ->
>>>>>>>>>> [Help 1]
>>>>>>>>>>
>>>>>>>>>> Beyond changing the versions, I do not know how to fix this. Can
>>>>>>>>>> anyone help?
>>>>>>>>>>
>>>>>>>>>> In general, is it possible to actually handle different CDH
>>>>>>>>>> versions in this project? One shouldn't have to do pom surgery to 
>>>>>>>>>> build
>>>>>>>>>> Phoenix for the most common platform.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Mon, Aug 18, 2014 at 5:15 PM, Russell Jurney <
>>>>>>>>>> [email protected]> wrote:
>>>>>>>>>>
>>>>>>>>>>> When I try to store data into Phoenix from Pig, I get this
>>>>>>>>>>> error. I am on CDH 5.1, and Phoenix 4.0.
>>>>>>>>>>>
>>>>>>>>>>> Anyone know how to resolve this issue?
>>>>>>>>>>>
>>>>>>>>>>> 2014-08-18 17:11:25,165 INFO 
>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigRecordReader:
>>>>>>>>>>>  Current split being processed 
>>>>>>>>>>> hdfs://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+845
>>>>>>>>>>> 2014-08-18 
>>>>>>>>>>> <http://cluster1-srv1.e8.com:8020/e8/prod/web_behavior/anomaly_profile.txt/2014/07/15/00/part-r-00000:0+8452014-08-18>
>>>>>>>>>>>  17:11:25,173 INFO org.apache.hadoop.mapred.TaskLogsTruncater: 
>>>>>>>>>>> Initializing logs' truncater with mapRetainSize=-1 and 
>>>>>>>>>>> reduceRetainSize=-1
>>>>>>>>>>> 2014-08-18 17:11:25,175 FATAL org.apache.hadoop.mapred.Child: Error 
>>>>>>>>>>> running child : java.lang.IncompatibleClassChangeError: Found 
>>>>>>>>>>> interface org.apache.hadoop.mapreduce.TaskAttemptContext, but class 
>>>>>>>>>>> was expected
>>>>>>>>>>>     at 
>>>>>>>>>>> org.apache.phoenix.pig.hadoop.PhoenixOutputFormat.getRecordWriter(PhoenixOutputFormat.java:65)
>>>>>>>>>>>     at 
>>>>>>>>>>> org.apache.pig.backend.hadoop.executionengine.mapReduceLayer.PigOutputFormat.getRecordWriter(PigOutputFormat.java:84)
>>>>>>>>>>>     at 
>>>>>>>>>>> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:548)
>>>>>>>>>>>     at 
>>>>>>>>>>> org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:653)
>>>>>>>>>>>     at org.apache.hadoop.mapred.MapTask.run(MapTask.java:330)
>>>>>>>>>>>     at org.apache.hadoop.mapred.Child$4.run(Child.java:268)
>>>>>>>>>>>     at java.security.AccessController.doPrivileged(Native Method)
>>>>>>>>>>>     at javax.security.auth.Subject.doAs(Subject.java:415)
>>>>>>>>>>>     at 
>>>>>>>>>>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1554)
>>>>>>>>>>>     at org.apache.hadoop.mapred.Child.main(Child.java:262)
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Russell Jurney twitter.com/rjurney [email protected]
>>>>>>>>>>> datasyndrome.com
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Russell Jurney twitter.com/rjurney [email protected]
>>>>>>>>>> datasyndrome.com
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Russell Jurney twitter.com/rjurney [email protected]
>>>>>>>>> datasyndrome.com
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Russell Jurney twitter.com/rjurney [email protected]
>>>>>>> datasyndrome.com
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Russell Jurney twitter.com/rjurney [email protected]
>>>>> datasyndrome.com
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Russell Jurney twitter.com/rjurney [email protected] datasyndrome
>>> .com
>>>
>>
>>
>>
>> --
>> Russell Jurney twitter.com/rjurney [email protected] datasyndrome.
>> com
>>
>
>
>
> --
> Best regards,
>
>    - Andy
>
> Problems worthy of attack prove their worth by hitting back. - Piet Hein
> (via Tom White)
>



-- 
Russell Jurney twitter.com/rjurney [email protected] datasyndrome.com

Reply via email to