Please do not use 4.12.0 release

2017-11-11 Thread James Taylor
FYI, the 4.12.0 release had a critical issue [1] that has been fixed in the
4.13.0 release. Please make sure you do not use the 4.12.0 release and
instead use the 4.13.0 release. Sorry for any inconvenience. More details
on the release may be found here [2].

Thanks,
James

[1] https://issues.apache.org/jira/browse/PHOENIX-4335
[2] https://blogs.apache.org/phoenix/entry/announcing-phoenix-4-13-released


[ANNOUNCE] Apache Phoenix 4.13 released

2017-11-11 Thread James Taylor
The Apache Phoenix team is pleased to announce the immediate availability
of the 4.13.0 release. Apache Phoenix enables SQL-based OLTP and
operational analytics for Apache Hadoop using Apache HBase as its backing
store and providing integration with other projects in the Apache ecosystem
such as Spark, Hive, Pig, Flume, and MapReduce. The 4.x releases are
compatible with HBase 0.98 and 1.3.

Highlights of the release include:

* Critical bug fix to prevent snapshot creation of SYSTEM.CATALOG when
connecting [1]
* Numerous bug fixes around handling of row deletion [2]
* Improvements to statistics collection [3]
* New COLLATION_KEY built-in function for linguistic sort [4]

Source and binary downloads are available here [5].

[1] https://issues.apache.org/jira/browse/PHOENIX-4335
[2] https://issues.apache.org/jira/issues/?jql=labels%20%3D%20rowDeletion
[3]
https://issues.apache.org/jira/issues/?jql=labels%20%3D%20statsCollection
[4] https://phoenix.apache.org/language/functions.html#collation_key
[5] http://phoenix.apache.org/download.html


Re: Spark & UpgradeInProgressException: Cluster is being concurrently upgraded from 4.11.x to 4.12.x

2017-11-11 Thread James Taylor
Hi Stepan,
We discussed whether or not we should continue with Phoenix releases for
HBase 1.1, but no one showed any interested in being the release manager
[1], so we concluded that we would stop doing them. It's important to
remember that the ASF is a volunteer effort and anyone can step up and take
on this responsibility. That's essentially how contributors build merit to
become committers and eventually PMC members and the project continues to
grow. If you're interested, I suggest you start a new DISCUSS thread on the
dev list and volunteer. Here's what would need to be done:
- cherry-pick changes from master between 4.12.0 and 4.13.0 release to
4.x-HBase-1.1 branch
- create a pull request with the above and get a +1 from a committer
- monitor the Jenkins job that'll run with these changes keeping a lookout
for any test failures
- assuming there are no test failures, follow the directions here [2] to
perform a release

Thanks,
James


[1]
https://lists.apache.org/thread.html/ae13def3c024603ce3cdde871223cbdbae0219b4efe93ed4e48f55d5@%3Cdev.phoenix.apache.org%3E
[2] https://phoenix.apache.org/release.html

On Sat, Nov 11, 2017 at 1:02 AM, stepan.migu...@firstlinesoftware.com <
stepan.migu...@firstlinesoftware.com> wrote:

>
>
> On 2017-11-10 22:36, Mujtaba Chohan  wrote:
> > Probably being hit by https://issues.apache.org/jira/browse/PHOENIX-4335
> .
> > Please upgrade to 4.13.0 which will be available by EOD today.
> >
> > On Fri, Nov 10, 2017 at 8:37 AM, Stepan Migunov <
> > stepan.migu...@firstlinesoftware.com> wrote:
> >
> > > Hi,
> > >
> > >
> > >
> > > I have just upgraded my cluster to Phoenix 4.12 and got an issue with
> > > tasks running on Spark 2.2 (yarn cluster mode). Any attempts to use
> method
> > > phoenixTableAsDataFrame to load data from existing database causes an
> > > exception (see below).
> > >
> > >
> > >
> > > The tasks worked fine on version 4.11. I have checked connection with
> > > sqlline - it works now and shows that version is 4.12. Moreover, I have
> > > noticed, that if limit the number of executors to one, the Spark's task
> > > executing successfully too!
> > >
> > >
> > >
> > > It looks like that executors running in parallel "interferes" each
> other’s
> > > and could not acquire version's mutex.
> > >
> > >
> > >
> > > Any suggestions please?
> > >
> > >
> > >
> > > *final Connection connection =
> > > ConnectionUtil.getInputConnection(configuration, overridingProps);*
> > >
> > > *User class threw exception: org.apache.spark.SparkException: Job
> aborted
> > > due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent
> > > failure: Lost task 0.3 in stage 0.0 (TID 36, n7701-hdp005, executor
> 26):
> > > java.lang.RuntimeException:
> > > org.apache.phoenix.exception.UpgradeInProgressException: Cluster is
> being
> > > concurrently upgraded from 4.11.x to 4.12.x. Please retry establishing
> > > connection.*
> > >
> > > *at
> > > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan(
> PhoenixInputFormat.java:201)*
> > >
> > > *at
> > > org.apache.phoenix.mapreduce.PhoenixInputFormat.createRecordReader(
> PhoenixInputFormat.java:76)*
> > >
> > > *at
> > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.
> scala:180)*
> > >
> > > *at
> > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.(
> NewHadoopRDD.scala:179)*
> > >
> > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)*
> > >
> > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)*
> > >
> > > *at org.apache.phoenix.spark.PhoenixRDD.compute(PhoenixRDD.scala:64)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > > *at
> > > org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > > *at
> > > org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > > *at
> > > org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > > *at
> > > org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > > *at
> > > org.apache.spark.rdd.MapPartitionsRDD.compute(
> MapPartitionsRDD.scala:38)*
> > >
> > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> > >
> > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> > >
> > 

Re: Spark & UpgradeInProgressException: Cluster is being concurrently upgraded from 4.11.x to 4.12.x

2017-11-11 Thread stepan.migu...@firstlinesoftware.com


On 2017-11-10 22:36, Mujtaba Chohan  wrote: 
> Probably being hit by https://issues.apache.org/jira/browse/PHOENIX-4335.
> Please upgrade to 4.13.0 which will be available by EOD today.
> 
> On Fri, Nov 10, 2017 at 8:37 AM, Stepan Migunov <
> stepan.migu...@firstlinesoftware.com> wrote:
> 
> > Hi,
> >
> >
> >
> > I have just upgraded my cluster to Phoenix 4.12 and got an issue with
> > tasks running on Spark 2.2 (yarn cluster mode). Any attempts to use method
> > phoenixTableAsDataFrame to load data from existing database causes an
> > exception (see below).
> >
> >
> >
> > The tasks worked fine on version 4.11. I have checked connection with
> > sqlline - it works now and shows that version is 4.12. Moreover, I have
> > noticed, that if limit the number of executors to one, the Spark's task
> > executing successfully too!
> >
> >
> >
> > It looks like that executors running in parallel "interferes" each other’s
> > and could not acquire version's mutex.
> >
> >
> >
> > Any suggestions please?
> >
> >
> >
> > *final Connection connection =
> > ConnectionUtil.getInputConnection(configuration, overridingProps);*
> >
> > *User class threw exception: org.apache.spark.SparkException: Job aborted
> > due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent
> > failure: Lost task 0.3 in stage 0.0 (TID 36, n7701-hdp005, executor 26):
> > java.lang.RuntimeException:
> > org.apache.phoenix.exception.UpgradeInProgressException: Cluster is being
> > concurrently upgraded from 4.11.x to 4.12.x. Please retry establishing
> > connection.*
> >
> > *at
> > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan(PhoenixInputFormat.java:201)*
> >
> > *at
> > org.apache.phoenix.mapreduce.PhoenixInputFormat.createRecordReader(PhoenixInputFormat.java:76)*
> >
> > *at
> > org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)*
> >
> > *at
> > org.apache.spark.rdd.NewHadoopRDD$$anon$1.(NewHadoopRDD.scala:179)*
> >
> > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)*
> >
> > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)*
> >
> > *at org.apache.phoenix.spark.PhoenixRDD.compute(PhoenixRDD.scala:64)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)*
> >
> > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)*
> >
> > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)*
> >
> > *at
> > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)*
> >
> > *at
> > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)*
> >
> > *at org.apache.spark.scheduler.Task.run(Task.scala:108)*
> >
> > *at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)*
> >
> > *at
> > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)*
> >
> > *at
> > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)*
> >
> > *at java.lang.Thread.run(Thread.java:745)*
> >
> > *Caused by: org.apache.phoenix.exception.UpgradeInProgressException:
> > Cluster is being concurrently upgraded from 4.11.x to 4.12.x. Please retry
> > establishing connection.*
> >
> > *at
> > org.apache.phoenix.query.ConnectionQueryServicesImpl.acquireUpgradeMutex(ConnectionQueryServicesImpl.java:3173)*
> >
> > *at
> > org.apache.phoenix.query.ConnectionQueryServicesImpl.upgradeSystemTables(ConnectionQueryServicesImpl.java:2567)*
> >
> > *at
> > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2440)*
> >
> > *at
> > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2360)*
> >
> > *at
> > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)*
> >
> > *at
> > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2360)*
> >
> > *at
> >