Please do not use 4.12.0 release
FYI, the 4.12.0 release had a critical issue [1] that has been fixed in the 4.13.0 release. Please make sure you do not use the 4.12.0 release and instead use the 4.13.0 release. Sorry for any inconvenience. More details on the release may be found here [2]. Thanks, James [1] https://issues.apache.org/jira/browse/PHOENIX-4335 [2] https://blogs.apache.org/phoenix/entry/announcing-phoenix-4-13-released
[ANNOUNCE] Apache Phoenix 4.13 released
The Apache Phoenix team is pleased to announce the immediate availability of the 4.13.0 release. Apache Phoenix enables SQL-based OLTP and operational analytics for Apache Hadoop using Apache HBase as its backing store and providing integration with other projects in the Apache ecosystem such as Spark, Hive, Pig, Flume, and MapReduce. The 4.x releases are compatible with HBase 0.98 and 1.3. Highlights of the release include: * Critical bug fix to prevent snapshot creation of SYSTEM.CATALOG when connecting [1] * Numerous bug fixes around handling of row deletion [2] * Improvements to statistics collection [3] * New COLLATION_KEY built-in function for linguistic sort [4] Source and binary downloads are available here [5]. [1] https://issues.apache.org/jira/browse/PHOENIX-4335 [2] https://issues.apache.org/jira/issues/?jql=labels%20%3D%20rowDeletion [3] https://issues.apache.org/jira/issues/?jql=labels%20%3D%20statsCollection [4] https://phoenix.apache.org/language/functions.html#collation_key [5] http://phoenix.apache.org/download.html
Re: Spark & UpgradeInProgressException: Cluster is being concurrently upgraded from 4.11.x to 4.12.x
Hi Stepan, We discussed whether or not we should continue with Phoenix releases for HBase 1.1, but no one showed any interested in being the release manager [1], so we concluded that we would stop doing them. It's important to remember that the ASF is a volunteer effort and anyone can step up and take on this responsibility. That's essentially how contributors build merit to become committers and eventually PMC members and the project continues to grow. If you're interested, I suggest you start a new DISCUSS thread on the dev list and volunteer. Here's what would need to be done: - cherry-pick changes from master between 4.12.0 and 4.13.0 release to 4.x-HBase-1.1 branch - create a pull request with the above and get a +1 from a committer - monitor the Jenkins job that'll run with these changes keeping a lookout for any test failures - assuming there are no test failures, follow the directions here [2] to perform a release Thanks, James [1] https://lists.apache.org/thread.html/ae13def3c024603ce3cdde871223cbdbae0219b4efe93ed4e48f55d5@%3Cdev.phoenix.apache.org%3E [2] https://phoenix.apache.org/release.html On Sat, Nov 11, 2017 at 1:02 AM, stepan.migu...@firstlinesoftware.com < stepan.migu...@firstlinesoftware.com> wrote: > > > On 2017-11-10 22:36, Mujtaba Chohanwrote: > > Probably being hit by https://issues.apache.org/jira/browse/PHOENIX-4335 > . > > Please upgrade to 4.13.0 which will be available by EOD today. > > > > On Fri, Nov 10, 2017 at 8:37 AM, Stepan Migunov < > > stepan.migu...@firstlinesoftware.com> wrote: > > > > > Hi, > > > > > > > > > > > > I have just upgraded my cluster to Phoenix 4.12 and got an issue with > > > tasks running on Spark 2.2 (yarn cluster mode). Any attempts to use > method > > > phoenixTableAsDataFrame to load data from existing database causes an > > > exception (see below). > > > > > > > > > > > > The tasks worked fine on version 4.11. I have checked connection with > > > sqlline - it works now and shows that version is 4.12. Moreover, I have > > > noticed, that if limit the number of executors to one, the Spark's task > > > executing successfully too! > > > > > > > > > > > > It looks like that executors running in parallel "interferes" each > other’s > > > and could not acquire version's mutex. > > > > > > > > > > > > Any suggestions please? > > > > > > > > > > > > *final Connection connection = > > > ConnectionUtil.getInputConnection(configuration, overridingProps);* > > > > > > *User class threw exception: org.apache.spark.SparkException: Job > aborted > > > due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent > > > failure: Lost task 0.3 in stage 0.0 (TID 36, n7701-hdp005, executor > 26): > > > java.lang.RuntimeException: > > > org.apache.phoenix.exception.UpgradeInProgressException: Cluster is > being > > > concurrently upgraded from 4.11.x to 4.12.x. Please retry establishing > > > connection.* > > > > > > *at > > > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan( > PhoenixInputFormat.java:201)* > > > > > > *at > > > org.apache.phoenix.mapreduce.PhoenixInputFormat.createRecordReader( > PhoenixInputFormat.java:76)* > > > > > > *at > > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD. > scala:180)* > > > > > > *at > > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.( > NewHadoopRDD.scala:179)* > > > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)* > > > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)* > > > > > > *at org.apache.phoenix.spark.PhoenixRDD.compute(PhoenixRDD.scala:64)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > > > *at > > > org.apache.spark.rdd.MapPartitionsRDD.compute( > MapPartitionsRDD.scala:38)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > > > *at > > > org.apache.spark.rdd.MapPartitionsRDD.compute( > MapPartitionsRDD.scala:38)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > > > *at > > > org.apache.spark.rdd.MapPartitionsRDD.compute( > MapPartitionsRDD.scala:38)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > > > *at > > > org.apache.spark.rdd.MapPartitionsRDD.compute( > MapPartitionsRDD.scala:38)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > > > *at > > > org.apache.spark.rdd.MapPartitionsRDD.compute( > MapPartitionsRDD.scala:38)* > > > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > >
Re: Spark & UpgradeInProgressException: Cluster is being concurrently upgraded from 4.11.x to 4.12.x
On 2017-11-10 22:36, Mujtaba Chohanwrote: > Probably being hit by https://issues.apache.org/jira/browse/PHOENIX-4335. > Please upgrade to 4.13.0 which will be available by EOD today. > > On Fri, Nov 10, 2017 at 8:37 AM, Stepan Migunov < > stepan.migu...@firstlinesoftware.com> wrote: > > > Hi, > > > > > > > > I have just upgraded my cluster to Phoenix 4.12 and got an issue with > > tasks running on Spark 2.2 (yarn cluster mode). Any attempts to use method > > phoenixTableAsDataFrame to load data from existing database causes an > > exception (see below). > > > > > > > > The tasks worked fine on version 4.11. I have checked connection with > > sqlline - it works now and shows that version is 4.12. Moreover, I have > > noticed, that if limit the number of executors to one, the Spark's task > > executing successfully too! > > > > > > > > It looks like that executors running in parallel "interferes" each otherâs > > and could not acquire version's mutex. > > > > > > > > Any suggestions please? > > > > > > > > *final Connection connection = > > ConnectionUtil.getInputConnection(configuration, overridingProps);* > > > > *User class threw exception: org.apache.spark.SparkException: Job aborted > > due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent > > failure: Lost task 0.3 in stage 0.0 (TID 36, n7701-hdp005, executor 26): > > java.lang.RuntimeException: > > org.apache.phoenix.exception.UpgradeInProgressException: Cluster is being > > concurrently upgraded from 4.11.x to 4.12.x. Please retry establishing > > connection.* > > > > *at > > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan(PhoenixInputFormat.java:201)* > > > > *at > > org.apache.phoenix.mapreduce.PhoenixInputFormat.createRecordReader(PhoenixInputFormat.java:76)* > > > > *at > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)* > > > > *at > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.(NewHadoopRDD.scala:179)* > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)* > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)* > > > > *at org.apache.phoenix.spark.PhoenixRDD.compute(PhoenixRDD.scala:64)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)* > > > > *at > > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)* > > > > *at org.apache.spark.scheduler.Task.run(Task.scala:108)* > > > > *at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)* > > > > *at > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)* > > > > *at > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)* > > > > *at java.lang.Thread.run(Thread.java:745)* > > > > *Caused by: org.apache.phoenix.exception.UpgradeInProgressException: > > Cluster is being concurrently upgraded from 4.11.x to 4.12.x. Please retry > > establishing connection.* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.acquireUpgradeMutex(ConnectionQueryServicesImpl.java:3173)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.upgradeSystemTables(ConnectionQueryServicesImpl.java:2567)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2440)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2360)* > > > > *at > > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2360)* > > > > *at > >