On 2017-11-10 22:36, Mujtaba Chohan <mujt...@apache.org> wrote: > Probably being hit by https://issues.apache.org/jira/browse/PHOENIX-4335. > Please upgrade to 4.13.0 which will be available by EOD today. > > On Fri, Nov 10, 2017 at 8:37 AM, Stepan Migunov < > stepan.migu...@firstlinesoftware.com> wrote: > > > Hi, > > > > > > > > I have just upgraded my cluster to Phoenix 4.12 and got an issue with > > tasks running on Spark 2.2 (yarn cluster mode). Any attempts to use method > > phoenixTableAsDataFrame to load data from existing database causes an > > exception (see below). > > > > > > > > The tasks worked fine on version 4.11. I have checked connection with > > sqlline - it works now and shows that version is 4.12. Moreover, I have > > noticed, that if limit the number of executors to one, the Spark's task > > executing successfully too! > > > > > > > > It looks like that executors running in parallel "interferes" each otherâs > > and could not acquire version's mutex. > > > > > > > > Any suggestions please? > > > > > > > > *final Connection connection = > > ConnectionUtil.getInputConnection(configuration, overridingProps);* > > > > *User class threw exception: org.apache.spark.SparkException: Job aborted > > due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent > > failure: Lost task 0.3 in stage 0.0 (TID 36, n7701-hdp005, executor 26): > > java.lang.RuntimeException: > > org.apache.phoenix.exception.UpgradeInProgressException: Cluster is being > > concurrently upgraded from 4.11.x to 4.12.x. Please retry establishing > > connection.* > > > > *at > > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan(PhoenixInputFormat.java:201)* > > > > *at > > org.apache.phoenix.mapreduce.PhoenixInputFormat.createRecordReader(PhoenixInputFormat.java:76)* > > > > *at > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.liftedTree1$1(NewHadoopRDD.scala:180)* > > > > *at > > org.apache.spark.rdd.NewHadoopRDD$$anon$1.<init>(NewHadoopRDD.scala:179)* > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:134)* > > > > *at org.apache.spark.rdd.NewHadoopRDD.compute(NewHadoopRDD.scala:69)* > > > > *at org.apache.phoenix.spark.PhoenixRDD.compute(PhoenixRDD.scala:64)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)* > > > > *at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:323)* > > > > *at org.apache.spark.rdd.RDD.iterator(RDD.scala:287)* > > > > *at > > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)* > > > > *at > > org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)* > > > > *at org.apache.spark.scheduler.Task.run(Task.scala:108)* > > > > *at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:335)* > > > > *at > > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)* > > > > *at > > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)* > > > > *at java.lang.Thread.run(Thread.java:745)* > > > > *Caused by: org.apache.phoenix.exception.UpgradeInProgressException: > > Cluster is being concurrently upgraded from 4.11.x to 4.12.x. Please retry > > establishing connection.* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.acquireUpgradeMutex(ConnectionQueryServicesImpl.java:3173)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.upgradeSystemTables(ConnectionQueryServicesImpl.java:2567)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2440)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl$12.call(ConnectionQueryServicesImpl.java:2360)* > > > > *at > > org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76)* > > > > *at > > org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2360)* > > > > *at > > org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:255)* > > > > *at > > org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:150)* > > > > *at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:221)* > > > > *at java.sql.DriverManager.getConnection(DriverManager.java:664)* > > > > *at java.sql.DriverManager.getConnection(DriverManager.java:208)* > > > > *at > > org.apache.phoenix.mapreduce.util.ConnectionUtil.getConnection(ConnectionUtil.java:98)* > > > > *at > > org.apache.phoenix.mapreduce.util.ConnectionUtil.getInputConnection(ConnectionUtil.java:57)* > > > > *at > > org.apache.phoenix.mapreduce.PhoenixInputFormat.getQueryPlan(PhoenixInputFormat.java:176)* > > > > *... 30 more* > > > Thank you for reply! Do you know when version 4.13 for HBASE 1.1 excpected?
Re: Spark & UpgradeInProgressException: Cluster is being concurrently upgraded from 4.11.x to 4.12.x
stepan.migu...@firstlinesoftware.com Sat, 11 Nov 2017 01:03:07 -0800
- Spark & UpgradeInProgressExceptio... Stepan Migunov
- Re: Spark & UpgradeInProgres... Mujtaba Chohan
- Re: Spark & UpgradeInProgres... stepan.migu...@firstlinesoftware.com
- Re: Spark & UpgradeInPro... James Taylor