Hi Hubert,

In the original log file, I see hbase-server and hbase-common are on the
spark command:

--jars 
*/usr/lib/hbase/lib/hbase-common-1.4.2.jar*,*/usr/lib/hbase/lib/hbase-server-1.4.2.jar*,/usr/lib/hbase/lib/hbase-client-1.4.2.jar,/usr/lib/hbase/lib/hbase-protocol-1.4.2.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.2.jar,/usr/lib/hbase/lib/htrace-core-3.1.0-incubating.jar,/usr/lib/hbase/lib/metrics-core-2.2.0.jar,/usr/lib/hbase/lib/hbase-hadoop-compat-1.4.2.jar,/usr/lib/hbase/lib/hbase-hadoop2-compat-1.4.2.jar,


Did you add other jars?


Best regards,

Shaofeng Shi 史少锋
Apache Kylin PMC
Work email: [email protected]
Kyligence Inc: https://kyligence.io/

Apache Kylin FAQ: https://kylin.apache.org/docs/gettingstarted/faq.html
Join Kylin user mail group: [email protected]
Join Kylin dev mail group: [email protected]




hubert stefani <[email protected]> 于2019年1月11日周五 下午6:20写道:

>  indeed. By adding the hbase-server and hbase common jars the pb seems to
> be fixed.
>
>
>     Le vendredi 11 janvier 2019 à 11:11:40 UTC+1, ShaoFeng Shi <
> [email protected]> a écrit :
>
>  It seems missing some HBase class; The HFile.class should be in
> hbase-server-<version>.jar, not sure whether it is EMR's package issue. You
> can unzip the HBase jar files to see which jar has the class, and then add
> it to spark/lib folder.
>
> java.lang.NoClassDefFoundError: Could not initialize class
> org.apache.hadoop.hbase.io.hfile.HFile
>     at
> org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.getNewWriter(HFileOutputFormat2.java:305)
>     at
> org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:229)
>     at
> org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2$1.write(HFileOutputFormat2.java:167)
>     at org.apache.spark.internal.io
> .HadoopMapReduceWriteConfigUtil.write(SparkHadoopWriter.scala:356)
>     at org.apache.spark.internal.io
> .SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:130)
>     at org.apache.spark.internal.io
> .SparkHadoopWriter$$anonfun$4.apply(SparkHadoopWriter.scala:127)
>     at
> org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1415)
>
>
>
> Best regards,
>
> Shaofeng Shi 史少锋
> Apache Kylin PMC
> Work email: [email protected]
> Kyligence Inc: https://kyligence.io/
>
> Apache Kylin FAQ: https://kylin.apache.org/docs/gettingstarted/faq.html
> Join Kylin user mail group: [email protected]
> Join Kylin dev mail group: [email protected]
>
>
>
>
> hubert stefani <[email protected]> 于2019年1月11日周五 下午4:52写道:
>
> > hello,
> >
> > we are testing the 2.6 RC and we are facing a systematic issue when
> > building cubes with spark engine (even with sample cube), whereas the
> > MapReduce engin succeeds.
> >
> > The job process fails at step #8 Step Name: Convert Cuboid Data to HFile
> > with the following error  (full log is available as attachment):
> >
> > ClassNotFoundException: org.apache.hadoop.hbase.metrics.MetricRegistry
> >
> > We run kylin on AWS EMR 5.13 (it failed also with 5.17).
> >
> > Do you have any idea of the reasons why it happens ?
> > Hubert
> >

Reply via email to