Hi Niranda,

I am running DAS with HBase and getting following error message when
running a spark query. I was able to fix this issue by adding the complete
file path of commons-codec_1.4.0.wso2v1.jar in add-to-spark-classpath.conf
file in the spark folder. Please look into the issue.

TID: [-1234] [] [2015-09-18 09:36:28,060]  INFO
{org.wso2.carbon.analytics.spark.core.internal.SparkAnalyticsExecutor} -
 Executed query: CREATE TEMPORARY TABLE memberstatus       USING
CarbonAnalytics       OPTIONS (tableName "STREAMING_DATA")
Time Elapsed: 0.548 seconds.
{org.wso2.carbon.analytics.spark.core.internal.SparkAnalyticsExecutor}
TID: [-1] [] [2015-09-18 09:36:31,979]  WARN
{org.apache.spark.scheduler.TaskSetManager} -  Lost task 0.0 in stage 0.0
(TID 0, 10.244.83.30): java.lang.NoClassDefFoundError:
org/apache/commons/codec/binary/Hex
at org.apache.hadoop.hbase.util.MD5Hash.getMD5AsHex(MD5Hash.java:64)
at
org.apache.hadoop.hbase.HRegionInfo.createRegionName(HRegionInfo.java:485)
at
org.apache.hadoop.hbase.HRegionInfo.createRegionName(HRegionInfo.java:415)
at org.apache.hadoop.hbase.HRegionInfo.<init>(HRegionInfo.java:340)
at org.apache.hadoop.hbase.HRegionInfo.convert(HRegionInfo.java:1036)
at org.apache.hadoop.hbase.HRegionInfo.parseFrom(HRegionInfo.java:1106)
at
org.apache.hadoop.hbase.HRegionInfo.parseFromOrNull(HRegionInfo.java:1073)
at
org.apache.hadoop.hbase.MetaTableAccessor.getHRegionInfo(MetaTableAccessor.java:795)
at
org.apache.hadoop.hbase.MetaTableAccessor.getRegionLocations(MetaTableAccessor.java:728)
at
org.apache.hadoop.hbase.MetaTableAccessor$1.visit(MetaTableAccessor.java:342)
at
org.apache.hadoop.hbase.MetaTableAccessor.fullScan(MetaTableAccessor.java:606)
at
org.apache.hadoop.hbase.MetaTableAccessor.tableExists(MetaTableAccessor.java:365)
at
org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:281)
at
org.wso2.carbon.analytics.datasource.hbase.HBaseAnalyticsRecordStore.tableExists(HBaseAnalyticsRecordStore.java:136)
at
org.wso2.carbon.analytics.datasource.hbase.HBaseAnalyticsRecordStore.get(HBaseAnalyticsRecordStore.java:352)
at
org.wso2.carbon.analytics.dataservice.core.AnalyticsDataServiceImpl.initIndexedTableStore(AnalyticsDataServiceImpl.java:218)
at
org.wso2.carbon.analytics.dataservice.core.AnalyticsDataServiceImpl.<init>(AnalyticsDataServiceImpl.java:153)
at
org.wso2.carbon.analytics.dataservice.core.AnalyticsServiceHolder.checkAndPopulateCustomAnalyticsDS(AnalyticsServiceHolder.java:75)
at
org.wso2.carbon.analytics.dataservice.core.AnalyticsServiceHolder.getAnalyticsDataService(AnalyticsServiceHolder.java:64)
at
org.wso2.carbon.analytics.spark.core.internal.ServiceHolder.getAnalyticsDataService(ServiceHolder.java:81)
at
org.wso2.carbon.analytics.spark.core.rdd.AnalyticsRDD.compute(AnalyticsRDD.java:82)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:35)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:277)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:244)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:63)
at org.apache.spark.scheduler.Task.run(Task.scala:70)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:213)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.ClassNotFoundException:
org.apache.commons.codec.binary.Hex
at java.net.URLClassLoader$1.run(URLClassLoader.java:366)
at java.net.URLClassLoader$1.run(URLClassLoader.java:355)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:354)
at java.lang.ClassLoader.loadClass(ClassLoader.java:425)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:308)
at java.lang.ClassLoader.loadClass(ClassLoader.java:358)
... 35 more
 {org.apache.spark.scheduler.TaskSetManager}
TID: [-1] [] [2015-09-18 09:36:32,364] ERROR
{org.apache.spark.scheduler.TaskSchedulerImpl} -  Lost executor 0 on
10.244.83.30: remote Rpc client disassociated
{org.apache.spark.scheduler.TaskSchedulerImpl}
TID: [-1] [] [2015-09-18 09:36:32,356]  WARN
{akka.remote.ReliableDeliverySupervisor} -  Association with remote system
[akka.tcp://[email protected]:13500] has failed, address is now
gated for [5000] ms. Reason is: [Disassociated].
{akka.remote.ReliableDeliverySupervisor}
TID: [-1] [] [2015-09-18 09:36:32,377]  WARN
{org.apache.spark.scheduler.TaskSetManager} -  Lost task 0.1 in stage 0.0
(TID 1, 10.244.83.30): ExecutorLostFailure (executor 0 lost)
{org.apache.spark.scheduler.TaskSetManager}
TID: [-1] [] [2015-09-18 09:36:32,361]  WARN
{akka.remote.ReliableDeliverySupervisor} -  Association with remote system
[akka.tcp://[email protected]:13500] has failed, address is now
gated for [5000] ms. Reason is: [Disassociated].
{akka.remote.ReliableDeliverySupervisor}
TID: [-1] [] [2015-09-18 09:36:32,385] ERROR
{org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend} -  Asked
to remove non-existent executor 0
{org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend}
TID: [-1] [] [2015-09-18 09:36:39,569] ERROR
{org.apache.spark.scheduler.TaskSchedulerImpl} -  Lost executor 1 on
10.244.83.30: remote Rpc client disassociated
{org.apache.spark.scheduler.TaskSchedulerImpl}
TID: [-1] [] [2015-09-18 09:36:39,569]  WARN
{akka.remote.ReliableDeliverySupervisor} -  Association with remote system
[akka.tcp://[email protected]:13500] has failed, address is now
gated for [5000] ms. Reason is: [Disassociated].
{akka.remote.ReliableDeliverySupervisor}
TID: [-1] [] [2015-09-18 09:36:39,569]  WARN
{akka.remote.ReliableDeliverySupervisor} -  Association with remote system
[akka.tcp://[email protected]:13500] has failed, address is now
gated for [5000] ms. Reason is: [Disassociated].
{akka.remote.ReliableDeliverySupervisor}
TID: [-1] [] [2015-09-18 09:36:39,596]  WARN
{org.apache.spark.scheduler.TaskSetManager} -  Lost task 0.3 in stage 0.0
(TID 3, 10.244.83.30): ExecutorLostFailure (executor 1 lost)
{org.apache.spark.scheduler.TaskSetManager}
TID: [-1] [] [2015-09-18 09:36:39,598] ERROR
{org.apache.spark.scheduler.TaskSetManager} -  Task 0 in stage 0.0 failed 4
times; aborting job {org.apache.spark.scheduler.TaskSetManager}
TID: [-1234] [] [2015-09-18 09:36:39,614]  INFO
{org.wso2.carbon.analytics.spark.core.internal.SparkAnalyticsExecutor} -
 Executed query: SELECT * from memberstatus
Time Elapsed: 11.55 seconds.
{org.wso2.carbon.analytics.spark.core.internal.SparkAnalyticsExecutor}
TID: [-1234] [] [2015-09-18 09:36:39,614] ERROR
{org.apache.axis2.rpc.receivers.RPCMessageReceiver} -  Job aborted due to
stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure:
Lost task 0.3 in stage 0.0 (TID 3, 10.244.83.30): ExecutorLostFailure
(executor 1 lost)
Driver stacktrace: {org.apache.axis2.rpc.receivers.RPCMessageReceiver}
java.lang.reflect.InvocationTargetException
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at
org.apache.axis2.rpc.receivers.RPCUtil.invokeServiceClass(RPCUtil.java:212)
at
org.apache.axis2.rpc.receivers.RPCMessageReceiver.invokeBusinessLogic(RPCMessageReceiver.java:117)
at
org.apache.axis2.receivers.AbstractInOutMessageReceiver.invokeBusinessLogic(AbstractInOutMessageReceiver.java:40)
at
org.apache.axis2.receivers.AbstractMessageReceiver.receive(AbstractMessageReceiver.java:110)
at org.apache.axis2.engine.AxisEngine.receive(AxisEngine.java:180)
at
org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:169)
at
org.apache.axis2.transport.local.LocalTransportReceiver.processMessage(LocalTransportReceiver.java:82)
at
org.wso2.carbon.core.transports.local.CarbonLocalTransportSender.finalizeSendWithToAddress(CarbonLocalTransportSender.java:45)
at
org.apache.axis2.transport.local.LocalTransportSender.invoke(LocalTransportSender.java:77)
at org.apache.axis2.engine.AxisEngine.send(AxisEngine.java:442)
at
org.apache.axis2.description.OutInAxisOperationClient.send(OutInAxisOperation.java:430)
at
org.apache.axis2.description.OutInAxisOperationClient.executeImpl(OutInAxisOperation.java:225)
at org.apache.axis2.client.OperationClient.execute(OperationClient.java:149)
at
org.wso2.carbon.analytics.spark.admin.stub.AnalyticsProcessorAdminServiceStub.execute(AnalyticsProcessorAdminServiceStub.java:912)
at
org.wso2.carbon.analytics.spark.ui.client.AnalyticsExecutionClient.executeScriptContent(AnalyticsExecutionClient.java:67)
at
org.apache.jsp.spark_002dmanagement.executeScript_005fajaxprocessor_jsp._jspService(executeScript_005fajaxprocessor_jsp.java:110)
at org.apache.jasper.runtime.HttpJspBase.service(HttpJspBase.java:70)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
at
org.apache.jasper.servlet.JspServletWrapper.service(JspServletWrapper.java:432)
at org.apache.jasper.servlet.JspServlet.serviceJspFile(JspServlet.java:395)
at org.apache.jasper.servlet.JspServlet.service(JspServlet.java:339)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
at org.wso2.carbon.ui.JspServlet.service(JspServlet.java:155)
at org.wso2.carbon.ui.TilesJspServlet.service(TilesJspServlet.java:80)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
at
org.eclipse.equinox.http.helper.ContextPathServletAdaptor.service(ContextPathServletAdaptor.java:37)
at
org.eclipse.equinox.http.servlet.internal.ServletRegistration.service(ServletRegistration.java:61)
at
org.eclipse.equinox.http.servlet.internal.ProxyServlet.processAlias(ProxyServlet.java:128)
at
org.eclipse.equinox.http.servlet.internal.ProxyServlet.service(ProxyServlet.java:68)
at javax.servlet.http.HttpServlet.service(HttpServlet.java:727)
at
org.wso2.carbon.tomcat.ext.servlet.DelegationServlet.service(DelegationServlet.java:68)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at
org.wso2.carbon.tomcat.ext.filter.CharacterSetFilter.doFilter(CharacterSetFilter.java:61)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122)
at
org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:504)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:170)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103)
at
org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99)
at
org.wso2.carbon.tomcat.ext.valves.CarbonTomcatValve$1.invoke(CarbonTomcatValve.java:47)
at
org.wso2.carbon.webapp.mgt.TenantLazyLoaderValve.invoke(TenantLazyLoaderValve.java:57)
at
org.wso2.carbon.event.receiver.core.internal.tenantmgt.TenantLazyLoaderValve.invoke(TenantLazyLoaderValve.java:48)
at
org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:47)
at
org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62)
at
org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:159)
at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:950)
at
org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:421)
at
org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1074)
at
org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:611)
at
org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1739)
at
org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:1698)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at
org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.spark.SparkException: Job aborted due to stage
failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task
0.3 in stage 0.0 (TID 3, 10.244.83.30): ExecutorLostFailure (executor 1
lost)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org
$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1273)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1264)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1263)
at
scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at
org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1263)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730)
at
org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:730)
at scala.Option.foreach(Option.scala:236)
at
org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:730)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1457)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1418)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
TID: [-1] [] [2015-09-18 09:36:39,600] ERROR
{org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend} -  Asked
to remove non-existent executor 1
{org.apache.spark.scheduler.cluster.SparkDeploySchedulerBackend}


Thank you!
-- 

*Pubudu Gunatilaka*
Software Engineer
WSO2, Inc.: http://wso2.com
lean.enterprise.middleware
mobile:  +94 77 4078049
_______________________________________________
Dev mailing list
[email protected]
http://wso2.org/cgi-bin/mailman/listinfo/dev

Reply via email to