Thank you Herman!

That was it.




-- 
Ruslan Dautkhanov

On Mon, Nov 14, 2016 at 7:23 AM, herman...@teeupdata.com <
herman...@teeupdata.com> wrote:

> You may check if you have hive-site.xml under zeppelin/spark  config
> folder…
>
> Thanks
> Herman.
>
>
>
> On Nov 14, 2016, at 03:07, Ruslan Dautkhanov <dautkha...@gmail.com> wrote:
>
> Dear Apache Zeppelin User group,
>
> Got Zeppelin running, but can't get %spark.sql interpreter running
> correctly,
> getting [1] in console output.
>
> Running latest Zeppelin (0.6.2), Spark 2.0, Hive 1.1, Hadoop 2.6, Java 7.
>
> My understanding is that Spark wants to initialize Hive Context, and Hive
> isn't getting its config correclty so tries to create a Derby local
> database
> in ZEPPELIN_HOME/bin/
> Nope, we don't use Derby for HMS and it's using a remote RDBMS, so clearly
> Hive/Spark isn't getting Hive settings correctly.
>
> Btw, adding
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> to zeppelin-env.sh didn't change anything.
>
> See [2] for zeppelin-env.sh .
>
>
> PS. I was trying to workaround this through setting
> export SPARK_CLASSPATH=$HIVE_HOME/hive/lib/*:/etc/hive/conf
> but it gives other problems - complains
> "Caused by: org.apache.spark.SparkException: Found both 
> *spark.driver.extraClassPath
> and* *SPARK_CLASSPATH*. *Use only the former*."
> Sidenote: that's interesting that spark-shell with the same SPARK_CLASSPATH
> gives opposite message -
> " *SPARK_CLASSPATH *was detected (set to '...').
> *This is deprecated in Spark 1.0*+. "
>
>
>
> Thank you,
> Ruslan Dautkhanov
>
>
> [0]
>
> export JAVA_HOME=/usr/java/java7
> export ZEPPELIN_MEM="-Xms1024m -Xmx2048m -XX:MaxPermSize=512m"
>
> export ZEPPELIN_LOG_DIR="/home/rdautkha/zeppelin/log"
> export ZEPPELIN_PID_DIR="/home/rdautkha/zeppelin/run"
> export ZEPPELIN_WAR_TEMPDIR="/home/rdautkha/zeppelin/tmp"
> export ZEPPELIN_NOTEBOOK_DIR="/home/rdautkha/zeppelin/notebooks"
>
> export SPARK_HOME=/opt/cloudera/parcels/SPARK2/lib/spark2
> export SPARK_SUBMIT_OPTIONS="--principal rdautkha...@corp.epsilon.com
> --keytab /home/rdautkha/.kt --conf spark.driver.memory=7g --conf
> spark.executor.cores=2 --conf spark.executor.memory=8g"
> export SPARK_APP_NAME="Zeppelin (test notebook 1)"
>
> export HADOOP_CONF_DIR=/etc/hadoop/conf
>
> export PYSPARK_PYTHON="/opt/cloudera/parcels/Anaconda/bin/python2"
> export PYTHONPATH="/opt/cloudera/parcels/SPARK2/lib/spark2/
> python:/opt/cloudera/parcels/SPARK2/lib/spark2/python/lib/
> py4j-0.10.3-src.zip"
> export MASTER="yarn-client"
>
> # these last three lines (and their combinations) were added in attempt to
> resolve the the problem:
> export SPARK_CLASSPATH=/opt/cloudera/parcels/CDH/lib/hive/lib/*:/
> etc/hive/conf
> export ZEPPELIN_SPARK_USEHIVECONTEXT=false
> export HIVE_CONF_DIR=/etc/hive/conf
>
>
>
> [1]
>
> Caused by: ERROR XBM0H: Directory 
> /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at 
> org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at 
> org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         ... 104 more
> ============= end nested exception, level (1) ===========
> ============= begin nested exception, level (2) ===========
> ERROR XBM0H: Directory /opt/zeppelin/zeppelin-0.6.2-bin-all/bin/metastore_db
> cannot be created.
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.iapi.error.StandardException.newException(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.
> StorageFactoryService$10.run(Unknown Source)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at 
> org.apache.derby.impl.services.monitor.StorageFactoryService.createServiceRoot(Unknown
> Source)
>         at 
> org.apache.derby.impl.services.monitor.BaseMonitor.bootService(Unknown
> Source)
>         at org.apache.derby.impl.services.monitor.BaseMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.services.monitor.FileMonitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.iapi.services.monitor.Monitor.
> createPersistentService(Unknown Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.createDatabase(Unknown
> Source)
>         at org.apache.derby.impl.jdbc.EmbedConnection.<init>(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.getNewEmbedConnection(Unknown
> Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.InternalDriver.connect(Unknown Source)
>         at org.apache.derby.jdbc.AutoloadedDriver.connect(Unknown Source)
>         at java.sql.DriverManager.getConnection(DriverManager.java:571)
>         at java.sql.DriverManager.getConnection(DriverManager.java:187)
>         at com.jolbox.bonecp.BoneCP.obtainRawInternalConnection(
> BoneCP.java:361)
>         at com.jolbox.bonecp.BoneCP.<init>(BoneCP.java:416)
>         at com.jolbox.bonecp.BoneCPDataSource.getConnection(
> BoneCPDataSource.java:120)
>         at org.datanucleus.store.rdbms.ConnectionFactoryImpl$
> ManagedConnectionImpl.getConnection(ConnectionFactoryImpl.java:501)
>         at org.datanucleus.store.rdbms.RDBMSStoreManager.<init>(
> RDBMSStoreManager.java:298)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(
> NativeConstructorAccessorImpl.java:57)
>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(
> DelegatingConstructorAccessorImpl.java:45)
>         at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>         at org.datanucleus.plugin.NonManagedPluginRegistry.
> createExecutableExtension(NonManagedPluginRegistry.java:631)
>         at org.datanucleus.plugin.PluginManager.createExecutableExtension(
> PluginManager.java:301)
>         at org.datanucleus.NucleusContext.createStoreManagerForPropertie
> s(NucleusContext.java:1187)
>         at org.datanucleus.NucleusContext.initialise(
> NucleusContext.java:356)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> freezeConfiguration(JDOPersistenceManagerFactory.java:775)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> createPersistenceManagerFactory(JDOPersistenceManagerFactory.java:333)
>         at org.datanucleus.api.jdo.JDOPersistenceManagerFactory.
> getPersistenceManagerFactory(JDOPersistenceManagerFactory.java:202)
>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>         at sun.reflect.NativeMethodAccessorImpl.invoke(
> NativeMethodAccessorImpl.java:57)
>         at sun.reflect.DelegatingMethodAccessorImpl.invoke(
> DelegatingMethodAccessorImpl.java:43)
>         at java.lang.reflect.Method.invoke(Method.java:606)
>         at javax.jdo.JDOHelper$16.run(JDOHelper.java:1965)
>         at java.security.AccessController.doPrivileged(Native Method)
>         at javax.jdo.JDOHelper.invoke(JDOHelper.java:1960)
>         at javax.jdo.JDOHelper.invokeGetPersistenceManagerFac
> toryOnImplementation(JDOHelper.java:1166)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:808)
>         at javax.jdo.JDOHelper.getPersistenceManagerFactory(
> JDOHelper.java:701)
>         at org.apache.hadoop.hive.metastore.ObjectStore.getPMF(
> ObjectStore.java:411)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> getPersistenceManager(ObjectStore.java:440)
>         at org.apache.hadoop.hive.metastore.ObjectStore.
> initialize(ObjectStore.java:335)
>         at org.apache.hadoop.hive.metastore.ObjectStore.setConf(
> ObjectStore.java:291)
>         at org.apache.hadoop.util.ReflectionUtils.setConf(
> ReflectionUtils.java:73)
>         at org.apache.hadoop.util.ReflectionUtils.newInstance(
> ReflectionUtils.java:133)
>         at org.apache.hadoop.hive.metastore.RawStoreProxy.<init>
> (RawStoreProxy.java:57)
>
>
>
>
>

Reply via email to