Created https://issues.apache.org/jira/browse/SPARK-11798



On Wed, Nov 18, 2015 at 9:42 AM, Josh Rosen <joshro...@databricks.com>
wrote:

> Can you file a JIRA issue to help me triage this further? Thanks!
>
> On Tue, Nov 17, 2015 at 4:08 PM Jeff Zhang <zjf...@gmail.com> wrote:
>
>> Sure, hive profile is enabled.
>>
>> On Wed, Nov 18, 2015 at 6:12 AM, Josh Rosen <joshro...@databricks.com>
>> wrote:
>>
>>> Is the Hive profile enabled? I think it may need to be turned on in
>>> order for those JARs to be deployed.
>>>
>>> On Tue, Nov 17, 2015 at 2:27 AM Jeff Zhang <zjf...@gmail.com> wrote:
>>>
>>>> BTW, After I revert  SPARK-7841, I can see all the jars under
>>>> lib_managed/jars
>>>>
>>>> On Tue, Nov 17, 2015 at 2:46 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>
>>>>> Hi Josh,
>>>>>
>>>>> I notice the comments in https://github.com/apache/spark/pull/9575 said
>>>>> that Datanucleus related jars will still be copied to
>>>>> lib_managed/jars. But I don't see any jars under lib_managed/jars.
>>>>> The weird thing is that I see the jars on another machine, but could not
>>>>> see jars on my laptop even after I delete the whole spark project and 
>>>>> start
>>>>> from scratch. Does it related with environments ? I try to add the
>>>>> following code in SparkBuild.scala to track the issue, it shows that the
>>>>> jars is empty. Any thoughts on that ?
>>>>>
>>>>>
>>>>> deployDatanucleusJars := {
>>>>>       val jars: Seq[File] = (fullClasspath in
>>>>> assembly).value.map(_.data)
>>>>>         .filter(_.getPath.contains("org.datanucleus"))
>>>>>       // this is what I added
>>>>>       println("*********************************************")
>>>>>       println("fullClasspath:"+fullClasspath)
>>>>>       println("assembly:"+assembly)
>>>>>       println("jars:"+jars.map(_.getAbsolutePath()).mkString(","))
>>>>>       //
>>>>>
>>>>>
>>>>> On Mon, Nov 16, 2015 at 4:51 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>
>>>>>> This is the exception I got
>>>>>>
>>>>>> 15/11/16 16:50:48 WARN metastore.HiveMetaStore: Retrying creating
>>>>>> default database after error: Class
>>>>>> org.datanucleus.api.jdo.JDOPersistenceManagerFactory was not found.
>>>>>> javax.jdo.JDOFatalUserException: Class
>>>>>> org.datanucleus.api.jdo.JDOPersistenceManagerFactory was not found.
>>>>>> at
>>>>>> javax.jdo.JDOHelper.invokeGetPersistenceManagerFactoryOnImplementation(JDOHelper.java:1175)
>>>>>> at
>>>>>> javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:808)
>>>>>> at
>>>>>> javax.jdo.JDOHelper.getPersistenceManagerFactory(JDOHelper.java:701)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.ObjectStore.getPMF(ObjectStore.java:365)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.ObjectStore.getPersistenceManager(ObjectStore.java:394)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.ObjectStore.initialize(ObjectStore.java:291)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.ObjectStore.setConf(ObjectStore.java:258)
>>>>>> at
>>>>>> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:73)
>>>>>> at
>>>>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:133)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.RawStoreProxy.<init>(RawStoreProxy.java:57)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.RawStoreProxy.getProxy(RawStoreProxy.java:66)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.newRawStore(HiveMetaStore.java:593)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.getMS(HiveMetaStore.java:571)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.createDefaultDB(HiveMetaStore.java:620)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStore$HMSHandler.init(HiveMetaStore.java:461)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.<init>(RetryingHMSHandler.java:66)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.RetryingHMSHandler.getProxy(RetryingHMSHandler.java:72)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStore.newRetryingHMSHandler(HiveMetaStore.java:5762)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(HiveMetaStoreClient.java:199)
>>>>>> at
>>>>>> org.apache.hadoop.hive.ql.metadata.SessionHiveMetaStoreClient.<init>(SessionHiveMetaStoreClient.java:74)
>>>>>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>>>>> Method)
>>>>>> at
>>>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)
>>>>>> at
>>>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>>>>> at java.lang.reflect.Constructor.newInstance(Constructor.java:526)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.MetaStoreUtils.newInstance(MetaStoreUtils.java:1521)
>>>>>> at
>>>>>> org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.<init>(RetryingMetaStoreClient.java:86)
>>>>>>
>>>>>> On Mon, Nov 16, 2015 at 4:47 PM, Jeff Zhang <zjf...@gmail.com> wrote:
>>>>>>
>>>>>>> It's about the datanucleus related jars which is needed by spark
>>>>>>> sql. Without these jars, I could not call data frame related api ( I 
>>>>>>> make
>>>>>>> HiveContext enabled)
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Nov 16, 2015 at 4:10 PM, Josh Rosen <
>>>>>>> joshro...@databricks.com> wrote:
>>>>>>>
>>>>>>>> As of https://github.com/apache/spark/pull/9575, Spark's build
>>>>>>>> will no longer place every dependency JAR into lib_managed. Can you say
>>>>>>>> more about how this affected spark-shell for you (maybe share a 
>>>>>>>> stacktrace)?
>>>>>>>>
>>>>>>>> On Mon, Nov 16, 2015 at 12:03 AM, Jeff Zhang <zjf...@gmail.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>>
>>>>>>>>> Sometimes, the jars under lib_managed is missing. And after I
>>>>>>>>> rebuild the spark, the jars under lib_managed is still not 
>>>>>>>>> downloaded. This
>>>>>>>>> would cause the spark-shell fail due to jars missing. Anyone has hit 
>>>>>>>>> this
>>>>>>>>> weird issue ?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Best Regards
>>>>>>>>>
>>>>>>>>> Jeff Zhang
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> --
>>>>>>> Best Regards
>>>>>>>
>>>>>>> Jeff Zhang
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> --
>>>>>> Best Regards
>>>>>>
>>>>>> Jeff Zhang
>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Best Regards
>>>>>
>>>>> Jeff Zhang
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Best Regards
>>>>
>>>> Jeff Zhang
>>>>
>>>
>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>


-- 
Best Regards

Jeff Zhang

Reply via email to