This log is zeppelin server log, the root log should be in the spark
interpreter log. The file name is something like this :
zeppelin-interpreter-spark*.log

Dave Boyd <db...@incadencecorp.com> 于2019年3月15日周五 上午9:31写道:

> Jeff:
>
>     Running a simple spark.version paragraph I sometimes get this:
>
> INFO [2019-03-15 01:12:18,720] ({pool-2-thread-49}
> RemoteInterpreter.java[call]:142) - Open RemoteInterpreter
> org.apache.zeppelin.spark.SparkInterpreter
>  INFO [2019-03-15 01:12:18,721] ({pool-2-thread-49}
> RemoteInterpreter.java[pushAngularObjectRegistryToRemote]:436) - Push local
> angular object registry from ZeppelinServer to remote interpreter group
> spark:shared_process
>  WARN [2019-03-15 01:13:30,593] ({pool-2-thread-49}
> NotebookServer.java[afterStatusChange]:2316) - Job
> 20190207-030535_192412278 is finished, status: ERROR, exception: null,
> result: %text java.lang.IllegalStateException: Spark context stopped while
> waiting for backend
>         at
> org.apache.spark.scheduler.TaskSchedulerImpl.waitBackendReady(TaskSchedulerImpl.scala:614)
>         at
> org.apache.spark.scheduler.TaskSchedulerImpl.postStartHook(TaskSchedulerImpl.scala:169)
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:567)
>         at org.apache.spark.SparkContext.<init>(SparkContext.scala:117)
>         at
> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2336)
>         at org.apache.spark.SparkContext.getOrCreate(SparkContext.scala)
>         at
> org.apache.zeppelin.spark.Spark2Shims.setupSparkListener(Spark2Shims.java:38)
>         at
> org.apache.zeppelin.spark.NewSparkInterpreter.open(NewSparkInterpreter.java:120)
>         at
> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:62)
>         at
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>         at
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:616)
>         at org.apache.zeppelin.scheduler.Job.run(Job.java:188)
>         at
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
>         at
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>         at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>         at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>         at
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>         at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>         at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>         at java.lang.Thread.run(Thread.java:748)
>
>  INFO [2019-03-15 01:13:30,598] ({pool-2-thread-49}
> VFSNotebookRepo.java[save]:196) - Saving note:2E4D6HQ3F
>  INFO [2019-03-15 01:13:30,600] ({pool-2-thread-49}
> SchedulerFactory.java[jobFinished]:120) - Job 20190207-030535_192412278
> finished by scheduler
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter-spark:shared_process-shared_session
>
> When I run this spark sql paragraph:
>
> // DataStore params to a hypothetical GeoMesa Accumulo table
> val dsParams = Map(
>   "instanceId" -> "oedl",
>   "zookeepers" -> "oedevnode00,oedevnode01,oedevnode02",
>   "user"       -> "oe_user",
>   "password"   -> "XXXXXXX",
>   "tableName"  -> "CoalesceSearch")
>
> // Create DataFrame using the "geomesa" format
> val docdataFrame =
> spark.read.format("geomesa").options(dsParams).option("geomesa.feature",
> "oedocumentrecordset").load()
> docdataFrame.createOrReplaceTempView("documentview")
>
> Here is the complete stack trace:
>
> INFO [2019-03-15 01:07:21,569] ({pool-2-thread-43}
> Paragraph.java[jobRun]:380) - Run paragraph [paragraph_id:
> 20190222-204451_856915056, interpreter: , note_id: 2E6X2CDWW, user:
> anonymous]
>  WARN [2019-03-15 01:07:27,098] ({pool-2-thread-43}
> NotebookServer.java[afterStatusChange]:2316) - Job
> 20190222-204451_856915056 is finished, status: ERROR, exception: null,
> result: %text java.lang.IllegalStateException: Cannot call methods on a
> stopped SparkContext.
> This stopped SparkContext was created at:
>
>
> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:860)
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> java.lang.reflect.Method.invoke(Method.java:498)
>
> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.spark2CreateContext(BaseSparkScalaInterpreter.scala:259)
>
> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.createSparkContext(BaseSparkScalaInterpreter.scala:178)
>
> org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:89)
>
> org.apache.zeppelin.spark.NewSparkInterpreter.open(NewSparkInterpreter.java:102)
> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:62)
>
> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>
> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:616)
> org.apache.zeppelin.scheduler.Job.run(Job.java:188)
> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
> java.util.concurrent.FutureTask.run(FutureTask.java:266)
>
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>
> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>
> The currently active SparkContext was created at:
>
> (No active SparkContext.)
>
>   at org.apache.spark.SparkContext.assertNotStopped(SparkContext.scala:100)
>   at
> org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:716)
>   at
> org.apache.spark.SparkContext$$anonfun$parallelize$1.apply(SparkContext.scala:715)
>   at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>   at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
>   at org.apache.spark.SparkContext.withScope(SparkContext.scala:701)
>   at org.apache.spark.SparkContext.parallelize(SparkContext.scala:715)
>   at
> org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:74)
>   at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
>   at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:114)
>   at
> org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:135)
>   at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
>   at
> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:132)
>   at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:113)
>   at
> org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:87)
>   at org.apache.spark.sql.Dataset.<init>(Dataset.scala:185)
>   at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:64)
>   at org.apache.spark.sql.Dataset.org
> $apache$spark$sql$Dataset$$withPlan(Dataset.scala:2822)
>   at
> org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:2605)
>   ... 47 elided
>
>  INFO [2019-03-15 01:07:27,118] ({pool-2-thread-43}
> VFSNotebookRepo.java[save]:196) - Saving note:2E6X2CDWW
>  INFO [2019-03-15 01:07:27,124] ({pool-2-thread-43}
> SchedulerFactory.java[jobFinished]:120) - Job 20190222-204451_856915056
> finished by scheduler
> org.apache.zeppelin.interpreter.remote.RemoteInterpreter-spark:shared_process-shared_session
> On 3/14/19 9:02 PM, Jeff Zhang wrote:
>
> Hi Dave,
>
> Could you paste the full stacktrace ? You can find it in the spark
> interpreter log file which is located in ZEPPELIN_HOME/logs
>
> Xun Liu <neliu...@163.com> 于2019年3月15日周五 上午8:21写道:
>
>> Hi
>>
>> You can first execute a simple statement in spark, through sparksql, to
>> see if it can run normally in YARN.
>> If sparksql is running without problems, check the zeppelin and spark on
>> yarn issues.
>>
>> Also, what do you use for zeppelin-0.7.4? zeppelin-0.8.2? Is it a branch
>> that you maintain yourself?
>>
>> 在 2019年3月15日,上午6:31,Dave Boyd <db...@incadencecorp.com> 写道:
>>
>> All:
>>
>>    I have some code that worked fine in Zeppelin 0.7.4 but I am having
>> issues in 0.8.2 when going from spark master of local to yarn-client.  Yarn
>> client worked in 0.7.4.
>>
>> When my master is set to local[*] it runs just fine.  However, as soon as
>> I switch to yarn-client I get the Cannot call methods on a stopped
>> SparkContext error.   In looking at my yarn logs everything creates fine
>> and the job finishes without an error.  The executors start just fine
>> from what I get out of yarn logs.
>>
>> Any suggestions on where to look?   This happens with any note that trys
>> to run spark.
>>
>> If I try this very simple code:
>>
>> // Spark Version
>> spark.version
>>
>> I get this error:
>>
>> java.lang.IllegalStateException: Spark context stopped while waiting for
>> backend at
>> org.apache.spark.scheduler.TaskSchedulerImpl.waitBackendReady(TaskSchedulerImpl.scala:614)
>> at
>> org.apache.spark.scheduler.TaskSchedulerImpl.postStartHook(TaskSchedulerImpl.scala:169)
>> at org.apache.spark.SparkContext.<init>(SparkContext.scala:567) at
>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2313) at
>> org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:868)
>> at
>> org.apache.spark.sql.SparkSession$Builder$$anonfun$6.apply(SparkSession.scala:860)
>> at scala.Option.getOrElse(Option.scala:121) at
>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:860)
>> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.lang.reflect.Method.invoke(Method.java:498) at
>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.spark2CreateContext(BaseSparkScalaInterpreter.scala:259)
>> at
>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.createSparkContext(BaseSparkScalaInterpreter.scala:178)
>> at
>> org.apache.zeppelin.spark.SparkScala211Interpreter.open(SparkScala211Interpreter.scala:89)
>> at
>> org.apache.zeppelin.spark.NewSparkInterpreter.open(NewSparkInterpreter.java:102)
>> at
>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:62)
>> at
>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:69)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:616)
>> at org.apache.zeppelin.scheduler.Job.run(Job.java:188) at
>> org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:140)
>> at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>> at java.util.concurrent.FutureTask.run(FutureTask.java:266) at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180)
>> at
>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293)
>> at
>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>> at
>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>> at java.lang.Thread.run(Thread.java:748)
>>
>> What am I missing?
>>
>> --
>> ========= mailto:db...@incadencecorp.com <db...@incadencecorp.com> 
>> ============
>> David W. Boyd
>> VP,  Data Solutions
>> 10432 Balls Ford, Suite 240
>> Manassas, VA 20109
>> office:   +1-703-552-2862
>> cell:     +1-703-402-7908
>> ============== http://www.incadencecorp.com/ ============
>> ISO/IEC JTC1 WG9, editor ISO/IEC 20547 Big Data Reference Architecture
>> Chair ANSI/INCITS TC Big Data
>> Co-chair NIST Big Data Public Working Group Reference Architecture
>> First Robotic Mentor - FRC, FTC - www.iliterobotics.org
>> Board Member- USSTEM Foundation - www.usstem.org
>>
>> The information contained in this message may be privileged
>> and/or confidential and protected from disclosure.
>> If the reader of this message is not the intended recipient
>> or an employee or agent responsible for delivering this message
>> to the intended recipient, you are hereby notified that any
>> dissemination, distribution or copying of this communication
>> is strictly prohibited.  If you have received this communication
>> in error, please notify the sender immediately by replying to
>> this message and deleting the material from any computer.
>>
>>
>>
>>
>>
>
> --
> Best Regards
>
> Jeff Zhang
>
> --
> ========= mailto:db...@incadencecorp.com <db...@incadencecorp.com> 
> ============
> David W. Boyd
> VP,  Data Solutions
> 10432 Balls Ford, Suite 240
> Manassas, VA 20109
> office:   +1-703-552-2862
> cell:     +1-703-402-7908
> ============== http://www.incadencecorp.com/ ============
> ISO/IEC JTC1 WG9, editor ISO/IEC 20547 Big Data Reference Architecture
> Chair ANSI/INCITS TC Big Data
> Co-chair NIST Big Data Public Working Group Reference Architecture
> First Robotic Mentor - FRC, FTC - www.iliterobotics.org
> Board Member- USSTEM Foundation - www.usstem.org
>
> The information contained in this message may be privileged
> and/or confidential and protected from disclosure.
> If the reader of this message is not the intended recipient
> or an employee or agent responsible for delivering this message
> to the intended recipient, you are hereby notified that any
> dissemination, distribution or copying of this communication
> is strictly prohibited.  If you have received this communication
> in error, please notify the sender immediately by replying to
> this message and deleting the material from any computer.
>
>
>
>

-- 
Best Regards

Jeff Zhang

Reply via email to