Hi, started history-server
Here is UI output:

   - *Event log directory:* file:/var/log/spark/applicationHistory/

No completed applications found!

Did you specify the correct logging directory? Please verify your setting
of spark.history.fs.logDirectory and whether you have the permissions to
access it.
It is also possible that your application did not run to completion or did
not stop the SparkContext.

Spark 1.2.0

I goto node where server runs and:

ls -la /var/log/spark/applicationHistory/

total 44

drwxrwxrwx 11 root      root    4096 Apr 17 14:50 .

drwxrwxrwx  3 cassandra root    4096 Apr 16 15:31 ..

drwxrwxrwx  2 vagrant   vagrant 4096 Apr 17 10:06 app-20150417100630-0000

drwxrwxrwx  2 vagrant   vagrant 4096 Apr 17 11:01 app-20150417110140-0001

drwxrwxrwx  2 vagrant   vagrant 4096 Apr 17 11:12 app-20150417111216-0002

drwxrwxrwx  2 vagrant   vagrant 4096 Apr 17 11:14 app-20150417111441-0003

drwxrwx---  2 vagrant   vagrant 4096 Apr 17 11:20 *app-20150417112028-0004*

drwxrwx---  2 vagrant   vagrant 4096 Apr 17 14:17 *app-20150417141733-0005*

drwxrwx---  2 vagrant   vagrant 4096 Apr 17 14:32 *app-20150417143237-0006*

drwxrwx---  2 vagrant   vagrant 4096 Apr 17 14:49 *app-20150417144902-0007*

drwxrwx---  2 vagrant   vagrant 4096 Apr 17 14:50 *app-20150417145025-0008*


So there are logs, but history-server doesn't want to display them.

I've checked workers, they are pointed to that dir also, I run app, I see
new log.


Here is history-server log output:

vagrant@dsenode01:/usr/lib/spark/logs$ cat
spark-root-org.apache.spark.deploy.history.HistoryServer-1-dsenode01.out

Spark assembly has been built with Hive, including Datanucleus jars on
classpath

Spark Command: java -cp
::/usr/lib/spark/sbin/../conf:/usr/lib/spark/lib/spark-assembly-1.2.0-hadoop2.4.0.jar:/usr/lib/spark/lib/datanucleus-api-jdo-3.2.6.jar:/usr/lib/spark/lib/datanucleus-rdbms-3.2.9.jar:/usr/lib/spark/lib/datanucleus-core-3.2.10.jar
-XX:MaxPermSize=128m -Dspark.akka.logLifecycleEvents=true
-Dspark.history.fs.logDirectory=/var/log/spark/applicationHistory
-Dspark.eventLog.enabled=true -Xms512m -Xmx512m
org.apache.spark.deploy.history.HistoryServer

========================================


Using Spark's default log4j profile:
org/apache/spark/log4j-defaults.properties

15/04/17 09:55:21 INFO HistoryServer: Registered signal handlers for [TERM,
HUP, INT]

15/04/17 09:55:21 INFO SecurityManager: Changing view acls to: root

15/04/17 09:55:21 INFO SecurityManager: Changing modify acls to: root

15/04/17 09:55:21 INFO SecurityManager: SecurityManager: authentication
disabled; ui acls disabled; users with view permissions: Set(root); users
with modify permissions: Set(root)

15/04/17 09:55:22 WARN NativeCodeLoader: Unable to load native-hadoop
library for your platform... using builtin-java classes where applicable

15/04/17 09:55:24 INFO Utils: Successfully started service on port 18080.

15/04/17 09:55:24 INFO HistoryServer: Started HistoryServer at
http://dsenode01:18080


What could be wrong with it?

Reply via email to