Github user tgravescs commented on a diff in the pull request:

    https://github.com/apache/spark/pull/7463#discussion_r34997465
  
    --- Diff: docs/running-on-yarn.md ---
    @@ -68,9 +68,9 @@ In YARN terminology, executors and application masters 
run inside "containers".
     
         yarn logs -applicationId <app ID>
         
    -will print out the contents of all log files from all containers from the 
given application. You can also view the container log files directly in HDFS 
using the HDFS shell or API. The directory where they are located can be found 
by looking at your YARN configs (`yarn.nodemanager.remote-app-log-dir` and 
`yarn.nodemanager.remote-app-log-dir-suffix`).
    +will print out the contents of all log files from all containers from the 
given application. You can also view the container log files directly in HDFS 
using the HDFS shell or API. The directory where they are located can be found 
by looking at your YARN configs (`yarn.nodemanager.remote-app-log-dir` and 
`yarn.nodemanager.remote-app-log-dir-suffix`). The logs are also available on 
Web UI. You need have both the Spark history server and the MapReduce history 
server running and configure `yarn.log.server.url` in `yarn-site.xml` properly. 
The log URL on the Spark history server UI will redirect you to the MapReduce 
history server to show the aggregated logs.
     
    -When log aggregation isn't turned on, logs are retained locally on each 
machine under `YARN_APP_LOGS_DIR`, which is usually configured to `/tmp/logs` 
or `$HADOOP_HOME/logs/userlogs` depending on the Hadoop version and 
installation. Viewing logs for a container requires going to the host that 
contains them and looking in this directory.  Subdirectories organize log files 
by application ID and container ID.
    +When log aggregation isn't turned on, logs are retained locally on each 
machine under `YARN_APP_LOGS_DIR`, which is usually configured to `/tmp/logs` 
or `$HADOOP_HOME/logs/userlogs` depending on the Hadoop version and 
installation. Viewing logs for a container requires going to the host that 
contains them and looking in this directory.  Subdirectories organize log files 
by application ID and container ID. The logs are also available on Web UI even 
without running the MapReduce history server. 
    --- End diff --
    
    again can we be a bit more specific here and make it  Spark Web UI or again 
spark web ui under the executors tab.   Perhaps also rephrase a bit:
     and doesn't require running the MapReduce history server.  



---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to