Your best option if you are using spark is here: https://spark.apache.org/docs/latest/monitoring.html#viewing-after-the-fact
Standard, you can have access to the log via: * the web ui * yarn log * or the file system As far as I know, there is no simple call rest api. The cli logs implementation of Yarn can be found here if you want to have a look: https://github.com/apache/hadoop/blob/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/LogsCLI.java Met vriendelijke groeten, Regards, Cordialement, Nicolas GERARD On Tue, 27 Nov 2018 at 23:11, Lian Jiang <jiangok2...@gmail.com> wrote: > Any idea? or I should ask another user group? Thanks. > > On Mon, Nov 26, 2018 at 2:02 PM Lian Jiang <jiangok2...@gmail.com> wrote: > >> On HDP3, I cannot get the full log of a failing spark job by using yarn >> api: >> >> curl -k -u guest:"" -X GET https:// >> myhost.com/gateway/ui/resourcemanager/v1/cluster/apps/ >> <https://confluence.oci.oraclecorp.com/display/BDW/dw-knox-prod1.us-phoenix-1.oracleiaas.com/gateway/ui/resourcemanager/v1/cluster/apps/> >> {applicationId} >> >> This means the job owner has to ssh to the cluster to run "yarn logs" >> command to get the full log. Is this expected? How can I get the full spark >> log without sshing to the cluster? Appreciate your help. >> >> >> >>