Please use the `yarn logs -applicationId <id>` to retrieve the
logs. If you have enabled log aggregation, this will give you all
container logs concatenated.
Cheers,
Max
On Wed, Jun 15, 2016 at 12:24 AM, Theofilos Kakantousis
<t...@kth.se <mailto:t...@kth.se>> wrote:
Hi Max,
The runBlocking(..) problem was due to a Netty dependency
issue on my project, it works fine now :)
To pinpoint the logging issue, I just ran a single flink job
on yarn as per the documentation "./bin/flink run -m
yarn-cluster -yn 2 ./examples/streaming/Iteration.jar" and I
have the same issue.During the job I can see in the
containers the taskmanager logs, and a sample output from the
taskmanager.out is the following:
"cat
/srv/hadoop/logs/userlogs/application_1465901188070_0037/container_1465901188070_0037_01_000002/taskmanager.out
2> ((49,1),3)
2> ((25,11),4)
2> ((46,44),2
.."
However, the yarn aggregated log contains only the jobmanager
output. Is this expected or could it indicate a problem with
my hadoop logging configuration not picking up taskmanager logs?
Cheers,
Theofilos
On 6/13/2016 12:13 PM, Maximilian Michels wrote:
Hi Theofilos,
Flink doesn't send the local client output to the Yarn
cluster. I
think this will only change once we move the entire
execution of the
Job to the cluster framework. All output of the actual
Flink job
should be within the JobManager or TaskManager logs.
There is something wrong with the network communication
if the Client
doesn't return from `runBlocking(..)`. Would be
interesting to take a
look at the logs to find out why that could be.
Cheers,
Max
On Sat, Jun 11, 2016 at 1:53 PM, Theofilos Kakantousis
<t...@kth.se <mailto:t...@kth.se>> wrote:
Hi Robert,
Thanks for the prompt reply. I'm using the
IterateExample from Flink
examples. In the yarn log I get entries for the
YarnJobManager and
ExecutionGraph, but I was wondering if there is a way
to push all the
logging that the client produces into the yarn log.
Including the System.out
calls. Is there a way to modify the example to use a
logging framework to
achieve it?
Also when I submit the program using the Client
runBlocking method, although
I see in the taskmanager and jobmanager log that the
application has
finished, the runBlocking method does not return.
Should I call it in a
separate thread?
Cheers,
Theofilos
On 2016-06-10 22:12, Robert Metzger wrote:
Hi Theofilos,
how exactly are you writing the application output?
Are you using a logging framework?
Are you writing the log statements from the open(),
map(), invoke() methods
or from some constructors? (I'm asking since
different parts are executed on
the cluster and locally).
On Fri, Jun 10, 2016 at 4:00 PM, Theofilos
Kakantousis <t...@kth.se <mailto:t...@kth.se>> wrote:
Hi all,
Flink 1.0.3
Hadoop 2.4.0
When running a job on a Flink Cluster on Yarn,
the application output is
not included in the Yarn log. Instead, it is only
printed in the stdout from
where I run my program. For the jobmanager, I'm
using the log4j.properties
file from the flink/conf directory. Yarn log
aggregation is enabled and the
YarnJobManager log is printed in the yarn log.
The application is submitted
by a Flink Client to the FlinkYarnCluster using a
PackagedProgram.
Is this expected behavior and if so, is there a
way to include the
application output in the Yarn aggregated log?
Thanks!
Cheers,
Theofilos