You should also see TaskManager output in the logs. I just verified this
using Flink 1.0.3 with Hadoop 2.7.1. I executed the Iterate example and it
aggregated correctly including the TaskManager logs.

I'm wondering, is there anything in the Hadoop logs of the
Resourcemanager/Nodemanager that could indicate a transfer failure?

On Wed, Jun 15, 2016 at 10:18 AM, Theofilos Kakantousis <t...@kth.se> wrote:

> Hi,
>
> By yarn aggregated log I mean Yarn log aggregation is enabled and the log
> I'm referring to is the one returned by `yarn logs -applicationId <id>`.
> When running a Spark job for example on the same setup, the yarn aggregated
> log contains all the information printed out by the application.
>
> Cheers,
> Theofilos
>
>
> On 6/15/2016 10:14 AM, Maximilian Michels wrote:
>
> Please use the `yarn logs -applicationId <id>` to retrieve the logs. If
> you have enabled log aggregation, this will give you all container logs
> concatenated.
>
> Cheers,
> Max
>
> On Wed, Jun 15, 2016 at 12:24 AM, Theofilos Kakantousis < <t...@kth.se>
> t...@kth.se> wrote:
>
>> Hi Max,
>>
>> The runBlocking(..) problem was due to a Netty dependency issue on my
>> project, it works fine now :)
>>
>> To pinpoint the logging issue, I just ran a single flink job on yarn as
>> per the documentation "./bin/flink run -m yarn-cluster -yn 2
>> ./examples/streaming/Iteration.jar" and I have the same issue.During the
>> job I can see in the containers the taskmanager logs, and a sample output
>> from the taskmanager.out is the following:
>> "cat
>> /srv/hadoop/logs/userlogs/application_1465901188070_0037/container_1465901188070_0037_01_000002/taskmanager.out
>>
>> 2> ((49,1),3)
>> 2> ((25,11),4)
>> 2> ((46,44),2
>> .."
>>
>> However, the yarn aggregated log contains only the jobmanager output. Is
>> this expected or could it indicate a problem with my hadoop logging
>> configuration not picking up taskmanager logs?
>>
>> Cheers,
>> Theofilos
>>
>>
>> On 6/13/2016 12:13 PM, Maximilian Michels wrote:
>>
>>> Hi Theofilos,
>>>
>>> Flink doesn't send the local client output to the Yarn cluster. I
>>> think this will only change once we move the entire execution of the
>>> Job to the cluster framework. All output of the actual Flink job
>>> should be within the JobManager or TaskManager logs.
>>>
>>> There is something wrong with the network communication if the Client
>>> doesn't return from `runBlocking(..)`. Would be interesting to take a
>>> look at the logs to find out why that could be.
>>>
>>> Cheers,
>>> Max
>>>
>>>
>>> On Sat, Jun 11, 2016 at 1:53 PM, Theofilos Kakantousis <t...@kth.se>
>>> wrote:
>>>
>>>> Hi Robert,
>>>>
>>>> Thanks for the prompt reply. I'm using the IterateExample from Flink
>>>> examples. In the yarn log I get entries for the YarnJobManager and
>>>> ExecutionGraph, but I was wondering if there is a way to push all the
>>>> logging that the client produces into the yarn log. Including the
>>>> System.out
>>>> calls. Is there a way to modify the example to use a logging framework
>>>> to
>>>> achieve it?
>>>>
>>>> Also when I submit the program using the Client runBlocking method,
>>>> although
>>>> I see in the taskmanager and jobmanager log that the application has
>>>> finished, the runBlocking method does not return. Should I call it in a
>>>> separate thread?
>>>>
>>>> Cheers,
>>>> Theofilos
>>>>
>>>> On 2016-06-10 22:12, Robert Metzger wrote:
>>>>
>>>> Hi Theofilos,
>>>>
>>>> how exactly are you writing the application output?
>>>> Are you using a logging framework?
>>>> Are you writing the log statements from the open(), map(), invoke()
>>>> methods
>>>> or from some constructors? (I'm asking since different parts are
>>>> executed on
>>>> the cluster and locally).
>>>>
>>>> On Fri, Jun 10, 2016 at 4:00 PM, Theofilos Kakantousis <t...@kth.se>
>>>> wrote:
>>>>
>>>>> Hi all,
>>>>>
>>>>> Flink 1.0.3
>>>>> Hadoop 2.4.0
>>>>>
>>>>> When running a job on a Flink Cluster on Yarn, the application output
>>>>> is
>>>>> not included in the Yarn log. Instead, it is only printed in the
>>>>> stdout from
>>>>> where I run my program.  For the jobmanager, I'm using the
>>>>> log4j.properties
>>>>> file from the flink/conf directory. Yarn log aggregation is enabled
>>>>> and the
>>>>> YarnJobManager log is printed in the yarn log. The application is
>>>>> submitted
>>>>> by a Flink Client to the FlinkYarnCluster using a PackagedProgram.
>>>>>
>>>>> Is this expected behavior and if so, is there a way to include the
>>>>> application output in the Yarn aggregated log? Thanks!
>>>>>
>>>>> Cheers,
>>>>> Theofilos
>>>>>
>>>>>
>>>>
>>
>
>

Reply via email to