The marksweep value is very high, the scavenge very low. If this helps ;-)



> Am 08.09.2015 um 11:27 schrieb Robert Metzger <rmetz...@apache.org>:
> 
> It is in the "Information" column: http://i.imgur.com/rzxxURR.png
> In the screenshot, the two GCs only spend 84 and 25 ms.
> 
>> On Tue, Sep 8, 2015 at 10:34 AM, Rico Bergmann <i...@ricobergmann.de> wrote:
>> Where can I find these information? I can see the memory usage and cpu load. 
>> But where are the information on the GC?
>> 
>> 
>> 
>>> Am 08.09.2015 um 09:34 schrieb Robert Metzger <rmetz...@apache.org>:
>>> 
>>> The webinterface of Flink has a tab for the TaskManagers. There, you can 
>>> also see how much time the JVM spend with garbage collection.
>>> Can you check whether the number of GC calls + the time spend goes up after 
>>> 30 minutes?
>>> 
>>>> On Tue, Sep 8, 2015 at 8:37 AM, Rico Bergmann <i...@ricobergmann.de> wrote:
>>>> Hi!
>>>> 
>>>> I also think it's a GC problem. In the KeySelector I don't instantiate any 
>>>> object. It's a simple toString method call. 
>>>> In the mapWindow I create new objects. But I'm doing the same in other map 
>>>> operators, too. They don't slow down the execution. Only with this 
>>>> construct the execution is slowed down. 
>>>> 
>>>> I watched on the memory footprint of my program. Once with the code 
>>>> construct I wrote and once without. The memory characteristic were the 
>>>> same. The CPU usage also ... 
>>>> 
>>>> I don't have an explanation. But I don't think it comes from my operator 
>>>> functions ...
>>>> 
>>>> Cheers Rico. 
>>>> 
>>>> 
>>>> 
>>>>> Am 07.09.2015 um 22:43 schrieb Martin Neumann <mneum...@sics.se>:
>>>>> 
>>>>> Hej,
>>>>> 
>>>>> This sounds like it could be a garbage collection problem. Do you 
>>>>> instantiate any classes inside any of the operators (e.g. in the 
>>>>> KeySelector). You can also try to run it locally and use something like 
>>>>> jstat to rule this out.
>>>>> 
>>>>> cheers Martin
>>>>> 
>>>>>> On Mon, Sep 7, 2015 at 12:00 PM, Rico Bergmann <i...@ricobergmann.de> 
>>>>>> wrote:
>>>>>> Hi!
>>>>>> 
>>>>>> While working with grouping and windowing I encountered a strange 
>>>>>> behavior. I'm doing:
>>>>>>> dataStream.groupBy(KeySelector).window(Time.of(x, 
>>>>>>> TimeUnit.SECONDS)).mapWindow(toString).flatten()
>>>>>> 
>>>>>> When I run the program containing this snippet it initially outputs data 
>>>>>> at a rate around 150 events per sec. (That is roughly the input rate for 
>>>>>> the program). After about 10-30 minutes the rate drops down below 5 
>>>>>> events per sec. This leads to event delivery offsets getting bigger and 
>>>>>> bigger ... 
>>>>>> 
>>>>>> Any explanation for this? I know you are reworking the streaming API. 
>>>>>> But it would be useful to know, why this happens ...
>>>>>> 
>>>>>> Cheers. Rico. 
> 

Reply via email to