[ 
https://issues.apache.org/jira/browse/GIRAPH-246?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Reisman updated GIRAPH-246:
-------------------------------

    Attachment: GIRAPH-246-NEW-FIX-2.patch

This includes one more extra call to progress() and is otherwise just like the 
first "NEW-FIX" patch.

Neither trunk nor any of these PredicateLock fixes allow std err and std out 
logs reach the HTML Mapper Detail pages for some reason, the old rebase does. 
The only times my old rebase blew up and "timed out" today is when Netty blew 
up under stress tests, as the Mapper Details revealed.

This had "timeouts" occasionally in the same circumstances but there was no log 
to confirm why (which is why I doubted the earlier "new fix" patch.) We can 
commit this and hope it was just Netty blowing up during stress tests, if we 
insist on using a PredicateLock solution. This as closely replicates the good 
behavior of the old rebase patch as any I've tested yet.

Alternately, I can upload the latest rebase of the original 246 and we can 
commit that and wait for a better fix, but if you guys are having trouble 
replicating this problem then chances are shooting in the dark for a fix to 
replace the rebase will not happen, and I know you don't like seeing the 
progress calls in BspServiceWorker.

Either way is fine at this point, I will be running a lot of code over the next 
few days/nights as I get clear windows to do so, but  I am 100% sure after 
running trunk a bunch today as I was doing A/B tests on these that the current 
fix does not work for us here. I'd love to pick one (the rebase or this) and 
commit, its at least a step forward from what we have which is just not 
functioning for us at all. users here are getting sick of manually adding 
patches  to run Giraph on our job queue ;)

                
> Periodic worker calls to context.progress() will prevent timeout on some 
> Hadoop clusters during barrier waits
> -------------------------------------------------------------------------------------------------------------
>
>                 Key: GIRAPH-246
>                 URL: https://issues.apache.org/jira/browse/GIRAPH-246
>             Project: Giraph
>          Issue Type: Improvement
>          Components: bsp
>    Affects Versions: 0.2.0
>            Reporter: Eli Reisman
>            Assignee: Eli Reisman
>            Priority: Minor
>              Labels: hadoop, patch
>             Fix For: 0.2.0
>
>         Attachments: GIRAPH-246-10.patch, GIRAPH-246-11.patch, 
> GIRAPH-246-1.patch, GIRAPH-246-2.patch, GIRAPH-246-3.patch, 
> GIRAPH-246-4.patch, GIRAPH-246-5.patch, GIRAPH-246-6.patch, 
> GIRAPH-246-7.patch, GIRAPH-246-7_rebase1.patch, GIRAPH-246-8.patch, 
> GIRAPH-246-9.patch, GIRAPH-246-NEW-FIX-2.patch, GIRAPH-246-NEW-FIX.patch
>
>
> This simple change creates a command-line configurable option in GiraphJob to 
> control the time between calls to context().progress() that allows workers to 
> avoid timeouts during long data load-ins in which some works complete their 
> input split reads much faster than others, or finish a super step faster. I 
> found this allowed jobs that were large-scale but with low memory overhead to 
> complete even when they would previously time out during runs on a Hadoop 
> cluster. Timeout is still possible when the worker crashes or runs out of 
> memory or has other GC or RPC trouble that is legitimate, but prevents 
> unintentional crashes when the worker is actually still healthy.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to