[ 
https://issues.apache.org/jira/browse/PIG-3356?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13684627#comment-13684627
 ] 

xplenty commented on PIG-3356:
------------------------------

A solution to this issue would be adding a call to:
PigStatusReporter.getInstance().progress();
for each file opened, for instance in the PigRecordReader in nextKeyValue().
this way Hadoop knows the pig job is still running.
                
> Timeout when processing large amounts of empty files.
> -----------------------------------------------------
>
>                 Key: PIG-3356
>                 URL: https://issues.apache.org/jira/browse/PIG-3356
>             Project: Pig
>          Issue Type: Bug
>          Components: impl
>    Affects Versions: 0.11.1
>            Reporter: xplenty
>            Priority: Minor
>              Labels: hadoop, progress, timeout,
>
> When processing large amounts of empty files (for example traversing empty 
> files on S3) takes a long time and Pig tasks don't report progress and end up 
> being killed for being unresponsive.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira

Reply via email to