[
https://issues.apache.org/jira/browse/PIG-4443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14595756#comment-14595756
]
Ángel Álvarez commented on PIG-4443:
------------------------------------
I have a script in PIG that loads data from Hive using
org.apache.hive.hcatalog.pig.HCatLoader. This script works fine in Pig 0.14,
but in Pig 0.15 I'm getting this error:
Requested data length 160452289 is longer than maximum configured RPC length
67108864
In Pig 0.14 I had to deal with this issue too, but I could always make it work
by reducing the number of splits in the Hive tables created by Sqoop (using no
more than 60 splits). Is there any special configuration needed?
> Write inputsplits in Tez to disk if the size is huge and option to compress
> pig input splits
> --------------------------------------------------------------------------------------------
>
> Key: PIG-4443
> URL: https://issues.apache.org/jira/browse/PIG-4443
> Project: Pig
> Issue Type: Bug
> Affects Versions: 0.14.0
> Reporter: Rohini Palaniswamy
> Assignee: Rohini Palaniswamy
> Fix For: 0.15.0
>
> Attachments: PIG-4443-1.patch, PIG-4443-Fix-TEZ-2192-2.patch,
> PIG-4443-Fix-TEZ-2192.patch
>
>
> Pig sets the input split information in user payload and when running against
> a table with 10s of 1000s of partitions, DAG submission fails with
> java.io.IOException: Requested data length 305844060 is longer than maximum
> configured RPC length 67108864
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)