[
https://issues.apache.org/jira/browse/HIVE-10837?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15123523#comment-15123523
]
David Watzke commented on HIVE-10837:
-------------------------------------
Cloudera recommends increasing the HiveServer2's heap size
http://www.cloudera.com/documentation/enterprise/latest/topics/admin_hos_troubleshooting.html
so that's what I've done (from 256M to 8G) and it seems that it helped.
BTW: I noticed that Cloudera increased the default heap size for hive roles in
CDH 5.5 which is maybe related to this.
> Running large queries (inserts) fails and crashes hiveserver2
> -------------------------------------------------------------
>
> Key: HIVE-10837
> URL: https://issues.apache.org/jira/browse/HIVE-10837
> Project: Hive
> Issue Type: Bug
> Environment: Hive 1.1.0 on RHEL with Cloudera (cdh5.4.0)
> Reporter: Patrick McAnneny
> Priority: Critical
>
> When running a large insert statement through beeline or pyhs2, a thrift
> error is returned and hiveserver2 crashes.
> I ran into this with large insert statements -- my initial failing query was
> around 6million characters. After further testing however it seems like the
> failure threshold is based on number of inserted rows rather than the query's
> size in characters. My testing shows the failure threshold between 199,000
> and 230,000 inserted rows.
> The thrift error is as follows:
> Error: org.apache.thrift.transport.TTransportException:
> java.net.SocketException: Broken pipe (state=08S01,code=0)
> Also note for anyone that tests this issue - when testing different queries I
> ran into https://issues.apache.org/jira/browse/HIVE-10836
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)