[ https://issues.apache.org/jira/browse/HIVE-7155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14015820#comment-14015820 ]
Thejas M Nair commented on HIVE-7155: ------------------------------------- [~shanyu] One concern I have with the default value of templeton.mapper.memory.mb being set to 1GB is that the cluster might be having a larger mapreduce.map.memory.mb value, but webhcat would end up having a smaller limit for it. I think we should take into consideration the current setting for mapreduce.map.memory.mb in the cluster. If the default value of mapreduce.map.memory.mb is greater than the value of templeton.mapper.memory.mb, I think we should use that. ie, templeton.mapper.memory.mb should be used for setting a higher value if necessary. What has been a good value for templeton.mapper.memory.mb in your experience ? > WebHCat controller job exceeds container memory limit > ----------------------------------------------------- > > Key: HIVE-7155 > URL: https://issues.apache.org/jira/browse/HIVE-7155 > Project: Hive > Issue Type: Bug > Components: WebHCat > Affects Versions: 0.13.0 > Reporter: shanyu zhao > Assignee: shanyu zhao > Attachments: HIVE-7155.patch > > > Submit a Hive query on a large table via WebHCat results in failure because > the WebHCat controller job is killed by Yarn since it exceeds the memory > limit (set by mapreduce.map.memory.mb, defaults to 1GB): > {code} > INSERT OVERWRITE TABLE Temp_InjusticeEvents_2014_03_01_00_00 SELECT * from > Stage_InjusticeEvents where LogTimestamp > '2014-03-01 00:00:00' and > LogTimestamp <= '2014-03-01 01:00:00'; > {code} > We could increase mapreduce.map.memory.mb to solve this problem, but this way > we are changing this setting system wise. > We need to provide a WebHCat configuration to overwrite > mapreduce.map.memory.mb when submitting the controller job. -- This message was sent by Atlassian JIRA (v6.2#6252)