[
https://issues.apache.org/jira/browse/HIVE-15097?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16490475#comment-16490475
]
zhiwen wang commented on HIVE-15097:
------------------------------------
turn off map join will make your query normal!!!
hive.auto.convert.join=false
hive.ignore.mapjoin.hint=false
> Error: java.lang.RuntimeException: Error in configuring object Caused by:
> java.lang.OutOfMemoryError: GC overhead limit exceeded
> ---------------------------------------------------------------------------------------------------------------------------------
>
> Key: HIVE-15097
> URL: https://issues.apache.org/jira/browse/HIVE-15097
> Project: Hive
> Issue Type: Bug
> Environment: hive 1.2.1
> Reporter: chen cong
> Priority: Major
>
> hive> CREATE TABLE user_tag_pre AS
> > SELECT A.userid , A.tag , A.single_tag_sumcount/B.sum_tag_sumcount_avg
> as pre
> > FROM user_tag_sumcount A , user_tag_avg B
> > WHERE A.userid = B.userid;
> Query ID = ubuntu_20161101091124_753d511e-c7b8-4d31-bf97-cf0c04203ddc
> Total jobs = 1
> Execution log at:
> /tmp/ubuntu/ubuntu_20161101091124_753d511e-c7b8-4d31-bf97-cf0c04203ddc.log
> 2016-11-01 09:11:28 Starting to launch local task to process map join;
> maximum memory = 477626368
> 2016-11-01 09:11:30 Processing rows: 200000 Hashtable size:
> 199999 Memory usage: 106513696 percentage: 0.223
> 2016-11-01 09:11:31 Processing rows: 300000 Hashtable size:
> 299999 Memory usage: 92688064 percentage: 0.194
> 2016-11-01 09:11:31 Processing rows: 400000 Hashtable size:
> 399999 Memory usage: 152547912 percentage: 0.319
> 2016-11-01 09:11:31 Processing rows: 500000 Hashtable size:
> 499999 Memory usage: 149501320 percentage: 0.313
> 2016-11-01 09:11:32 Processing rows: 600000 Hashtable size:
> 599999 Memory usage: 170749448 percentage: 0.357
> 2016-11-01 09:11:32 Dump the side-table for tag: 1 with group count:
> 675226 into file:
> file:/tmp/ubuntu/9b365431-aea2-41fa-bebe-602435a22612/hive_2016-11-01_09-11-2
>
> 4_333_5456525982690482779-1/-local-10003/HashTable-Stage-4/MapJoin-mapfile01--.hashtable
> 2016-11-01 09:11:33 Uploaded 1 File to:
> file:/tmp/ubuntu/9b365431-aea2-41fa-bebe-602435a22612/hive_2016-11-01_09-11-24_333_5456525982690482779-1/-local-10003/HashTa
> ble-Stage-4/MapJoin-mapfile01--.hashtable
> (22302233 bytes)
> 2016-11-01 09:11:33 End of local task; Time Taken: 5.334 sec.
> Execution completed successfully
> MapredLocal task succeeded
> Launching Job 1 out of 1
> Number of reduce tasks is set to 0 since there's no reduce operator
> Starting Job = job_1477989556419_0006, Tracking URL =
> http://master:8088/proxy/application_1477989556419_0006/
> Kill Command = /home/ubuntu/cloud/hadoop-2.7.2/bin/hadoop job -kill
> job_1477989556419_0006
> Hadoop job information for Stage-4: number of mappers: 1; number of reducers: > 0
> 2016-11-01 09:11:43,429 Stage-4 map = 0%, reduce = 0%
> 2016-11-01 09:12:44,257 Stage-4 map = 0%, reduce = 0%
> 2016-11-01 09:13:44,706 Stage-4 map = 0%, reduce = 0%
> 2016-11-01 09:14:32,819 Stage-4 map = 100%, reduce = 0%
> Ended Job = job_1477989556419_0006 with errors
> Error during job, obtaining debugging information...
> Examining task ID: task_1477989556419_0006_m_000000 (and more) from job
> job_1477989556419_0006
> Task with the most failures(4):
> -----
> Task ID:
> task_1477989556419_0006_m_000000
> URL:
>
> http://master:8088/taskdetails.jsp?jobid=job_1477989556419_0006&tipid=task_1477989556419_0006_m_000000
> -----
> Diagnostic Messages for this Task:
> Error: java.lang.RuntimeException: Error in configuring object
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:112)
> at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:78)
> at
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:449)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1657)
> at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:158)
> Caused by: java.lang.reflect.InvocationTargetException
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
> ... 9 more
> Caused by: java.lang.RuntimeException: Error in configuring object
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:112)
> at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:78)
> at
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
> at org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:38)
> ... 14 more
> Caused by: java.lang.reflect.InvocationTargetException
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
> ... 17 more
> Caused by: java.lang.OutOfMemoryError: GC overhead limit exceeded
> at
> org.apache.hadoop.hive.ql.exec.persistence.MapJoinTableContainerSerDe.load(MapJoinTableContainerSerDe.java:82)
> at
> org.apache.hadoop.hive.ql.exec.mr.HashTableLoader.load(HashTableLoader.java:98)
> at
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.loadHashTable(MapJoinOperator.java:288)
> at
> org.apache.hadoop.hive.ql.exec.MapJoinOperator$1.call(MapJoinOperator.java:173)
> at
> org.apache.hadoop.hive.ql.exec.MapJoinOperator$1.call(MapJoinOperator.java:169)
> at
> org.apache.hadoop.hive.ql.exec.mr.ObjectCache.retrieve(ObjectCache.java:55)
> at
> org.apache.hadoop.hive.ql.exec.mr.ObjectCache.retrieveAsync(ObjectCache.java:63)
> at
> org.apache.hadoop.hive.ql.exec.MapJoinOperator.initializeOp(MapJoinOperator.java:166)
> at
> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:362)
> at
> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:481)
> at
> org.apache.hadoop.hive.ql.exec.Operator.initializeChildren(Operator.java:438)
> at
> org.apache.hadoop.hive.ql.exec.Operator.initialize(Operator.java:375)
> at
> org.apache.hadoop.hive.ql.exec.mr.ExecMapper.configure(ExecMapper.java:131)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
> at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:78)
> at
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
> at org.apache.hadoop.mapred.MapRunner.configure(MapRunner.java:38)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498)
> at
> org.apache.hadoop.util.ReflectionUtils.setJobConf(ReflectionUtils.java:109)
> at
> org.apache.hadoop.util.ReflectionUtils.setConf(ReflectionUtils.java:78)
> at
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:136)
> at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:449)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:343)
> at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:164)
> at java.security.AccessController.doPrivileged(Native Method)
> FAILED: Execution Error, return code 2 from
> org.apache.hadoop.hive.ql.exec.mr.MapRedTask
> MapReduce Jobs Launched:
> Stage-Stage-4: Map: 1 HDFS Read: 0 HDFS Write: 0 FAIL
> Total MapReduce CPU Time Spent: 0 msec
> How can I solve it ? Need help!!!
--
This message was sent by Atlassian JIRA
(v7.6.3#76005)